2025-12-04T09:19:23.2413923Z Current runner version: '2.330.0' 2025-12-04T09:19:23.2419761Z Runner name: 'i-0a8d72b6d567bdb22' 2025-12-04T09:19:23.2420559Z Runner group name: 'default' 2025-12-04T09:19:23.2421349Z Machine name: 'ip-10-0-27-140' 2025-12-04T09:19:23.2423997Z ##[group]GITHUB_TOKEN Permissions 2025-12-04T09:19:23.2426234Z Contents: read 2025-12-04T09:19:23.2426735Z Metadata: read 2025-12-04T09:19:23.2427238Z ##[endgroup] 2025-12-04T09:19:23.2429064Z Secret source: Actions 2025-12-04T09:19:23.2429841Z Prepare workflow directory 2025-12-04T09:19:23.2892092Z Prepare all required actions 2025-12-04T09:19:23.2925239Z Getting action download info 2025-12-04T09:19:23.6026556Z Download action repository 'pytorch/test-infra@main' (SHA:39aa74d619174326f4e2fb0e216151c2f29d9ffd) 2025-12-04T09:19:26.4463987Z Download action repository 'pytorch/pytorch@main' (SHA:7716da9fb23f27a65b41f9f016a2afadf281c18f) 2025-12-04T09:19:43.4764858Z Download action repository 'actions/setup-python@a26af69be951a213d495a4c3e4e4022e16d87065' (SHA:a26af69be951a213d495a4c3e4e4022e16d87065) 2025-12-04T09:19:43.8318842Z Download action repository 'aws-actions/configure-aws-credentials@ececac1a45f3b08a01d2dd070d28d111c5fe6722' (SHA:ececac1a45f3b08a01d2dd070d28d111c5fe6722) 2025-12-04T09:19:44.1397683Z Download action repository 'aws-actions/amazon-ecr-login@062b18b96a7aff071d4dc91bc00c4c1a7945b076' (SHA:062b18b96a7aff071d4dc91bc00c4c1a7945b076) 2025-12-04T09:19:44.3481021Z Download action repository 'seemethere/download-artifact-s3@1da556a7aa0a088e3153970611f6c432d58e80e6' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-12-04T09:19:44.6205561Z Download action repository 'seemethere/upload-artifact-s3@baba72d0712b404f646cebe0730933554ebce96a' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-12-04T09:19:44.8998031Z Getting action download info 2025-12-04T09:19:45.0152029Z Download action repository 'actions/checkout@v4' (SHA:34e114876b0b11c390a56381ad16ebd13914f8d5) 2025-12-04T09:19:45.3872453Z Getting action download info 2025-12-04T09:19:45.5388142Z Download action repository 'nick-fields/retry@v3.0.0' (SHA:7152eba30c6575329ac0576536151aca5a72780e) 2025-12-04T09:19:45.7727496Z Getting action download info 2025-12-04T09:19:45.9194711Z Download action repository 'nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482' (SHA:3e91a01664abd3c5cd539100d10d33b9c5b68482) 2025-12-04T09:19:46.1444987Z Getting action download info 2025-12-04T09:19:46.2808666Z Uses: pytorch/pytorch/.github/workflows/_linux-test.yml@refs/heads/main (ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32) 2025-12-04T09:19:46.2812211Z ##[group] Inputs 2025-12-04T09:19:46.2812520Z build-environment: linux-jammy-cuda12.8-py3.12-gcc11 2025-12-04T09:19:46.2818493Z test-matrix: {"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_language_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu", "unstable": "unstable"}]} 2025-12-04T09:19:46.2824935Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:19:46.2825555Z sync-tag: 2025-12-04T09:19:46.2826208Z timeout-minutes: 240 2025-12-04T09:19:46.2826408Z use-gha: 2025-12-04T09:19:46.2826566Z dashboard-tag: 2025-12-04T09:19:46.2826744Z s3-bucket: gha-artifacts 2025-12-04T09:19:46.2826955Z aws-role-to-assume: 2025-12-04T09:19:46.2827426Z disable-monitor: false 2025-12-04T09:19:46.2827672Z monitor-log-interval: 5 2025-12-04T09:19:46.2827898Z monitor-data-collect-interval: 1 2025-12-04T09:19:46.2828121Z ##[endgroup] 2025-12-04T09:19:46.2828513Z Complete job name: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu, unstable) 2025-12-04T09:19:46.3432829Z A job started hook has been configured by the self-hosted runner administrator 2025-12-04T09:19:46.3525795Z ##[group]Run '/home/ec2-user/runner-scripts/before_job.sh' 2025-12-04T09:19:46.3536064Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:19:46.3536599Z ##[endgroup] 2025-12-04T09:19:47.6931690Z Runner Type: linux.g6.4xlarge.experimental.nvidia.gpu 2025-12-04T09:19:47.6932122Z Instance Type: g6.4xlarge 2025-12-04T09:19:47.6932327Z AMI Name: unknown 2025-12-04T09:19:47.6987453Z AMI ID: ami-08982f1c5bf93d976 2025-12-04T09:19:52.5226876Z ##[group]Run pytorch/test-infra/.github/actions/setup-ssh@main 2025-12-04T09:19:52.5227238Z with: 2025-12-04T09:19:52.5227753Z github-secret: *** 2025-12-04T09:19:52.5228281Z instructions: All testing is done inside the container, to start an interactive session run: docker exec -it $(docker container ps --format '{{.ID}}') bash 2025-12-04T09:19:52.5228851Z activate-with-label: false 2025-12-04T09:19:52.5229062Z label: with-ssh 2025-12-04T09:19:52.5229238Z remove-existing-keys: true 2025-12-04T09:19:52.5229438Z fail-silently: true 2025-12-04T09:19:52.5229627Z env: 2025-12-04T09:19:52.5229772Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:19:52.5229968Z ##[endgroup] 2025-12-04T09:19:52.6527332Z Please see https://github.com/pytorch/pytorch/wiki/Debugging-using-with-ssh-for-Github-Actions for more info. 2025-12-04T09:19:52.6528714Z Not on pull request and ciflow reference could not be extracted, skipping adding ssh keys 2025-12-04T09:19:52.6672251Z ##[group]Run pytorch/pytorch/.github/actions/checkout-pytorch@main 2025-12-04T09:19:52.6672620Z with: 2025-12-04T09:19:52.6672783Z no-sudo: true 2025-12-04T09:19:52.6672975Z submodules: recursive 2025-12-04T09:19:52.6673171Z fetch-depth: 0 2025-12-04T09:19:52.6673339Z env: 2025-12-04T09:19:52.6673492Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:19:52.6673682Z ##[endgroup] 2025-12-04T09:19:52.6737095Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:19:52.6737824Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:19:52.6749924Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:19:52.6750220Z env: 2025-12-04T09:19:52.6750401Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:19:52.6750630Z ##[endgroup] 2025-12-04T09:19:52.6833132Z ##[group]Run # Use all available CPUs for fetching 2025-12-04T09:19:52.6833455Z # Use all available CPUs for fetching 2025-12-04T09:19:52.6833706Z cd "${GITHUB_WORKSPACE}" 2025-12-04T09:19:52.6833946Z git config --global fetch.parallel 0 2025-12-04T09:19:52.6834226Z git config --global submodule.fetchJobs 0 2025-12-04T09:19:52.6834473Z  2025-12-04T09:19:52.6834784Z # Clean workspace. The default checkout action should also do this, but 2025-12-04T09:19:52.6835142Z # do it here as well just in case 2025-12-04T09:19:52.6835383Z if [[ -d .git ]]; then 2025-12-04T09:19:52.6835597Z  if [ -z "${NO_SUDO}" ]; then 2025-12-04T09:19:52.6835825Z  sudo git clean -ffdx 2025-12-04T09:19:52.6836027Z  else 2025-12-04T09:19:52.6836195Z  git clean -ffdx 2025-12-04T09:19:52.6836370Z  fi 2025-12-04T09:19:52.6836521Z fi 2025-12-04T09:19:52.6843641Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:19:52.6843913Z env: 2025-12-04T09:19:52.6844070Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:19:52.6844254Z NO_SUDO: true 2025-12-04T09:19:52.6844408Z ##[endgroup] 2025-12-04T09:19:52.6961053Z ##[group]Run actions/checkout@v4 2025-12-04T09:19:52.6961270Z with: 2025-12-04T09:19:52.6961459Z ref: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:19:52.6961704Z fetch-depth: 0 2025-12-04T09:19:52.6961870Z submodules: recursive 2025-12-04T09:19:52.6962056Z show-progress: false 2025-12-04T09:19:52.6962250Z repository: pytorch/pytorch 2025-12-04T09:19:52.6962597Z token: *** 2025-12-04T09:19:52.6962756Z ssh-strict: true 2025-12-04T09:19:52.6962932Z ssh-user: git 2025-12-04T09:19:52.6963104Z persist-credentials: true 2025-12-04T09:19:52.6963294Z clean: true 2025-12-04T09:19:52.6963470Z sparse-checkout-cone-mode: true 2025-12-04T09:19:52.6963686Z fetch-tags: false 2025-12-04T09:19:52.6963847Z lfs: false 2025-12-04T09:19:52.6964007Z set-safe-directory: true 2025-12-04T09:19:52.6964210Z env: 2025-12-04T09:19:52.6964361Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:19:52.6964546Z ##[endgroup] 2025-12-04T09:19:52.7960305Z Syncing repository: pytorch/pytorch 2025-12-04T09:19:52.7961438Z ##[group]Getting Git version info 2025-12-04T09:19:52.7961815Z Working directory is '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-12-04T09:19:52.7962323Z [command]/usr/bin/git version 2025-12-04T09:19:52.8173762Z git version 2.50.1 2025-12-04T09:19:52.8197771Z ##[endgroup] 2025-12-04T09:19:52.8207224Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/7a13691b-58dc-402f-8d9e-9ddca497db93/.gitconfig' 2025-12-04T09:19:52.8227031Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/7a13691b-58dc-402f-8d9e-9ddca497db93' before making global git config changes 2025-12-04T09:19:52.8227980Z Adding repository directory to the temporary git global config as a safe directory 2025-12-04T09:19:52.8248808Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:19:52.8278200Z Deleting the contents of '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-12-04T09:19:52.8281512Z ##[group]Initializing the repository 2025-12-04T09:19:52.8284990Z [command]/usr/bin/git init /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:19:52.8357733Z hint: Using 'master' as the name for the initial branch. This default branch name 2025-12-04T09:19:52.8358264Z hint: is subject to change. To configure the initial branch name to use in all 2025-12-04T09:19:52.8358772Z hint: of your new repositories, which will suppress this warning, call: 2025-12-04T09:19:52.8359425Z hint: 2025-12-04T09:19:52.8359723Z hint: git config --global init.defaultBranch 2025-12-04T09:19:52.8360054Z hint: 2025-12-04T09:19:52.8360358Z hint: Names commonly chosen instead of 'master' are 'main', 'trunk' and 2025-12-04T09:19:52.8360884Z hint: 'development'. The just-created branch can be renamed via this command: 2025-12-04T09:19:52.8361257Z hint: 2025-12-04T09:19:52.8361469Z hint: git branch -m 2025-12-04T09:19:52.8361706Z hint: 2025-12-04T09:19:52.8362032Z hint: Disable this message with "git config set advice.defaultBranchName false" 2025-12-04T09:19:52.8368656Z Initialized empty Git repository in /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/ 2025-12-04T09:19:52.8378306Z [command]/usr/bin/git remote add origin https://github.com/pytorch/pytorch 2025-12-04T09:19:52.8421526Z ##[endgroup] 2025-12-04T09:19:52.8421966Z ##[group]Disabling automatic garbage collection 2025-12-04T09:19:52.8424694Z [command]/usr/bin/git config --local gc.auto 0 2025-12-04T09:19:52.8452776Z ##[endgroup] 2025-12-04T09:19:52.8453136Z ##[group]Setting up auth 2025-12-04T09:19:52.8458705Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-12-04T09:19:52.8488354Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-12-04T09:19:52.8879350Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-12-04T09:19:52.8908924Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-12-04T09:19:52.9281319Z [command]/usr/bin/git config --local --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:19:52.9312224Z [command]/usr/bin/git submodule foreach --recursive git config --local --show-origin --name-only --get-regexp remote.origin.url 2025-12-04T09:19:52.9645501Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-12-04T09:19:52.9698335Z ##[endgroup] 2025-12-04T09:19:52.9698765Z ##[group]Fetching the repository 2025-12-04T09:19:52.9705784Z [command]/usr/bin/git -c protocol.version=2 fetch --prune --no-recurse-submodules origin +refs/heads/*:refs/remotes/origin/* +refs/tags/*:refs/tags/* 2025-12-04T09:20:35.0488910Z From https://github.com/pytorch/pytorch 2025-12-04T09:20:35.0489963Z * [new branch] 2.6.0.dev20241004+ -> origin/2.6.0.dev20241004+ 2025-12-04T09:20:35.0490521Z * [new branch] 2.9.1 -> origin/2.9.1 2025-12-04T09:20:35.0491042Z * [new branch] AaronWang04_addmmfusion_perftest -> origin/AaronWang04_addmmfusion_perftest 2025-12-04T09:20:35.0491531Z * [new branch] Flamefire-patch-1 -> origin/Flamefire-patch-1 2025-12-04T09:20:35.0493484Z * [new branch] HDCharles-2.6.0-release-notes -> origin/HDCharles-2.6.0-release-notes 2025-12-04T09:20:35.0495244Z * [new branch] HOPrintFunc -> origin/HOPrintFunc 2025-12-04T09:20:35.0498220Z * [new branch] IvanKobzarev/stack/1 -> origin/IvanKobzarev/stack/1 2025-12-04T09:20:35.0501173Z * [new branch] NicoshevSVE128 -> origin/NicoshevSVE128 2025-12-04T09:20:35.0502706Z * [new branch] PR-AOTInductorNoneBug -> origin/PR-AOTInductorNoneBug 2025-12-04T09:20:35.0504772Z * [new branch] PR-AOTInductorNoneBugFix -> origin/PR-AOTInductorNoneBugFix 2025-12-04T09:20:35.0506448Z * [new branch] PR-FixConfigsIssue -> origin/PR-FixConfigsIssue 2025-12-04T09:20:35.0508264Z * [new branch] PR-NoneBugFix-viable -> origin/PR-NoneBugFix-viable 2025-12-04T09:20:35.0510127Z * [new branch] PR-ResetToZero -> origin/PR-ResetToZero 2025-12-04T09:20:35.0512147Z * [new branch] Update-Flash-Packaging -> origin/Update-Flash-Packaging 2025-12-04T09:20:35.0513812Z * [new branch] VLA_exp -> origin/VLA_exp 2025-12-04T09:20:35.0515770Z * [new branch] activation_bench -> origin/activation_bench 2025-12-04T09:20:35.0518104Z * [new branch] addmm-heuristic -> origin/addmm-heuristic 2025-12-04T09:20:35.0520616Z * [new branch] adi/onednn_aarch64 -> origin/adi/onednn_aarch64 2025-12-04T09:20:35.0522394Z * [new branch] adi/test -> origin/adi/test 2025-12-04T09:20:35.0524201Z * [new branch] adi/test_bgemm -> origin/adi/test_bgemm 2025-12-04T09:20:35.0526036Z * [new branch] adi/test_m8g -> origin/adi/test_m8g 2025-12-04T09:20:35.0527821Z * [new branch] adi/test_onednn -> origin/adi/test_onednn 2025-12-04T09:20:35.0529608Z * [new branch] adi/test_onednn_v3.9 -> origin/adi/test_onednn_v3.9 2025-12-04T09:20:35.0531375Z * [new branch] adi/test_presve_change -> origin/adi/test_presve_change 2025-12-04T09:20:35.0533047Z * [new branch] adi/test_timm -> origin/adi/test_timm 2025-12-04T09:20:35.0535236Z * [new branch] adi/testpresve_change -> origin/adi/testpresve_change 2025-12-04T09:20:35.0538348Z * [new branch] aditew01/test/vec_bf16 -> origin/aditew01/test/vec_bf16 2025-12-04T09:20:35.0540235Z * [new branch] ah-globalfeedback-hook -> origin/ah-globalfeedback-hook 2025-12-04T09:20:35.0542361Z * [new branch] albanD-patch-1 -> origin/albanD-patch-1 2025-12-04T09:20:35.0544165Z * [new branch] also-surround-shimh -> origin/also-surround-shimh 2025-12-04T09:20:35.0546686Z * [new branch] angelayi/aot_compile -> origin/angelayi/aot_compile 2025-12-04T09:20:35.0548577Z * [new branch] angelayi/aoti_additional_files -> origin/angelayi/aoti_additional_files 2025-12-04T09:20:35.0550291Z * [new branch] angelayi/benchmark -> origin/angelayi/benchmark 2025-12-04T09:20:35.0552065Z * [new branch] angelayi/change_pytree_serialization -> origin/angelayi/change_pytree_serialization 2025-12-04T09:20:35.0553785Z * [new branch] angelayi/cpp_loader -> origin/angelayi/cpp_loader 2025-12-04T09:20:35.0555549Z * [new branch] angelayi/inductor_const -> origin/angelayi/inductor_const 2025-12-04T09:20:35.0557223Z * [new branch] angelayi/lstm -> origin/angelayi/lstm 2025-12-04T09:20:35.0559451Z * [new branch] angelayi/no_so_weight -> origin/angelayi/no_so_weight 2025-12-04T09:20:35.0561669Z * [new branch] angelayi/scan_layers -> origin/angelayi/scan_layers 2025-12-04T09:20:35.0563561Z * [new branch] angelayi/side_eff -> origin/angelayi/side_eff 2025-12-04T09:20:35.0565400Z * [new branch] angelayi/state_dict -> origin/angelayi/state_dict 2025-12-04T09:20:35.0567345Z * [new branch] angelayi/symint_input -> origin/angelayi/symint_input 2025-12-04T09:20:35.0569343Z * [new branch] angelayi/symm_mem -> origin/angelayi/symm_mem 2025-12-04T09:20:35.0571287Z * [new branch] angelayi/test_cpp -> origin/angelayi/test_cpp 2025-12-04T09:20:35.0573261Z * [new branch] angelayi/torch_size -> origin/angelayi/torch_size 2025-12-04T09:20:35.0575016Z * [new branch] annotate_assert -> origin/annotate_assert 2025-12-04T09:20:35.0576888Z * [new branch] annotate_fallback_kernel -> origin/annotate_fallback_kernel 2025-12-04T09:20:35.0578758Z * [new branch] annotation_deepcopy -> origin/annotation_deepcopy 2025-12-04T09:20:35.0580751Z * [new branch] annotation_dynamo -> origin/annotation_dynamo 2025-12-04T09:20:35.0582539Z * [new branch] aot_eager_stack_trace -> origin/aot_eager_stack_trace 2025-12-04T09:20:35.0584282Z * [new branch] aoti-cuda-alloc -> origin/aoti-cuda-alloc 2025-12-04T09:20:35.0586133Z * [new branch] aoti_const_device -> origin/aoti_const_device 2025-12-04T09:20:35.0587891Z * [new branch] aoti_fqn_name_interface -> origin/aoti_fqn_name_interface 2025-12-04T09:20:35.0589700Z * [new branch] aoti_package_weights_binary -> origin/aoti_package_weights_binary 2025-12-04T09:20:35.0591582Z * [new branch] aoti_target_windows -> origin/aoti_target_windows 2025-12-04T09:20:35.0594656Z * [new branch] arsh/feat/inductor_check_profiling -> origin/arsh/feat/inductor_check_profiling 2025-12-04T09:20:35.0596333Z * [new branch] async_tp -> origin/async_tp 2025-12-04T09:20:35.0598256Z * [new branch] atalman-inductor-perf-cu124 -> origin/atalman-inductor-perf-cu124 2025-12-04T09:20:35.0600670Z * [new branch] atalman-inductor-perf-cu124.1 -> origin/atalman-inductor-perf-cu124.1 2025-12-04T09:20:35.0602648Z * [new branch] atalman-patch-2 -> origin/atalman-patch-2 2025-12-04T09:20:35.0604509Z * [new branch] atalman-patch-3 -> origin/atalman-patch-3 2025-12-04T09:20:35.0606356Z * [new branch] atalman-patch-4 -> origin/atalman-patch-4 2025-12-04T09:20:35.0608273Z * [new branch] atalman-patch-5 -> origin/atalman-patch-5 2025-12-04T09:20:35.0610102Z * [new branch] atalman-patch-6 -> origin/atalman-patch-6 2025-12-04T09:20:35.0611846Z * [new branch] atalman-patch-7 -> origin/atalman-patch-7 2025-12-04T09:20:35.0613746Z * [new branch] atalman-patch-8 -> origin/atalman-patch-8 2025-12-04T09:20:35.0615590Z * [new branch] atalman_inductor_2.3.1 -> origin/atalman_inductor_2.3.1 2025-12-04T09:20:35.0617388Z * [new branch] atalman_inductor_2.4.0 -> origin/atalman_inductor_2.4.0 2025-12-04T09:20:35.0619332Z * [new branch] atalman_inductor_2.4.x -> origin/atalman_inductor_2.4.x 2025-12-04T09:20:35.0621261Z * [new branch] attention_benchmarking_clean -> origin/attention_benchmarking_clean 2025-12-04T09:20:35.0623680Z * [new branch] bahuang/dt_fix_scalar_add -> origin/bahuang/dt_fix_scalar_add 2025-12-04T09:20:35.0625361Z * [new branch] bahuang/fix_debug_mode -> origin/bahuang/fix_debug_mode 2025-12-04T09:20:35.0627141Z * [new branch] bahuang/fix_expand -> origin/bahuang/fix_expand 2025-12-04T09:20:35.0628857Z * [new branch] bahuang/test -> origin/bahuang/test 2025-12-04T09:20:35.0631521Z * [new branch] base/1.5 -> origin/base/1.5 2025-12-04T09:20:35.0633578Z * [new branch] batching_sdpa_efficient_attention -> origin/batching_sdpa_efficient_attention 2025-12-04T09:20:35.0635311Z * [new branch] bench_scaled_mm_ops -> origin/bench_scaled_mm_ops 2025-12-04T09:20:35.0637333Z * [new branch] benchmark-updates -> origin/benchmark-updates 2025-12-04T09:20:35.0639013Z * [new branch] benchmarking-script -> origin/benchmarking-script 2025-12-04T09:20:35.0641464Z * [new branch] bertmaher/pinbump26 -> origin/bertmaher/pinbump26 2025-12-04T09:20:35.0643878Z * [new branch] bertrand/cutlass -> origin/bertrand/cutlass 2025-12-04T09:20:35.0646378Z * [new branch] bf/bug-static-input -> origin/bf/bug-static-input 2025-12-04T09:20:35.0647999Z * [new branch] bf/cg-backend -> origin/bf/cg-backend 2025-12-04T09:20:35.0649711Z * [new branch] bf/cg-nccl-test -> origin/bf/cg-nccl-test 2025-12-04T09:20:35.0651665Z * [new branch] bf/cg-remove-check -> origin/bf/cg-remove-check 2025-12-04T09:20:35.0653421Z * [new branch] bf/clean-torchbench-hf -> origin/bf/clean-torchbench-hf 2025-12-04T09:20:35.0655064Z * [new branch] bf/combo-debug-log -> origin/bf/combo-debug-log 2025-12-04T09:20:35.0656787Z * [new branch] bf/cudagraph -> origin/bf/cudagraph 2025-12-04T09:20:35.0659072Z * [new branch] bf/cudagraph-disable-input-mutation -> origin/bf/cudagraph-disable-input-mutation 2025-12-04T09:20:35.0661278Z * [new branch] bf/cudagraph-enable-input-mutation-support-benchmark -> origin/bf/cudagraph-enable-input-mutation-support-benchmark 2025-12-04T09:20:35.0662630Z * [new branch] bf/cudagraph-partition -> origin/bf/cudagraph-partition 2025-12-04T09:20:35.0664719Z * [new branch] bf/donated-buffer-bench -> origin/bf/donated-buffer-bench 2025-12-04T09:20:35.0666552Z * [new branch] bf/dynamo-partition -> origin/bf/dynamo-partition 2025-12-04T09:20:35.0668363Z * [new branch] bf/lite -> origin/bf/lite 2025-12-04T09:20:35.0670144Z * [new branch] bf/pa-non-divisible -> origin/bf/pa-non-divisible 2025-12-04T09:20:35.0672421Z * [new branch] bf/partition-cache-free-symbols -> origin/bf/partition-cache-free-symbols 2025-12-04T09:20:35.0674277Z * [new branch] bf/partition-memory-plan -> origin/bf/partition-memory-plan 2025-12-04T09:20:35.0676148Z * [new branch] bf/partition-move-cpu -> origin/bf/partition-move-cpu 2025-12-04T09:20:35.0678052Z * [new branch] bf/partition-view-fallback -> origin/bf/partition-view-fallback 2025-12-04T09:20:35.0679806Z * [new branch] bf/remove-check-55b0c39d -> origin/bf/remove-check-55b0c39d 2025-12-04T09:20:35.0681639Z * [new branch] bf/timm-nov-26-2025 -> origin/bf/timm-nov-26-2025 2025-12-04T09:20:35.0683450Z * [new branch] bf/transformer-pin-4-57-3 -> origin/bf/transformer-pin-4-57-3 2025-12-04T09:20:35.0685279Z * [new branch] bisect_perf_hf_T5_3acc6eac492 -> origin/bisect_perf_hf_T5_3acc6eac492 2025-12-04T09:20:35.0687049Z * [new branch] bisect_perf_hf_T5_3fcf66f61fb -> origin/bisect_perf_hf_T5_3fcf66f61fb 2025-12-04T09:20:35.0688846Z * [new branch] bisect_perf_hf_T5_4009d154129 -> origin/bisect_perf_hf_T5_4009d154129 2025-12-04T09:20:35.0690581Z * [new branch] bisect_perf_hf_T5_40d0740e73d -> origin/bisect_perf_hf_T5_40d0740e73d 2025-12-04T09:20:35.0692294Z * [new branch] bisect_perf_hf_T5_5268754e -> origin/bisect_perf_hf_T5_5268754e 2025-12-04T09:20:35.0694101Z * [new branch] bisect_perf_hf_T5_7d89a8d385c -> origin/bisect_perf_hf_T5_7d89a8d385c 2025-12-04T09:20:35.0695825Z * [new branch] bisect_perf_hf_T5_b7a25c1ee7c -> origin/bisect_perf_hf_T5_b7a25c1ee7c 2025-12-04T09:20:35.0697579Z * [new branch] bisect_perf_hf_T5_c25b201583f -> origin/bisect_perf_hf_T5_c25b201583f 2025-12-04T09:20:35.0699457Z * [new branch] bisect_perf_hf_T5_c93e57efac0 -> origin/bisect_perf_hf_T5_c93e57efac0 2025-12-04T09:20:35.0701665Z * [new branch] bisect_perf_hf_T5_ca9813ea149 -> origin/bisect_perf_hf_T5_ca9813ea149 2025-12-04T09:20:35.0703252Z * [new branch] bisect_perf_hf_T5_d65f194a -> origin/bisect_perf_hf_T5_d65f194a 2025-12-04T09:20:35.0704984Z * [new branch] bisect_perf_hf_T5_da94ab0b -> origin/bisect_perf_hf_T5_da94ab0b 2025-12-04T09:20:35.0706813Z * [new branch] bisect_perf_hf_T5_da94ab0b_new -> origin/bisect_perf_hf_T5_da94ab0b_new 2025-12-04T09:20:35.0708508Z * [new branch] bisect_perf_hf_T5_db4e8a1d8a8 -> origin/bisect_perf_hf_T5_db4e8a1d8a8 2025-12-04T09:20:35.0710273Z * [new branch] bisect_perf_hf_T5_e0d97e936a2 -> origin/bisect_perf_hf_T5_e0d97e936a2 2025-12-04T09:20:35.0712113Z * [new branch] bisect_perf_hf_T5_f23621ec563 -> origin/bisect_perf_hf_T5_f23621ec563 2025-12-04T09:20:35.0714911Z * [new branch] brister/fx_device_type -> origin/brister/fx_device_type 2025-12-04T09:20:35.0716671Z * [new branch] brister/test_inductor_all_fx -> origin/brister/test_inductor_all_fx 2025-12-04T09:20:35.0718484Z * [new branch] brister/tiled_reduction_no_numel_check -> origin/brister/tiled_reduction_no_numel_check 2025-12-04T09:20:35.0720255Z * [new branch] bwd-backup -> origin/bwd-backup 2025-12-04T09:20:35.0722162Z * [new branch] c57382a49 -> origin/c57382a49 2025-12-04T09:20:35.0723910Z * [new branch] ca_0431d47eaa -> origin/ca_0431d47eaa 2025-12-04T09:20:35.0725650Z * [new branch] ca_fix_0431d47eaa -> origin/ca_fix_0431d47eaa 2025-12-04T09:20:35.0728159Z * [new branch] camyllh/test_setup_hooks_push -> origin/camyllh/test_setup_hooks_push 2025-12-04T09:20:35.0730078Z * [new branch] cccclai-patch-1 -> origin/cccclai-patch-1 2025-12-04T09:20:35.0732008Z * [new branch] cherry-pick-159969-by-pytorch_bot_bot_ -> origin/cherry-pick-159969-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0733815Z * [new branch] cherry-pick-160586-by-pytorch_bot_bot_ -> origin/cherry-pick-160586-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0735718Z * [new branch] cherry-pick-162208-by-pytorch_bot_bot_ -> origin/cherry-pick-162208-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0737521Z * [new branch] cherry-pick-163169-by-pytorch_bot_bot_ -> origin/cherry-pick-163169-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0739477Z * [new branch] cherry-pick-165086-by-pytorch_bot_bot_ -> origin/cherry-pick-165086-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0741523Z * [new branch] cherry-pick-165514-by-pytorch_bot_bot_ -> origin/cherry-pick-165514-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0743278Z * [new branch] cherry-pick-165601-by-pytorch_bot_bot_ -> origin/cherry-pick-165601-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0745115Z * [new branch] cherry-pick-165667-by-pytorch_bot_bot_ -> origin/cherry-pick-165667-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0747014Z * [new branch] cherry-pick-165815-by-pytorch_bot_bot_ -> origin/cherry-pick-165815-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0748844Z * [new branch] cherry-pick-165922-by-pytorch_bot_bot_ -> origin/cherry-pick-165922-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0750727Z * [new branch] cherry-pick-166148-by-pytorch_bot_bot_ -> origin/cherry-pick-166148-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0752580Z * [new branch] cherry-pick-166181-by-pytorch_bot_bot_ -> origin/cherry-pick-166181-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0754396Z * [new branch] cherry-pick-166404-by-pytorch_bot_bot_ -> origin/cherry-pick-166404-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0756285Z * [new branch] cherry-pick-166427-by-pytorch_bot_bot_ -> origin/cherry-pick-166427-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0758212Z * [new branch] cherry-pick-166480-by-pytorch_bot_bot_ -> origin/cherry-pick-166480-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0760001Z * [new branch] cherry-pick-166570-by-pytorch_bot_bot_ -> origin/cherry-pick-166570-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0761837Z * [new branch] cherry-pick-166993-by-pytorch_bot_bot_ -> origin/cherry-pick-166993-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0763659Z * [new branch] cherry-pick-167111-by-pytorch_bot_bot_ -> origin/cherry-pick-167111-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0765566Z * [new branch] cherry-pick-167478-by-pytorch_bot_bot_ -> origin/cherry-pick-167478-by-pytorch_bot_bot_ 2025-12-04T09:20:35.0767322Z * [new branch] cherry_pick_166036_166040 -> origin/cherry_pick_166036_166040 2025-12-04T09:20:35.0769197Z * [new branch] cherry_pick_166457 -> origin/cherry_pick_166457 2025-12-04T09:20:35.0771241Z * [new branch] cherrypick_166338 -> origin/cherrypick_166338 2025-12-04T09:20:35.0773118Z * [new branch] cherrypick_166458 -> origin/cherrypick_166458 2025-12-04T09:20:35.0774845Z * [new branch] cherrypick_166586 -> origin/cherrypick_166586 2025-12-04T09:20:35.0776603Z * [new branch] cherrypick_166956 -> origin/cherrypick_166956 2025-12-04T09:20:35.0778474Z * [new branch] ci_attn -> origin/ci_attn 2025-12-04T09:20:35.0780467Z * [new branch] codex-testing -> origin/codex-testing 2025-12-04T09:20:35.0783195Z * [new branch] codex/add-check_memory_overlap-helper-functions -> origin/codex/add-check_memory_overlap-helper-functions 2025-12-04T09:20:35.0784800Z * [new branch] codex/fix-issue-121219-in-pytorch -> origin/codex/fix-issue-121219-in-pytorch 2025-12-04T09:20:35.0787044Z * [new branch] codex/investigate-segfaults-in-get_tensor_storage_id -> origin/codex/investigate-segfaults-in-get_tensor_storage_id 2025-12-04T09:20:35.0789089Z * [new branch] codex/refactor-lintrunner-config-to-use-uv-run -> origin/codex/refactor-lintrunner-config-to-use-uv-run 2025-12-04T09:20:35.0790706Z * [new branch] compatiblpy39util -> origin/compatiblpy39util 2025-12-04T09:20:35.0792439Z * [new branch] cond_hop_device -> origin/cond_hop_device 2025-12-04T09:20:35.0794210Z * [new branch] context_test -> origin/context_test 2025-12-04T09:20:35.0796836Z * [new branch] copilot/code-style-cleanup-python-pip -> origin/copilot/code-style-cleanup-python-pip 2025-12-04T09:20:35.0799133Z * [new branch] cpio/fix_new_ami_tests -> origin/cpio/fix_new_ami_tests 2025-12-04T09:20:35.0800999Z * [new branch] cpp-docs-dependency-upgrade -> origin/cpp-docs-dependency-upgrade 2025-12-04T09:20:35.0803482Z * [new branch] crpa/typo-in-inductor_comm_lowering -> origin/crpa/typo-in-inductor_comm_lowering 2025-12-04T09:20:35.0805853Z * [new branch] csl/always_produce_xml -> origin/csl/always_produce_xml 2025-12-04T09:20:35.0807548Z * [new branch] csl/build_test_more_procs -> origin/csl/build_test_more_procs 2025-12-04T09:20:35.0809240Z * [new branch] csl/build_test_more_procs2 -> origin/csl/build_test_more_procs2 2025-12-04T09:20:35.0810968Z * [new branch] csl/clean_up -> origin/csl/clean_up 2025-12-04T09:20:35.0812838Z * [new branch] csl/fix_retry_segfault_exit -> origin/csl/fix_retry_segfault_exit 2025-12-04T09:20:35.0814506Z * [new branch] csl/katex -> origin/csl/katex 2025-12-04T09:20:35.0816487Z * [new branch] csl/larger_runner -> origin/csl/larger_runner 2025-12-04T09:20:35.0818664Z * [new branch] csl/lint_testing -> origin/csl/lint_testing 2025-12-04T09:20:35.0820851Z * [new branch] csl/lint_thing -> origin/csl/lint_thing 2025-12-04T09:20:35.0822877Z * [new branch] csl/lintrunner_stuff -> origin/csl/lintrunner_stuff 2025-12-04T09:20:35.0824723Z * [new branch] csl/manually_gen_json -> origin/csl/manually_gen_json 2025-12-04T09:20:35.0826413Z * [new branch] csl/mps_sharding -> origin/csl/mps_sharding 2025-12-04T09:20:35.0828215Z * [new branch] csl/multistage_docker -> origin/csl/multistage_docker 2025-12-04T09:20:35.0830027Z * [new branch] csl/print_timing -> origin/csl/print_timing 2025-12-04T09:20:35.0831786Z * [new branch] csl/remove_experiment -> origin/csl/remove_experiment 2025-12-04T09:20:35.0833689Z * [new branch] csl/remove_maybe_unused_var -> origin/csl/remove_maybe_unused_var 2025-12-04T09:20:35.0835588Z * [new branch] csl/remove_repo_specific_autolabel -> origin/csl/remove_repo_specific_autolabel 2025-12-04T09:20:35.0837383Z * [new branch] csl/remove_run_parallel -> origin/csl/remove_run_parallel 2025-12-04T09:20:35.0839051Z * [new branch] csl/remove_unused_vars -> origin/csl/remove_unused_vars 2025-12-04T09:20:35.0840863Z * [new branch] csl/revert_open -> origin/csl/revert_open 2025-12-04T09:20:35.0842627Z * [new branch] csl/skip_build -> origin/csl/skip_build 2025-12-04T09:20:35.0844415Z * [new branch] csl/smaller_avx_amx_runenrs -> origin/csl/smaller_avx_amx_runenrs 2025-12-04T09:20:35.0846136Z * [new branch] csl/td_job_level -> origin/csl/td_job_level 2025-12-04T09:20:35.0847983Z * [new branch] csl/test_cuda_build_large_runner -> origin/csl/test_cuda_build_large_runner 2025-12-04T09:20:35.0850111Z * [new branch] csl/test_owners_autograd_dispatch_nn -> origin/csl/test_owners_autograd_dispatch_nn 2025-12-04T09:20:35.0851716Z * [new branch] csl/test_owners_higher_confidence -> origin/csl/test_owners_higher_confidence 2025-12-04T09:20:35.0853425Z * [new branch] csl/upload_json_running -> origin/csl/upload_json_running 2025-12-04T09:20:35.0855280Z * [new branch] csl/win_sccache -> origin/csl/win_sccache 2025-12-04T09:20:35.0857027Z * [new branch] csl/xml_stuff -> origin/csl/xml_stuff 2025-12-04T09:20:35.0858836Z * [new branch] cublasrelax2 -> origin/cublasrelax2 2025-12-04T09:20:35.0860800Z * [new branch] cuda_mempool -> origin/cuda_mempool 2025-12-04T09:20:35.0862560Z * [new branch] custom_lowering_dict -> origin/custom_lowering_dict 2025-12-04T09:20:35.0865507Z * [new branch] d4l3k/debug_plane_frtrace -> origin/d4l3k/debug_plane_frtrace 2025-12-04T09:20:35.0867896Z * [new branch] daxia6/2.8o3 -> origin/daxia6/2.8o3 2025-12-04T09:20:35.0869719Z * [new branch] debug-guard -> origin/debug-guard 2025-12-04T09:20:35.0871817Z * [new branch] delete-quant-docs -> origin/delete-quant-docs 2025-12-04T09:20:35.0877275Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 2025-12-04T09:20:35.0879213Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.1 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.1 2025-12-04T09:20:35.0881524Z * [new branch] desertfire/test_cpp_wrapper -> origin/desertfire/test_cpp_wrapper 2025-12-04T09:20:35.0883148Z * [new branch] desertfire/triton-cpu-for-aarch64 -> origin/desertfire/triton-cpu-for-aarch64 2025-12-04T09:20:35.0886072Z * [new branch] dev/dhruva/flex_attn_opt -> origin/dev/dhruva/flex_attn_opt 2025-12-04T09:20:35.0888725Z * [new branch] dev/joona/MPSNDArrayAdd -> origin/dev/joona/MPSNDArrayAdd 2025-12-04T09:20:35.0890817Z * [new branch] dev/joona/Unranked -> origin/dev/joona/Unranked 2025-12-04T09:20:35.0892761Z * [new branch] dev/joona/cat -> origin/dev/joona/cat 2025-12-04T09:20:35.0894426Z * [new branch] dev/joona/embeddingbag -> origin/dev/joona/embeddingbag 2025-12-04T09:20:35.0896215Z * [new branch] dev/joona/fix_sdpa_memtest -> origin/dev/joona/fix_sdpa_memtest 2025-12-04T09:20:35.0898193Z * [new branch] dev/joona/getTensorsString -> origin/dev/joona/getTensorsString 2025-12-04T09:20:35.0900280Z * [new branch] dev/joona/mps_linear_macos14 -> origin/dev/joona/mps_linear_macos14 2025-12-04T09:20:35.0902527Z * [new branch] dev/joona/scalar_clamp -> origin/dev/joona/scalar_clamp 2025-12-04T09:20:35.0904716Z * [new branch] dev/joona/sdpa -> origin/dev/joona/sdpa 2025-12-04T09:20:35.0907136Z * [new branch] dev/joona/sdpa_api -> origin/dev/joona/sdpa_api 2025-12-04T09:20:35.0909014Z * [new branch] dev/joona/type_inf -> origin/dev/joona/type_inf 2025-12-04T09:20:35.0911086Z * [new branch] dev/joona/ulpAssertClose -> origin/dev/joona/ulpAssertClose 2025-12-04T09:20:35.0912926Z * [new branch] dev/joona/upsize3d -> origin/dev/joona/upsize3d 2025-12-04T09:20:35.0914676Z * [new branch] disp_counter -> origin/disp_counter 2025-12-04T09:20:35.0916516Z * [new branch] divyanshk-patch-1 -> origin/divyanshk-patch-1 2025-12-04T09:20:35.0918380Z * [new branch] docs -> origin/docs 2025-12-04T09:20:35.0920259Z * [new branch] documentation -> origin/documentation 2025-12-04T09:20:35.0922069Z * [new branch] eager_model_benchmarks -> origin/eager_model_benchmarks 2025-12-04T09:20:35.0924573Z * [new branch] embg/test_inductor_ci_control -> origin/embg/test_inductor_ci_control 2025-12-04T09:20:35.0926279Z * [new branch] embg/triton_l2_prefetch_128B -> origin/embg/triton_l2_prefetch_128B 2025-12-04T09:20:35.0927925Z * [new branch] embg/triton_l2_prefetch_256B -> origin/embg/triton_l2_prefetch_256B 2025-12-04T09:20:35.0929798Z * [new branch] eqy-patch-1 -> origin/eqy-patch-1 2025-12-04T09:20:35.0931644Z * [new branch] eqy-patch-2 -> origin/eqy-patch-2 2025-12-04T09:20:35.0933475Z * [new branch] eqy-patch-3 -> origin/eqy-patch-3 2025-12-04T09:20:35.0935309Z * [new branch] eqy-patch-4 -> origin/eqy-patch-4 2025-12-04T09:20:35.0937113Z * [new branch] eqy-patch-5 -> origin/eqy-patch-5 2025-12-04T09:20:35.0938884Z * [new branch] eqy-patch-6 -> origin/eqy-patch-6 2025-12-04T09:20:35.0941487Z * [new branch] exclamaforte/amd-ma -> origin/exclamaforte/amd-ma 2025-12-04T09:20:35.0943219Z * [new branch] exclamaforte/combo-kernels-perf-run -> origin/exclamaforte/combo-kernels-perf-run 2025-12-04T09:20:35.0944835Z * [new branch] exclamaforte/do_bench_refactor -> origin/exclamaforte/do_bench_refactor 2025-12-04T09:20:35.0946628Z * [new branch] exclamaforte/enable-mem-dep-fusion -> origin/exclamaforte/enable-mem-dep-fusion 2025-12-04T09:20:35.0948486Z * [new branch] exclamaforte/fix-exhaustive-autotuning -> origin/exclamaforte/fix-exhaustive-autotuning 2025-12-04T09:20:35.0950546Z * [new branch] exclamaforte/fix-trace-parsing-fx-svg -> origin/exclamaforte/fix-trace-parsing-fx-svg 2025-12-04T09:20:35.0952767Z * [new branch] exclamaforte/force-pointwise-cat-perf-run -> origin/exclamaforte/force-pointwise-cat-perf-run 2025-12-04T09:20:35.0954444Z * [new branch] exclamaforte/fusion-data -> origin/exclamaforte/fusion-data 2025-12-04T09:20:35.0956537Z * [new branch] exclamaforte/gemm-benchmark-run -> origin/exclamaforte/gemm-benchmark-run 2025-12-04T09:20:35.0958186Z * [new branch] exclamaforte/gemm-export-model -> origin/exclamaforte/gemm-export-model 2025-12-04T09:20:35.0959953Z * [new branch] exclamaforte/gemm-model -> origin/exclamaforte/gemm-model 2025-12-04T09:20:35.0961979Z * [new branch] exclamaforte/gemm-model-all-data-collection -> origin/exclamaforte/gemm-model-all-data-collection 2025-12-04T09:20:35.0963618Z * [new branch] exclamaforte/gemm-to-amd -> origin/exclamaforte/gemm-to-amd 2025-12-04T09:20:35.0965406Z * [new branch] exclamaforte/just-gemm-model -> origin/exclamaforte/just-gemm-model 2025-12-04T09:20:35.0967367Z * [new branch] exclamaforte/just-gemm-model-no-refactor -> origin/exclamaforte/just-gemm-model-no-refactor 2025-12-04T09:20:35.0969166Z * [new branch] exclamaforte/profile-diff-algo -> origin/exclamaforte/profile-diff-algo 2025-12-04T09:20:35.0971091Z * [new branch] exclamaforte/profiler-visualization -> origin/exclamaforte/profiler-visualization 2025-12-04T09:20:35.0973074Z * [new branch] exclamaforte/test_cpp_wrapper_mode -> origin/exclamaforte/test_cpp_wrapper_mode 2025-12-04T09:20:35.0974919Z * [new branch] exclamaforte/update-autotune-configs -> origin/exclamaforte/update-autotune-configs 2025-12-04T09:20:35.0976838Z * [new branch] exclamaforte/update-autotune-configs-2 -> origin/exclamaforte/update-autotune-configs-2 2025-12-04T09:20:35.0978547Z * [new branch] exec -> origin/exec 2025-12-04T09:20:35.0980654Z * [new branch] experimental-mosaic -> origin/experimental-mosaic 2025-12-04T09:20:35.0982622Z * [new branch] export-D61047529 -> origin/export-D61047529 2025-12-04T09:20:35.0984391Z * [new branch] export-D71412006 -> origin/export-D71412006 2025-12-04T09:20:35.0986325Z * [new branch] export-D73042989 -> origin/export-D73042989 2025-12-04T09:20:35.0988154Z * [new branch] export-D78957093 -> origin/export-D78957093 2025-12-04T09:20:35.0989909Z * [new branch] export-D78996107 -> origin/export-D78996107 2025-12-04T09:20:35.0991760Z * [new branch] export-D80823877 -> origin/export-D80823877 2025-12-04T09:20:35.0993620Z * [new branch] export-D80958642 -> origin/export-D80958642 2025-12-04T09:20:35.0995445Z * [new branch] export-D81054193 -> origin/export-D81054193 2025-12-04T09:20:35.0997159Z * [new branch] export-D81204584 -> origin/export-D81204584 2025-12-04T09:20:35.0998894Z * [new branch] export-D81429090 -> origin/export-D81429090 2025-12-04T09:20:35.1000882Z * [new branch] export-D82250826 -> origin/export-D82250826 2025-12-04T09:20:35.1002608Z * [new branch] export-D82253817 -> origin/export-D82253817 2025-12-04T09:20:35.1004403Z * [new branch] export-D83541846 -> origin/export-D83541846 2025-12-04T09:20:35.1006198Z * [new branch] export-D83627170 -> origin/export-D83627170 2025-12-04T09:20:35.1008013Z * [new branch] export-D83766701 -> origin/export-D83766701 2025-12-04T09:20:35.1009741Z * [new branch] export-D83768878 -> origin/export-D83768878 2025-12-04T09:20:35.1011594Z * [new branch] export-D83769447 -> origin/export-D83769447 2025-12-04T09:20:35.1013412Z * [new branch] export-D84089824 -> origin/export-D84089824 2025-12-04T09:20:35.1015201Z * [new branch] export-D84213020 -> origin/export-D84213020 2025-12-04T09:20:35.1017375Z * [new branch] export-D84373821 -> origin/export-D84373821 2025-12-04T09:20:35.1019327Z * [new branch] export-D84612194 -> origin/export-D84612194 2025-12-04T09:20:35.1021074Z * [new branch] export-D84890985 -> origin/export-D84890985 2025-12-04T09:20:35.1022811Z * [new branch] export-D85122326 -> origin/export-D85122326 2025-12-04T09:20:35.1025014Z * [new branch] export-D86256198 -> origin/export-D86256198 2025-12-04T09:20:35.1026747Z * [new branch] export-D86460608 -> origin/export-D86460608 2025-12-04T09:20:35.1028646Z * [new branch] export-D86474796 -> origin/export-D86474796 2025-12-04T09:20:35.1030608Z * [new branch] export-D86712396 -> origin/export-D86712396 2025-12-04T09:20:35.1032419Z * [new branch] export-D87022129 -> origin/export-D87022129 2025-12-04T09:20:35.1034279Z * [new branch] export-D87838959 -> origin/export-D87838959 2025-12-04T09:20:35.1036182Z * [new branch] export-D88319437 -> origin/export-D88319437 2025-12-04T09:20:35.1038171Z * [new branch] exported-model-train-idempotent -> origin/exported-model-train-idempotent 2025-12-04T09:20:35.1040513Z * [new branch] ezyang-titan-october -> origin/ezyang-titan-october 2025-12-04T09:20:35.1042306Z * [new branch] ezyang-titan-october2 -> origin/ezyang-titan-october2 2025-12-04T09:20:35.1044004Z * [new branch] ezyang-war -> origin/ezyang-war 2025-12-04T09:20:35.1046523Z * [new branch] ezyang/wip-aot-descriptors -> origin/ezyang/wip-aot-descriptors 2025-12-04T09:20:35.1048255Z * [new branch] fa_u8_brgemm -> origin/fa_u8_brgemm 2025-12-04T09:20:35.1050828Z * [new branch] fadeputr/sequence_fbgemm -> origin/fadeputr/sequence_fbgemm 2025-12-04T09:20:35.1052600Z * [new branch] fastmath_baseline -> origin/fastmath_baseline 2025-12-04T09:20:35.1055126Z * [new branch] fbcode/warm -> origin/fbcode/warm 2025-12-04T09:20:35.1057020Z * [new branch] fca -> origin/fca 2025-12-04T09:20:35.1058761Z * [new branch] fca2_ca5984c -> origin/fca2_ca5984c 2025-12-04T09:20:35.1060677Z * [new branch] fca5 -> origin/fca5 2025-12-04T09:20:35.1062996Z * [new branch] feature/justknobs-cpp -> origin/feature/justknobs-cpp 2025-12-04T09:20:35.1064824Z * [new branch] feature/numa-forkserver -> origin/feature/numa-forkserver 2025-12-04T09:20:35.1066928Z * [new branch] ffast_math_baseline -> origin/ffast_math_baseline 2025-12-04T09:20:35.1068685Z * [new branch] ffast_math_target -> origin/ffast_math_target 2025-12-04T09:20:35.1071215Z * [new branch] findhao/base_commit -> origin/findhao/base_commit 2025-12-04T09:20:35.1073258Z * [new branch] findhao/base_commit1 -> origin/findhao/base_commit1 2025-12-04T09:20:35.1074939Z * [new branch] findhao/multistream2 -> origin/findhao/multistream2 2025-12-04T09:20:35.1076577Z * [new branch] findhao/multistream5 -> origin/findhao/multistream5 2025-12-04T09:20:35.1078230Z * [new branch] findhao/multistream6 -> origin/findhao/multistream6 2025-12-04T09:20:35.1079927Z * [new branch] findhao/operatorbench3 -> origin/findhao/operatorbench3 2025-12-04T09:20:35.1081630Z * [new branch] findhao/operatorbench5 -> origin/findhao/operatorbench5 2025-12-04T09:20:35.1083269Z * [new branch] findhao/tritonparse -> origin/findhao/tritonparse 2025-12-04T09:20:35.1085123Z * [new branch] fix-ck-gemm-template-format -> origin/fix-ck-gemm-template-format 2025-12-04T09:20:35.1087110Z * [new branch] fix-config-ignore -> origin/fix-config-ignore 2025-12-04T09:20:35.1088777Z * [new branch] fix-dict-guard -> origin/fix-dict-guard 2025-12-04T09:20:35.1090656Z * [new branch] fix_addmm_issue -> origin/fix_addmm_issue 2025-12-04T09:20:35.1092440Z * [new branch] fix_amd_missing_cluster_dims -> origin/fix_amd_missing_cluster_dims 2025-12-04T09:20:35.1094227Z * [new branch] fix_bench_bwd_pass -> origin/fix_bench_bwd_pass 2025-12-04T09:20:35.1095998Z * [new branch] fix_mem_profiler_config -> origin/fix_mem_profiler_config 2025-12-04T09:20:35.1097772Z * [new branch] fix_nvrtc_discovery -> origin/fix_nvrtc_discovery 2025-12-04T09:20:35.1099577Z * [new branch] fix_op_runner -> origin/fix_op_runner 2025-12-04T09:20:35.1101620Z * [new branch] fix_ubn_159469 -> origin/fix_ubn_159469 2025-12-04T09:20:35.1103422Z * [new branch] fixes-triage -> origin/fixes-triage 2025-12-04T09:20:35.1105119Z * [new branch] fixflashinfer -> origin/fixflashinfer 2025-12-04T09:20:35.1106915Z * [new branch] flash_decoding_cpu -> origin/flash_decoding_cpu 2025-12-04T09:20:35.1108724Z * [new branch] flex-flash -> origin/flex-flash 2025-12-04T09:20:35.1110497Z * [new branch] flex_attention_functorch_grad -> origin/flex_attention_functorch_grad 2025-12-04T09:20:35.1112532Z * [new branch] flex_flash -> origin/flex_flash 2025-12-04T09:20:35.1115053Z * [new branch] fmassa/fix_memeff_sharding_rule -> origin/fmassa/fix_memeff_sharding_rule 2025-12-04T09:20:35.1116838Z * [new branch] fmassa/tests_comm_compute_scheduler -> origin/fmassa/tests_comm_compute_scheduler 2025-12-04T09:20:35.1118524Z * [new branch] forkserver_fix -> origin/forkserver_fix 2025-12-04T09:20:35.1120359Z * [new branch] fsdp2_trace_rules -> origin/fsdp2_trace_rules 2025-12-04T09:20:35.1122318Z * [new branch] fx_cpp -> origin/fx_cpp 2025-12-04T09:20:35.1124719Z * [new branch] fy/fix-win -> origin/fy/fix-win 2025-12-04T09:20:35.1126646Z * [new branch] galv-patch-1 -> origin/galv-patch-1 2025-12-04T09:20:35.1129259Z * [new branch] galv/cudagraphs-conditional-nodes-4 -> origin/galv/cudagraphs-conditional-nodes-4 2025-12-04T09:20:35.1131734Z * [new branch] georgehong/cmakelists-patch -> origin/georgehong/cmakelists-patch 2025-12-04T09:20:35.1135166Z * [new branch] gh/AlnisM/1/base -> origin/gh/AlnisM/1/base 2025-12-04T09:20:35.1136877Z * [new branch] gh/AlnisM/1/head -> origin/gh/AlnisM/1/head 2025-12-04T09:20:35.1139928Z * [new branch] gh/EikanWang/67/base -> origin/gh/EikanWang/67/base 2025-12-04T09:20:35.1141819Z * [new branch] gh/EikanWang/67/head -> origin/gh/EikanWang/67/head 2025-12-04T09:20:35.1144977Z * [new branch] gh/Gasoonjia/1/base -> origin/gh/Gasoonjia/1/base 2025-12-04T09:20:35.1146733Z * [new branch] gh/Gasoonjia/1/head -> origin/gh/Gasoonjia/1/head 2025-12-04T09:20:35.1149661Z * [new branch] gh/H-Huang/131/base -> origin/gh/H-Huang/131/base 2025-12-04T09:20:35.1151377Z * [new branch] gh/H-Huang/131/head -> origin/gh/H-Huang/131/head 2025-12-04T09:20:35.1153086Z * [new branch] gh/H-Huang/131/orig -> origin/gh/H-Huang/131/orig 2025-12-04T09:20:35.1155433Z * [new branch] gh/H-Huang/132/base -> origin/gh/H-Huang/132/base 2025-12-04T09:20:35.1157215Z * [new branch] gh/H-Huang/132/head -> origin/gh/H-Huang/132/head 2025-12-04T09:20:35.1158993Z * [new branch] gh/H-Huang/132/orig -> origin/gh/H-Huang/132/orig 2025-12-04T09:20:35.1161505Z * [new branch] gh/H-Huang/180/base -> origin/gh/H-Huang/180/base 2025-12-04T09:20:35.1163160Z * [new branch] gh/H-Huang/180/head -> origin/gh/H-Huang/180/head 2025-12-04T09:20:35.1164874Z * [new branch] gh/H-Huang/180/orig -> origin/gh/H-Huang/180/orig 2025-12-04T09:20:35.1167146Z * [new branch] gh/H-Huang/182/base -> origin/gh/H-Huang/182/base 2025-12-04T09:20:35.1168942Z * [new branch] gh/H-Huang/182/head -> origin/gh/H-Huang/182/head 2025-12-04T09:20:35.1170684Z * [new branch] gh/H-Huang/182/orig -> origin/gh/H-Huang/182/orig 2025-12-04T09:20:35.1173602Z * [new branch] gh/H-Huang/226/base -> origin/gh/H-Huang/226/base 2025-12-04T09:20:35.1175242Z * [new branch] gh/H-Huang/226/head -> origin/gh/H-Huang/226/head 2025-12-04T09:20:35.1176989Z * [new branch] gh/H-Huang/226/orig -> origin/gh/H-Huang/226/orig 2025-12-04T09:20:35.1179468Z * [new branch] gh/H-Huang/228/base -> origin/gh/H-Huang/228/base 2025-12-04T09:20:35.1181271Z * [new branch] gh/H-Huang/228/head -> origin/gh/H-Huang/228/head 2025-12-04T09:20:35.1182955Z * [new branch] gh/H-Huang/228/orig -> origin/gh/H-Huang/228/orig 2025-12-04T09:20:35.1185840Z * [new branch] gh/IvanKobzarev/150/base -> origin/gh/IvanKobzarev/150/base 2025-12-04T09:20:35.1187680Z * [new branch] gh/IvanKobzarev/150/head -> origin/gh/IvanKobzarev/150/head 2025-12-04T09:20:35.1189321Z * [new branch] gh/IvanKobzarev/150/orig -> origin/gh/IvanKobzarev/150/orig 2025-12-04T09:20:35.1191781Z * [new branch] gh/IvanKobzarev/157/base -> origin/gh/IvanKobzarev/157/base 2025-12-04T09:20:35.1193614Z * [new branch] gh/IvanKobzarev/157/head -> origin/gh/IvanKobzarev/157/head 2025-12-04T09:20:35.1195375Z * [new branch] gh/IvanKobzarev/157/orig -> origin/gh/IvanKobzarev/157/orig 2025-12-04T09:20:35.1197844Z * [new branch] gh/IvanKobzarev/159/base -> origin/gh/IvanKobzarev/159/base 2025-12-04T09:20:35.1199589Z * [new branch] gh/IvanKobzarev/159/head -> origin/gh/IvanKobzarev/159/head 2025-12-04T09:20:35.1201567Z * [new branch] gh/IvanKobzarev/159/orig -> origin/gh/IvanKobzarev/159/orig 2025-12-04T09:20:35.1204006Z * [new branch] gh/IvanKobzarev/162/base -> origin/gh/IvanKobzarev/162/base 2025-12-04T09:20:35.1205899Z * [new branch] gh/IvanKobzarev/162/head -> origin/gh/IvanKobzarev/162/head 2025-12-04T09:20:35.1207652Z * [new branch] gh/IvanKobzarev/162/orig -> origin/gh/IvanKobzarev/162/orig 2025-12-04T09:20:35.1210077Z * [new branch] gh/IvanKobzarev/163/base -> origin/gh/IvanKobzarev/163/base 2025-12-04T09:20:35.1211757Z * [new branch] gh/IvanKobzarev/163/head -> origin/gh/IvanKobzarev/163/head 2025-12-04T09:20:35.1213498Z * [new branch] gh/IvanKobzarev/163/orig -> origin/gh/IvanKobzarev/163/orig 2025-12-04T09:20:35.1215931Z * [new branch] gh/IvanKobzarev/166/base -> origin/gh/IvanKobzarev/166/base 2025-12-04T09:20:35.1217662Z * [new branch] gh/IvanKobzarev/166/head -> origin/gh/IvanKobzarev/166/head 2025-12-04T09:20:35.1219472Z * [new branch] gh/IvanKobzarev/166/orig -> origin/gh/IvanKobzarev/166/orig 2025-12-04T09:20:35.1221914Z * [new branch] gh/IvanKobzarev/167/base -> origin/gh/IvanKobzarev/167/base 2025-12-04T09:20:35.1223646Z * [new branch] gh/IvanKobzarev/167/head -> origin/gh/IvanKobzarev/167/head 2025-12-04T09:20:35.1225409Z * [new branch] gh/IvanKobzarev/167/orig -> origin/gh/IvanKobzarev/167/orig 2025-12-04T09:20:35.1227792Z * [new branch] gh/IvanKobzarev/168/base -> origin/gh/IvanKobzarev/168/base 2025-12-04T09:20:35.1229667Z * [new branch] gh/IvanKobzarev/168/head -> origin/gh/IvanKobzarev/168/head 2025-12-04T09:20:35.1231305Z * [new branch] gh/IvanKobzarev/168/orig -> origin/gh/IvanKobzarev/168/orig 2025-12-04T09:20:35.1233684Z * [new branch] gh/IvanKobzarev/169/base -> origin/gh/IvanKobzarev/169/base 2025-12-04T09:20:35.1235510Z * [new branch] gh/IvanKobzarev/169/head -> origin/gh/IvanKobzarev/169/head 2025-12-04T09:20:35.1237244Z * [new branch] gh/IvanKobzarev/169/orig -> origin/gh/IvanKobzarev/169/orig 2025-12-04T09:20:35.1239552Z * [new branch] gh/IvanKobzarev/170/base -> origin/gh/IvanKobzarev/170/base 2025-12-04T09:20:35.1241294Z * [new branch] gh/IvanKobzarev/170/head -> origin/gh/IvanKobzarev/170/head 2025-12-04T09:20:35.1242973Z * [new branch] gh/IvanKobzarev/170/orig -> origin/gh/IvanKobzarev/170/orig 2025-12-04T09:20:35.1245540Z * [new branch] gh/IvanKobzarev/171/base -> origin/gh/IvanKobzarev/171/base 2025-12-04T09:20:35.1247331Z * [new branch] gh/IvanKobzarev/171/head -> origin/gh/IvanKobzarev/171/head 2025-12-04T09:20:35.1249083Z * [new branch] gh/IvanKobzarev/171/orig -> origin/gh/IvanKobzarev/171/orig 2025-12-04T09:20:35.1251464Z * [new branch] gh/IvanKobzarev/172/base -> origin/gh/IvanKobzarev/172/base 2025-12-04T09:20:35.1253341Z * [new branch] gh/IvanKobzarev/172/head -> origin/gh/IvanKobzarev/172/head 2025-12-04T09:20:35.1255044Z * [new branch] gh/IvanKobzarev/172/orig -> origin/gh/IvanKobzarev/172/orig 2025-12-04T09:20:35.1257458Z * [new branch] gh/IvanKobzarev/173/base -> origin/gh/IvanKobzarev/173/base 2025-12-04T09:20:35.1259287Z * [new branch] gh/IvanKobzarev/173/head -> origin/gh/IvanKobzarev/173/head 2025-12-04T09:20:35.1261096Z * [new branch] gh/IvanKobzarev/173/orig -> origin/gh/IvanKobzarev/173/orig 2025-12-04T09:20:35.1263509Z * [new branch] gh/IvanKobzarev/174/base -> origin/gh/IvanKobzarev/174/base 2025-12-04T09:20:35.1265357Z * [new branch] gh/IvanKobzarev/174/head -> origin/gh/IvanKobzarev/174/head 2025-12-04T09:20:35.1267110Z * [new branch] gh/IvanKobzarev/174/orig -> origin/gh/IvanKobzarev/174/orig 2025-12-04T09:20:35.1269548Z * [new branch] gh/IvanKobzarev/175/base -> origin/gh/IvanKobzarev/175/base 2025-12-04T09:20:35.1271497Z * [new branch] gh/IvanKobzarev/175/head -> origin/gh/IvanKobzarev/175/head 2025-12-04T09:20:35.1273422Z * [new branch] gh/IvanKobzarev/175/orig -> origin/gh/IvanKobzarev/175/orig 2025-12-04T09:20:35.1275935Z * [new branch] gh/IvanKobzarev/176/base -> origin/gh/IvanKobzarev/176/base 2025-12-04T09:20:35.1277732Z * [new branch] gh/IvanKobzarev/176/head -> origin/gh/IvanKobzarev/176/head 2025-12-04T09:20:35.1279437Z * [new branch] gh/IvanKobzarev/176/orig -> origin/gh/IvanKobzarev/176/orig 2025-12-04T09:20:35.1282129Z * [new branch] gh/IvanKobzarev/177/base -> origin/gh/IvanKobzarev/177/base 2025-12-04T09:20:35.1283924Z * [new branch] gh/IvanKobzarev/177/head -> origin/gh/IvanKobzarev/177/head 2025-12-04T09:20:35.1285646Z * [new branch] gh/IvanKobzarev/177/orig -> origin/gh/IvanKobzarev/177/orig 2025-12-04T09:20:35.1288186Z * [new branch] gh/IvanKobzarev/178/base -> origin/gh/IvanKobzarev/178/base 2025-12-04T09:20:35.1289997Z * [new branch] gh/IvanKobzarev/178/head -> origin/gh/IvanKobzarev/178/head 2025-12-04T09:20:35.1291848Z * [new branch] gh/IvanKobzarev/178/orig -> origin/gh/IvanKobzarev/178/orig 2025-12-04T09:20:35.1294334Z * [new branch] gh/IvanKobzarev/179/base -> origin/gh/IvanKobzarev/179/base 2025-12-04T09:20:35.1296069Z * [new branch] gh/IvanKobzarev/179/head -> origin/gh/IvanKobzarev/179/head 2025-12-04T09:20:35.1298062Z * [new branch] gh/IvanKobzarev/179/orig -> origin/gh/IvanKobzarev/179/orig 2025-12-04T09:20:35.1300585Z * [new branch] gh/IvanKobzarev/180/base -> origin/gh/IvanKobzarev/180/base 2025-12-04T09:20:35.1302324Z * [new branch] gh/IvanKobzarev/180/head -> origin/gh/IvanKobzarev/180/head 2025-12-04T09:20:35.1304090Z * [new branch] gh/IvanKobzarev/180/orig -> origin/gh/IvanKobzarev/180/orig 2025-12-04T09:20:35.1306693Z * [new branch] gh/IvanKobzarev/181/base -> origin/gh/IvanKobzarev/181/base 2025-12-04T09:20:35.1308452Z * [new branch] gh/IvanKobzarev/181/head -> origin/gh/IvanKobzarev/181/head 2025-12-04T09:20:35.1310246Z * [new branch] gh/IvanKobzarev/181/orig -> origin/gh/IvanKobzarev/181/orig 2025-12-04T09:20:35.1312823Z * [new branch] gh/IvanKobzarev/182/base -> origin/gh/IvanKobzarev/182/base 2025-12-04T09:20:35.1314549Z * [new branch] gh/IvanKobzarev/182/head -> origin/gh/IvanKobzarev/182/head 2025-12-04T09:20:35.1316267Z * [new branch] gh/IvanKobzarev/182/orig -> origin/gh/IvanKobzarev/182/orig 2025-12-04T09:20:35.1318910Z * [new branch] gh/IvanKobzarev/183/base -> origin/gh/IvanKobzarev/183/base 2025-12-04T09:20:35.1320713Z * [new branch] gh/IvanKobzarev/183/head -> origin/gh/IvanKobzarev/183/head 2025-12-04T09:20:35.1322610Z * [new branch] gh/IvanKobzarev/183/orig -> origin/gh/IvanKobzarev/183/orig 2025-12-04T09:20:35.1325070Z * [new branch] gh/IvanKobzarev/184/base -> origin/gh/IvanKobzarev/184/base 2025-12-04T09:20:35.1326887Z * [new branch] gh/IvanKobzarev/184/head -> origin/gh/IvanKobzarev/184/head 2025-12-04T09:20:35.1328717Z * [new branch] gh/IvanKobzarev/184/orig -> origin/gh/IvanKobzarev/184/orig 2025-12-04T09:20:35.1331553Z * [new branch] gh/NikhilAPatel/1/base -> origin/gh/NikhilAPatel/1/base 2025-12-04T09:20:35.1333410Z * [new branch] gh/NikhilAPatel/1/head -> origin/gh/NikhilAPatel/1/head 2025-12-04T09:20:35.1335635Z * [new branch] gh/NikhilAPatel/2/base -> origin/gh/NikhilAPatel/2/base 2025-12-04T09:20:35.1337414Z * [new branch] gh/NikhilAPatel/2/head -> origin/gh/NikhilAPatel/2/head 2025-12-04T09:20:35.1340113Z * [new branch] gh/NikhilAPatel/4/base -> origin/gh/NikhilAPatel/4/base 2025-12-04T09:20:35.1341908Z * [new branch] gh/NikhilAPatel/4/head -> origin/gh/NikhilAPatel/4/head 2025-12-04T09:20:35.1344310Z * [new branch] gh/NikhilAPatel/5/base -> origin/gh/NikhilAPatel/5/base 2025-12-04T09:20:35.1346115Z * [new branch] gh/NikhilAPatel/5/head -> origin/gh/NikhilAPatel/5/head 2025-12-04T09:20:35.1347921Z * [new branch] gh/NikhilAPatel/5/orig -> origin/gh/NikhilAPatel/5/orig 2025-12-04T09:20:35.1350749Z * [new branch] gh/PaliC/17/base -> origin/gh/PaliC/17/base 2025-12-04T09:20:35.1352544Z * [new branch] gh/PaliC/17/head -> origin/gh/PaliC/17/head 2025-12-04T09:20:35.1354256Z * [new branch] gh/PaliC/17/orig -> origin/gh/PaliC/17/orig 2025-12-04T09:20:35.1356674Z * [new branch] gh/PaliC/18/base -> origin/gh/PaliC/18/base 2025-12-04T09:20:35.1358382Z * [new branch] gh/PaliC/18/head -> origin/gh/PaliC/18/head 2025-12-04T09:20:35.1360222Z * [new branch] gh/PaliC/18/orig -> origin/gh/PaliC/18/orig 2025-12-04T09:20:35.1362694Z * [new branch] gh/PaliC/20/base -> origin/gh/PaliC/20/base 2025-12-04T09:20:35.1364479Z * [new branch] gh/PaliC/20/head -> origin/gh/PaliC/20/head 2025-12-04T09:20:35.1366220Z * [new branch] gh/PaliC/20/orig -> origin/gh/PaliC/20/orig 2025-12-04T09:20:35.1368545Z * [new branch] gh/PaliC/21/base -> origin/gh/PaliC/21/base 2025-12-04T09:20:35.1370522Z * [new branch] gh/PaliC/21/head -> origin/gh/PaliC/21/head 2025-12-04T09:20:35.1372449Z * [new branch] gh/PaliC/21/orig -> origin/gh/PaliC/21/orig 2025-12-04T09:20:35.1374743Z * [new branch] gh/PaliC/23/base -> origin/gh/PaliC/23/base 2025-12-04T09:20:35.1376450Z * [new branch] gh/PaliC/23/head -> origin/gh/PaliC/23/head 2025-12-04T09:20:35.1378218Z * [new branch] gh/PaliC/23/orig -> origin/gh/PaliC/23/orig 2025-12-04T09:20:35.1380690Z * [new branch] gh/PaliC/24/base -> origin/gh/PaliC/24/base 2025-12-04T09:20:35.1382439Z * [new branch] gh/PaliC/24/head -> origin/gh/PaliC/24/head 2025-12-04T09:20:35.1384151Z * [new branch] gh/PaliC/24/orig -> origin/gh/PaliC/24/orig 2025-12-04T09:20:35.1386449Z * [new branch] gh/PaliC/25/head -> origin/gh/PaliC/25/head 2025-12-04T09:20:35.1388244Z * [new branch] gh/PaliC/25/next -> origin/gh/PaliC/25/next 2025-12-04T09:20:35.1390000Z * [new branch] gh/PaliC/25/orig -> origin/gh/PaliC/25/orig 2025-12-04T09:20:35.1392330Z * [new branch] gh/PaliC/26/head -> origin/gh/PaliC/26/head 2025-12-04T09:20:35.1394025Z * [new branch] gh/PaliC/26/next -> origin/gh/PaliC/26/next 2025-12-04T09:20:35.1395753Z * [new branch] gh/PaliC/26/orig -> origin/gh/PaliC/26/orig 2025-12-04T09:20:35.1398108Z * [new branch] gh/PaliC/27/next -> origin/gh/PaliC/27/next 2025-12-04T09:20:35.1400418Z * [new branch] gh/PaliC/28/head -> origin/gh/PaliC/28/head 2025-12-04T09:20:35.1402054Z * [new branch] gh/PaliC/28/next -> origin/gh/PaliC/28/next 2025-12-04T09:20:35.1403842Z * [new branch] gh/PaliC/28/orig -> origin/gh/PaliC/28/orig 2025-12-04T09:20:35.1406216Z * [new branch] gh/PaliC/29/head -> origin/gh/PaliC/29/head 2025-12-04T09:20:35.1407839Z * [new branch] gh/PaliC/29/next -> origin/gh/PaliC/29/next 2025-12-04T09:20:35.1409606Z * [new branch] gh/PaliC/29/orig -> origin/gh/PaliC/29/orig 2025-12-04T09:20:35.1411948Z * [new branch] gh/PaliC/30/head -> origin/gh/PaliC/30/head 2025-12-04T09:20:35.1428263Z * [new branch] gh/PaliC/30/next -> origin/gh/PaliC/30/next 2025-12-04T09:20:35.1428881Z * [new branch] gh/PaliC/30/orig -> origin/gh/PaliC/30/orig 2025-12-04T09:20:35.1429275Z * [new branch] gh/PaliC/31/head -> origin/gh/PaliC/31/head 2025-12-04T09:20:35.1429655Z * [new branch] gh/PaliC/31/next -> origin/gh/PaliC/31/next 2025-12-04T09:20:35.1430030Z * [new branch] gh/PaliC/31/orig -> origin/gh/PaliC/31/orig 2025-12-04T09:20:35.1430438Z * [new branch] gh/PaulZhang12/25/base -> origin/gh/PaulZhang12/25/base 2025-12-04T09:20:35.1430856Z * [new branch] gh/PaulZhang12/25/head -> origin/gh/PaulZhang12/25/head 2025-12-04T09:20:35.1431270Z * [new branch] gh/PaulZhang12/25/orig -> origin/gh/PaulZhang12/25/orig 2025-12-04T09:20:35.1431675Z * [new branch] gh/PaulZhang12/28/base -> origin/gh/PaulZhang12/28/base 2025-12-04T09:20:35.1432289Z * [new branch] gh/PaulZhang12/28/head -> origin/gh/PaulZhang12/28/head 2025-12-04T09:20:35.1433970Z * [new branch] gh/PaulZhang12/28/orig -> origin/gh/PaulZhang12/28/orig 2025-12-04T09:20:35.1436556Z * [new branch] gh/PaulZhang12/31/base -> origin/gh/PaulZhang12/31/base 2025-12-04T09:20:35.1439742Z * [new branch] gh/PaulZhang12/31/head -> origin/gh/PaulZhang12/31/head 2025-12-04T09:20:35.1440752Z * [new branch] gh/PaulZhang12/31/orig -> origin/gh/PaulZhang12/31/orig 2025-12-04T09:20:35.1442473Z * [new branch] gh/PaulZhang12/37/base -> origin/gh/PaulZhang12/37/base 2025-12-04T09:20:35.1444282Z * [new branch] gh/PaulZhang12/37/head -> origin/gh/PaulZhang12/37/head 2025-12-04T09:20:35.1446036Z * [new branch] gh/PaulZhang12/37/orig -> origin/gh/PaulZhang12/37/orig 2025-12-04T09:20:35.1448394Z * [new branch] gh/PaulZhang12/40/base -> origin/gh/PaulZhang12/40/base 2025-12-04T09:20:35.1450188Z * [new branch] gh/PaulZhang12/40/head -> origin/gh/PaulZhang12/40/head 2025-12-04T09:20:35.1451865Z * [new branch] gh/PaulZhang12/40/orig -> origin/gh/PaulZhang12/40/orig 2025-12-04T09:20:35.1454290Z * [new branch] gh/PaulZhang12/42/base -> origin/gh/PaulZhang12/42/base 2025-12-04T09:20:35.1456076Z * [new branch] gh/PaulZhang12/42/head -> origin/gh/PaulZhang12/42/head 2025-12-04T09:20:35.1458482Z * [new branch] gh/PaulZhang12/43/base -> origin/gh/PaulZhang12/43/base 2025-12-04T09:20:35.1460355Z * [new branch] gh/PaulZhang12/43/head -> origin/gh/PaulZhang12/43/head 2025-12-04T09:20:35.1462065Z * [new branch] gh/PaulZhang12/43/orig -> origin/gh/PaulZhang12/43/orig 2025-12-04T09:20:35.1464399Z * [new branch] gh/PaulZhang12/44/base -> origin/gh/PaulZhang12/44/base 2025-12-04T09:20:35.1466133Z * [new branch] gh/PaulZhang12/44/head -> origin/gh/PaulZhang12/44/head 2025-12-04T09:20:35.1468565Z * [new branch] gh/PaulZhang12/45/base -> origin/gh/PaulZhang12/45/base 2025-12-04T09:20:35.1470249Z * [new branch] gh/PaulZhang12/45/head -> origin/gh/PaulZhang12/45/head 2025-12-04T09:20:35.1473188Z * [new branch] gh/PaulZhang12/45/orig -> origin/gh/PaulZhang12/45/orig 2025-12-04T09:20:35.1475584Z * [new branch] gh/PaulZhang12/46/base -> origin/gh/PaulZhang12/46/base 2025-12-04T09:20:35.1477342Z * [new branch] gh/PaulZhang12/46/head -> origin/gh/PaulZhang12/46/head 2025-12-04T09:20:35.1479130Z * [new branch] gh/PaulZhang12/46/orig -> origin/gh/PaulZhang12/46/orig 2025-12-04T09:20:35.1481579Z * [new branch] gh/PaulZhang12/47/base -> origin/gh/PaulZhang12/47/base 2025-12-04T09:20:35.1483381Z * [new branch] gh/PaulZhang12/47/head -> origin/gh/PaulZhang12/47/head 2025-12-04T09:20:35.1485142Z * [new branch] gh/PaulZhang12/47/orig -> origin/gh/PaulZhang12/47/orig 2025-12-04T09:20:35.1487496Z * [new branch] gh/PaulZhang12/48/base -> origin/gh/PaulZhang12/48/base 2025-12-04T09:20:35.1489238Z * [new branch] gh/PaulZhang12/48/head -> origin/gh/PaulZhang12/48/head 2025-12-04T09:20:35.1490952Z * [new branch] gh/PaulZhang12/48/orig -> origin/gh/PaulZhang12/48/orig 2025-12-04T09:20:35.1493786Z * [new branch] gh/SamGinzburg/11/base -> origin/gh/SamGinzburg/11/base 2025-12-04T09:20:35.1495539Z * [new branch] gh/SamGinzburg/11/head -> origin/gh/SamGinzburg/11/head 2025-12-04T09:20:35.1498415Z * [new branch] gh/SherlockNoMad/1/base -> origin/gh/SherlockNoMad/1/base 2025-12-04T09:20:35.1500367Z * [new branch] gh/SherlockNoMad/1/head -> origin/gh/SherlockNoMad/1/head 2025-12-04T09:20:35.1502839Z * [new branch] gh/SherlockNoMad/10/base -> origin/gh/SherlockNoMad/10/base 2025-12-04T09:20:35.1504661Z * [new branch] gh/SherlockNoMad/10/head -> origin/gh/SherlockNoMad/10/head 2025-12-04T09:20:35.1506480Z * [new branch] gh/SherlockNoMad/10/orig -> origin/gh/SherlockNoMad/10/orig 2025-12-04T09:20:35.1508864Z * [new branch] gh/SherlockNoMad/11/base -> origin/gh/SherlockNoMad/11/base 2025-12-04T09:20:35.1510602Z * [new branch] gh/SherlockNoMad/11/head -> origin/gh/SherlockNoMad/11/head 2025-12-04T09:20:35.1512517Z * [new branch] gh/SherlockNoMad/11/orig -> origin/gh/SherlockNoMad/11/orig 2025-12-04T09:20:35.1514638Z * [new branch] gh/SherlockNoMad/12/base -> origin/gh/SherlockNoMad/12/base 2025-12-04T09:20:35.1516358Z * [new branch] gh/SherlockNoMad/12/head -> origin/gh/SherlockNoMad/12/head 2025-12-04T09:20:35.1518090Z * [new branch] gh/SherlockNoMad/12/orig -> origin/gh/SherlockNoMad/12/orig 2025-12-04T09:20:35.1520557Z * [new branch] gh/SherlockNoMad/15/base -> origin/gh/SherlockNoMad/15/base 2025-12-04T09:20:35.1522261Z * [new branch] gh/SherlockNoMad/15/head -> origin/gh/SherlockNoMad/15/head 2025-12-04T09:20:35.1524084Z * [new branch] gh/SherlockNoMad/15/orig -> origin/gh/SherlockNoMad/15/orig 2025-12-04T09:20:35.1526430Z * [new branch] gh/SherlockNoMad/17/base -> origin/gh/SherlockNoMad/17/base 2025-12-04T09:20:35.1528195Z * [new branch] gh/SherlockNoMad/17/head -> origin/gh/SherlockNoMad/17/head 2025-12-04T09:20:35.1529918Z * [new branch] gh/SherlockNoMad/17/orig -> origin/gh/SherlockNoMad/17/orig 2025-12-04T09:20:35.1532647Z * [new branch] gh/SherlockNoMad/18/base -> origin/gh/SherlockNoMad/18/base 2025-12-04T09:20:35.1534489Z * [new branch] gh/SherlockNoMad/18/head -> origin/gh/SherlockNoMad/18/head 2025-12-04T09:20:35.1536266Z * [new branch] gh/SherlockNoMad/18/orig -> origin/gh/SherlockNoMad/18/orig 2025-12-04T09:20:35.1538517Z * [new branch] gh/SherlockNoMad/19/base -> origin/gh/SherlockNoMad/19/base 2025-12-04T09:20:35.1540440Z * [new branch] gh/SherlockNoMad/19/head -> origin/gh/SherlockNoMad/19/head 2025-12-04T09:20:35.1542084Z * [new branch] gh/SherlockNoMad/19/orig -> origin/gh/SherlockNoMad/19/orig 2025-12-04T09:20:35.1544338Z * [new branch] gh/SherlockNoMad/2/base -> origin/gh/SherlockNoMad/2/base 2025-12-04T09:20:35.1546041Z * [new branch] gh/SherlockNoMad/2/head -> origin/gh/SherlockNoMad/2/head 2025-12-04T09:20:35.1548280Z * [new branch] gh/SherlockNoMad/20/base -> origin/gh/SherlockNoMad/20/base 2025-12-04T09:20:35.1550185Z * [new branch] gh/SherlockNoMad/20/head -> origin/gh/SherlockNoMad/20/head 2025-12-04T09:20:35.1551879Z * [new branch] gh/SherlockNoMad/20/orig -> origin/gh/SherlockNoMad/20/orig 2025-12-04T09:20:35.1554464Z * [new branch] gh/SherlockNoMad/21/base -> origin/gh/SherlockNoMad/21/base 2025-12-04T09:20:35.1556380Z * [new branch] gh/SherlockNoMad/21/head -> origin/gh/SherlockNoMad/21/head 2025-12-04T09:20:35.1558076Z * [new branch] gh/SherlockNoMad/21/orig -> origin/gh/SherlockNoMad/21/orig 2025-12-04T09:20:35.1560354Z * [new branch] gh/SherlockNoMad/3/base -> origin/gh/SherlockNoMad/3/base 2025-12-04T09:20:35.1562002Z * [new branch] gh/SherlockNoMad/3/head -> origin/gh/SherlockNoMad/3/head 2025-12-04T09:20:35.1564221Z * [new branch] gh/SherlockNoMad/4/base -> origin/gh/SherlockNoMad/4/base 2025-12-04T09:20:35.1565956Z * [new branch] gh/SherlockNoMad/4/head -> origin/gh/SherlockNoMad/4/head 2025-12-04T09:20:35.1568166Z * [new branch] gh/SherlockNoMad/5/base -> origin/gh/SherlockNoMad/5/base 2025-12-04T09:20:35.1569914Z * [new branch] gh/SherlockNoMad/5/head -> origin/gh/SherlockNoMad/5/head 2025-12-04T09:20:35.1573558Z * [new branch] gh/Sidharth123-cpu/24/base -> origin/gh/Sidharth123-cpu/24/base 2025-12-04T09:20:35.1575825Z * [new branch] gh/Sidharth123-cpu/25/base -> origin/gh/Sidharth123-cpu/25/base 2025-12-04T09:20:35.1578050Z * [new branch] gh/Sidharth123-cpu/26/base -> origin/gh/Sidharth123-cpu/26/base 2025-12-04T09:20:35.1580579Z * [new branch] gh/Sidharth123-cpu/27/base -> origin/gh/Sidharth123-cpu/27/base 2025-12-04T09:20:35.1583693Z * [new branch] gh/StrongerXi/1/base -> origin/gh/StrongerXi/1/base 2025-12-04T09:20:35.1585346Z * [new branch] gh/StrongerXi/1/head -> origin/gh/StrongerXi/1/head 2025-12-04T09:20:35.1587707Z * [new branch] gh/StrongerXi/71/base -> origin/gh/StrongerXi/71/base 2025-12-04T09:20:35.1589461Z * [new branch] gh/StrongerXi/71/head -> origin/gh/StrongerXi/71/head 2025-12-04T09:20:35.1591714Z * [new branch] gh/StrongerXi/72/base -> origin/gh/StrongerXi/72/base 2025-12-04T09:20:35.1593388Z * [new branch] gh/StrongerXi/72/head -> origin/gh/StrongerXi/72/head 2025-12-04T09:20:35.1595773Z * [new branch] gh/StrongerXi/73/base -> origin/gh/StrongerXi/73/base 2025-12-04T09:20:35.1597495Z * [new branch] gh/StrongerXi/73/head -> origin/gh/StrongerXi/73/head 2025-12-04T09:20:35.1599333Z * [new branch] gh/StrongerXi/73/orig -> origin/gh/StrongerXi/73/orig 2025-12-04T09:20:35.1602243Z * [new branch] gh/XilunWu/160/base -> origin/gh/XilunWu/160/base 2025-12-04T09:20:35.1603909Z * [new branch] gh/XilunWu/160/head -> origin/gh/XilunWu/160/head 2025-12-04T09:20:35.1605678Z * [new branch] gh/XilunWu/160/orig -> origin/gh/XilunWu/160/orig 2025-12-04T09:20:35.1608033Z * [new branch] gh/XilunWu/163/base -> origin/gh/XilunWu/163/base 2025-12-04T09:20:35.1609800Z * [new branch] gh/XilunWu/163/head -> origin/gh/XilunWu/163/head 2025-12-04T09:20:35.1611653Z * [new branch] gh/XilunWu/163/orig -> origin/gh/XilunWu/163/orig 2025-12-04T09:20:35.1614122Z * [new branch] gh/XilunWu/168/base -> origin/gh/XilunWu/168/base 2025-12-04T09:20:35.1615892Z * [new branch] gh/XilunWu/168/head -> origin/gh/XilunWu/168/head 2025-12-04T09:20:35.1617594Z * [new branch] gh/XilunWu/168/orig -> origin/gh/XilunWu/168/orig 2025-12-04T09:20:35.1620004Z * [new branch] gh/XilunWu/169/base -> origin/gh/XilunWu/169/base 2025-12-04T09:20:35.1621849Z * [new branch] gh/XilunWu/169/head -> origin/gh/XilunWu/169/head 2025-12-04T09:20:35.1623594Z * [new branch] gh/XilunWu/169/orig -> origin/gh/XilunWu/169/orig 2025-12-04T09:20:35.1625843Z * [new branch] gh/XilunWu/170/base -> origin/gh/XilunWu/170/base 2025-12-04T09:20:35.1627606Z * [new branch] gh/XilunWu/170/head -> origin/gh/XilunWu/170/head 2025-12-04T09:20:35.1629327Z * [new branch] gh/XilunWu/170/orig -> origin/gh/XilunWu/170/orig 2025-12-04T09:20:35.1631707Z * [new branch] gh/XilunWu/171/base -> origin/gh/XilunWu/171/base 2025-12-04T09:20:35.1633435Z * [new branch] gh/XilunWu/171/head -> origin/gh/XilunWu/171/head 2025-12-04T09:20:35.1635239Z * [new branch] gh/XilunWu/171/orig -> origin/gh/XilunWu/171/orig 2025-12-04T09:20:35.1637543Z * [new branch] gh/XilunWu/173/base -> origin/gh/XilunWu/173/base 2025-12-04T09:20:35.1639354Z * [new branch] gh/XilunWu/173/head -> origin/gh/XilunWu/173/head 2025-12-04T09:20:35.1641123Z * [new branch] gh/XilunWu/173/orig -> origin/gh/XilunWu/173/orig 2025-12-04T09:20:35.1643439Z * [new branch] gh/XilunWu/175/base -> origin/gh/XilunWu/175/base 2025-12-04T09:20:35.1645315Z * [new branch] gh/XilunWu/175/head -> origin/gh/XilunWu/175/head 2025-12-04T09:20:35.1647115Z * [new branch] gh/XilunWu/175/orig -> origin/gh/XilunWu/175/orig 2025-12-04T09:20:35.1649528Z * [new branch] gh/XilunWu/176/base -> origin/gh/XilunWu/176/base 2025-12-04T09:20:35.1651215Z * [new branch] gh/XilunWu/176/head -> origin/gh/XilunWu/176/head 2025-12-04T09:20:35.1653133Z * [new branch] gh/XilunWu/176/orig -> origin/gh/XilunWu/176/orig 2025-12-04T09:20:35.1655937Z * [new branch] gh/XuehaiPan/14/base -> origin/gh/XuehaiPan/14/base 2025-12-04T09:20:35.1657676Z * [new branch] gh/XuehaiPan/14/head -> origin/gh/XuehaiPan/14/head 2025-12-04T09:20:35.1659476Z * [new branch] gh/XuehaiPan/14/orig -> origin/gh/XuehaiPan/14/orig 2025-12-04T09:20:35.1661855Z * [new branch] gh/XuehaiPan/179/base -> origin/gh/XuehaiPan/179/base 2025-12-04T09:20:35.1663640Z * [new branch] gh/XuehaiPan/179/head -> origin/gh/XuehaiPan/179/head 2025-12-04T09:20:35.1665553Z * [new branch] gh/XuehaiPan/179/orig -> origin/gh/XuehaiPan/179/orig 2025-12-04T09:20:35.1667913Z * [new branch] gh/XuehaiPan/249/base -> origin/gh/XuehaiPan/249/base 2025-12-04T09:20:35.1669635Z * [new branch] gh/XuehaiPan/249/head -> origin/gh/XuehaiPan/249/head 2025-12-04T09:20:35.1671629Z * [new branch] gh/XuehaiPan/249/orig -> origin/gh/XuehaiPan/249/orig 2025-12-04T09:20:35.1673969Z * [new branch] gh/XuehaiPan/253/base -> origin/gh/XuehaiPan/253/base 2025-12-04T09:20:35.1675745Z * [new branch] gh/XuehaiPan/253/head -> origin/gh/XuehaiPan/253/head 2025-12-04T09:20:35.1677529Z * [new branch] gh/XuehaiPan/253/orig -> origin/gh/XuehaiPan/253/orig 2025-12-04T09:20:35.1679912Z * [new branch] gh/XuehaiPan/254/base -> origin/gh/XuehaiPan/254/base 2025-12-04T09:20:35.1681772Z * [new branch] gh/XuehaiPan/254/head -> origin/gh/XuehaiPan/254/head 2025-12-04T09:20:35.1683544Z * [new branch] gh/XuehaiPan/254/orig -> origin/gh/XuehaiPan/254/orig 2025-12-04T09:20:35.1685889Z * [new branch] gh/XuehaiPan/255/base -> origin/gh/XuehaiPan/255/base 2025-12-04T09:20:35.1687655Z * [new branch] gh/XuehaiPan/255/head -> origin/gh/XuehaiPan/255/head 2025-12-04T09:20:35.1689420Z * [new branch] gh/XuehaiPan/255/orig -> origin/gh/XuehaiPan/255/orig 2025-12-04T09:20:35.1691812Z * [new branch] gh/XuehaiPan/271/base -> origin/gh/XuehaiPan/271/base 2025-12-04T09:20:35.1693518Z * [new branch] gh/XuehaiPan/271/head -> origin/gh/XuehaiPan/271/head 2025-12-04T09:20:35.1695258Z * [new branch] gh/XuehaiPan/271/orig -> origin/gh/XuehaiPan/271/orig 2025-12-04T09:20:35.1697590Z * [new branch] gh/XuehaiPan/343/base -> origin/gh/XuehaiPan/343/base 2025-12-04T09:20:35.1699415Z * [new branch] gh/XuehaiPan/343/head -> origin/gh/XuehaiPan/343/head 2025-12-04T09:20:35.1701218Z * [new branch] gh/XuehaiPan/343/orig -> origin/gh/XuehaiPan/343/orig 2025-12-04T09:20:35.1703670Z * [new branch] gh/XuehaiPan/347/base -> origin/gh/XuehaiPan/347/base 2025-12-04T09:20:35.1705455Z * [new branch] gh/XuehaiPan/347/head -> origin/gh/XuehaiPan/347/head 2025-12-04T09:20:35.1707192Z * [new branch] gh/XuehaiPan/347/orig -> origin/gh/XuehaiPan/347/orig 2025-12-04T09:20:35.1709507Z * [new branch] gh/XuehaiPan/348/base -> origin/gh/XuehaiPan/348/base 2025-12-04T09:20:35.1711249Z * [new branch] gh/XuehaiPan/348/head -> origin/gh/XuehaiPan/348/head 2025-12-04T09:20:35.1712963Z * [new branch] gh/XuehaiPan/348/orig -> origin/gh/XuehaiPan/348/orig 2025-12-04T09:20:35.1715451Z * [new branch] gh/XuehaiPan/350/base -> origin/gh/XuehaiPan/350/base 2025-12-04T09:20:35.1717191Z * [new branch] gh/XuehaiPan/350/head -> origin/gh/XuehaiPan/350/head 2025-12-04T09:20:35.1718959Z * [new branch] gh/XuehaiPan/350/orig -> origin/gh/XuehaiPan/350/orig 2025-12-04T09:20:35.1721454Z * [new branch] gh/XuehaiPan/365/base -> origin/gh/XuehaiPan/365/base 2025-12-04T09:20:35.1723147Z * [new branch] gh/XuehaiPan/365/head -> origin/gh/XuehaiPan/365/head 2025-12-04T09:20:35.1724946Z * [new branch] gh/XuehaiPan/365/orig -> origin/gh/XuehaiPan/365/orig 2025-12-04T09:20:35.1727356Z * [new branch] gh/XuehaiPan/366/base -> origin/gh/XuehaiPan/366/base 2025-12-04T09:20:35.1729119Z * [new branch] gh/XuehaiPan/366/head -> origin/gh/XuehaiPan/366/head 2025-12-04T09:20:35.1731512Z * [new branch] gh/XuehaiPan/370/base -> origin/gh/XuehaiPan/370/base 2025-12-04T09:20:35.1733223Z * [new branch] gh/XuehaiPan/370/head -> origin/gh/XuehaiPan/370/head 2025-12-04T09:20:35.1735024Z * [new branch] gh/XuehaiPan/370/orig -> origin/gh/XuehaiPan/370/orig 2025-12-04T09:20:35.1737430Z * [new branch] gh/XuehaiPan/390/base -> origin/gh/XuehaiPan/390/base 2025-12-04T09:20:35.1739658Z * [new branch] gh/XuehaiPan/390/head -> origin/gh/XuehaiPan/390/head 2025-12-04T09:20:35.1741401Z * [new branch] gh/XuehaiPan/390/orig -> origin/gh/XuehaiPan/390/orig 2025-12-04T09:20:35.1743788Z * [new branch] gh/XuehaiPan/391/base -> origin/gh/XuehaiPan/391/base 2025-12-04T09:20:35.1745517Z * [new branch] gh/XuehaiPan/391/head -> origin/gh/XuehaiPan/391/head 2025-12-04T09:20:35.1747305Z * [new branch] gh/XuehaiPan/391/orig -> origin/gh/XuehaiPan/391/orig 2025-12-04T09:20:35.1749648Z * [new branch] gh/XuehaiPan/392/base -> origin/gh/XuehaiPan/392/base 2025-12-04T09:20:35.1751418Z * [new branch] gh/XuehaiPan/392/head -> origin/gh/XuehaiPan/392/head 2025-12-04T09:20:35.1753194Z * [new branch] gh/XuehaiPan/392/orig -> origin/gh/XuehaiPan/392/orig 2025-12-04T09:20:35.1756060Z * [new branch] gh/XuehaiPan/394/base -> origin/gh/XuehaiPan/394/base 2025-12-04T09:20:35.1757838Z * [new branch] gh/XuehaiPan/394/head -> origin/gh/XuehaiPan/394/head 2025-12-04T09:20:35.1759566Z * [new branch] gh/XuehaiPan/394/orig -> origin/gh/XuehaiPan/394/orig 2025-12-04T09:20:35.1761985Z * [new branch] gh/XuehaiPan/397/base -> origin/gh/XuehaiPan/397/base 2025-12-04T09:20:35.1763754Z * [new branch] gh/XuehaiPan/397/head -> origin/gh/XuehaiPan/397/head 2025-12-04T09:20:35.1765537Z * [new branch] gh/XuehaiPan/397/orig -> origin/gh/XuehaiPan/397/orig 2025-12-04T09:20:35.1767995Z * [new branch] gh/XuehaiPan/398/base -> origin/gh/XuehaiPan/398/base 2025-12-04T09:20:35.1769708Z * [new branch] gh/XuehaiPan/398/head -> origin/gh/XuehaiPan/398/head 2025-12-04T09:20:35.1771695Z * [new branch] gh/XuehaiPan/398/orig -> origin/gh/XuehaiPan/398/orig 2025-12-04T09:20:35.1775441Z * [new branch] gh/XuehaiPan/399/base -> origin/gh/XuehaiPan/399/base 2025-12-04T09:20:35.1777199Z * [new branch] gh/XuehaiPan/399/head -> origin/gh/XuehaiPan/399/head 2025-12-04T09:20:35.1778958Z * [new branch] gh/XuehaiPan/399/orig -> origin/gh/XuehaiPan/399/orig 2025-12-04T09:20:35.1781544Z * [new branch] gh/XuehaiPan/400/base -> origin/gh/XuehaiPan/400/base 2025-12-04T09:20:35.1783269Z * [new branch] gh/XuehaiPan/400/head -> origin/gh/XuehaiPan/400/head 2025-12-04T09:20:35.1785007Z * [new branch] gh/XuehaiPan/400/orig -> origin/gh/XuehaiPan/400/orig 2025-12-04T09:20:35.1787987Z * [new branch] gh/ZhiweiYan-96/39/base -> origin/gh/ZhiweiYan-96/39/base 2025-12-04T09:20:35.1789681Z * [new branch] gh/ZhiweiYan-96/39/head -> origin/gh/ZhiweiYan-96/39/head 2025-12-04T09:20:35.1791570Z * [new branch] gh/ZhiweiYan-96/39/orig -> origin/gh/ZhiweiYan-96/39/orig 2025-12-04T09:20:35.1794083Z * [new branch] gh/ZhiweiYan-96/44/base -> origin/gh/ZhiweiYan-96/44/base 2025-12-04T09:20:35.1795707Z * [new branch] gh/ZhiweiYan-96/44/head -> origin/gh/ZhiweiYan-96/44/head 2025-12-04T09:20:35.1797974Z * [new branch] gh/ZhiweiYan-96/45/base -> origin/gh/ZhiweiYan-96/45/base 2025-12-04T09:20:35.1799731Z * [new branch] gh/ZhiweiYan-96/45/head -> origin/gh/ZhiweiYan-96/45/head 2025-12-04T09:20:35.1802184Z * [new branch] gh/ZhiweiYan-96/49/base -> origin/gh/ZhiweiYan-96/49/base 2025-12-04T09:20:35.1803932Z * [new branch] gh/ZhiweiYan-96/49/head -> origin/gh/ZhiweiYan-96/49/head 2025-12-04T09:20:35.1806255Z * [new branch] gh/ZhiweiYan-96/62/base -> origin/gh/ZhiweiYan-96/62/base 2025-12-04T09:20:35.1807983Z * [new branch] gh/ZhiweiYan-96/62/head -> origin/gh/ZhiweiYan-96/62/head 2025-12-04T09:20:35.1810437Z * [new branch] gh/ZhiweiYan-96/66/base -> origin/gh/ZhiweiYan-96/66/base 2025-12-04T09:20:35.1812343Z * [new branch] gh/ZhiweiYan-96/66/head -> origin/gh/ZhiweiYan-96/66/head 2025-12-04T09:20:35.1814640Z * [new branch] gh/ZhiweiYan-96/67/base -> origin/gh/ZhiweiYan-96/67/base 2025-12-04T09:20:35.1816393Z * [new branch] gh/ZhiweiYan-96/67/head -> origin/gh/ZhiweiYan-96/67/head 2025-12-04T09:20:35.1818604Z * [new branch] gh/ZhiweiYan-96/68/base -> origin/gh/ZhiweiYan-96/68/base 2025-12-04T09:20:35.1820425Z * [new branch] gh/ZhiweiYan-96/68/head -> origin/gh/ZhiweiYan-96/68/head 2025-12-04T09:20:35.1822084Z * [new branch] gh/ZhiweiYan-96/68/orig -> origin/gh/ZhiweiYan-96/68/orig 2025-12-04T09:20:35.1824973Z * [new branch] gh/aakhundov/1/base -> origin/gh/aakhundov/1/base 2025-12-04T09:20:35.1826829Z * [new branch] gh/aakhundov/1/head -> origin/gh/aakhundov/1/head 2025-12-04T09:20:35.1829048Z * [new branch] gh/aakhundov/2/base -> origin/gh/aakhundov/2/base 2025-12-04T09:20:35.1830717Z * [new branch] gh/aakhundov/2/head -> origin/gh/aakhundov/2/head 2025-12-04T09:20:35.1833323Z * [new branch] gh/aditew01/openblas -> origin/gh/aditew01/openblas 2025-12-04T09:20:35.1835055Z * [new branch] gh/aditew01/sbgemm -> origin/gh/aditew01/sbgemm 2025-12-04T09:20:35.1836814Z * [new branch] gh/aditew01/vecbf16 -> origin/gh/aditew01/vecbf16 2025-12-04T09:20:35.1839639Z * [new branch] gh/albanD/4/base -> origin/gh/albanD/4/base 2025-12-04T09:20:35.1841358Z * [new branch] gh/albanD/4/head -> origin/gh/albanD/4/head 2025-12-04T09:20:35.1843010Z * [new branch] gh/albanD/4/orig -> origin/gh/albanD/4/orig 2025-12-04T09:20:35.1845559Z * [new branch] gh/alexbrauckmann/paddedtensor_faketensor_init -> origin/gh/alexbrauckmann/paddedtensor_faketensor_init 2025-12-04T09:20:35.1848317Z * [new branch] gh/alexsamardzic/12/base -> origin/gh/alexsamardzic/12/base 2025-12-04T09:20:35.1850081Z * [new branch] gh/alexsamardzic/12/head -> origin/gh/alexsamardzic/12/head 2025-12-04T09:20:35.1851944Z * [new branch] gh/alexsamardzic/12/orig -> origin/gh/alexsamardzic/12/orig 2025-12-04T09:20:35.1854353Z * [new branch] gh/alexsamardzic/14/base -> origin/gh/alexsamardzic/14/base 2025-12-04T09:20:35.1856163Z * [new branch] gh/alexsamardzic/14/head -> origin/gh/alexsamardzic/14/head 2025-12-04T09:20:35.1857895Z * [new branch] gh/alexsamardzic/14/orig -> origin/gh/alexsamardzic/14/orig 2025-12-04T09:20:35.1860363Z * [new branch] gh/alexsamardzic/15/base -> origin/gh/alexsamardzic/15/base 2025-12-04T09:20:35.1862114Z * [new branch] gh/alexsamardzic/15/head -> origin/gh/alexsamardzic/15/head 2025-12-04T09:20:35.1863967Z * [new branch] gh/alexsamardzic/15/orig -> origin/gh/alexsamardzic/15/orig 2025-12-04T09:20:35.1866841Z * [new branch] gh/amjames/18/base -> origin/gh/amjames/18/base 2025-12-04T09:20:35.1868572Z * [new branch] gh/amjames/18/head -> origin/gh/amjames/18/head 2025-12-04T09:20:35.1870423Z * [new branch] gh/amjames/18/orig -> origin/gh/amjames/18/orig 2025-12-04T09:20:35.1873598Z * [new branch] gh/andrewor14/35/base -> origin/gh/andrewor14/35/base 2025-12-04T09:20:35.1875506Z * [new branch] gh/andrewor14/35/head -> origin/gh/andrewor14/35/head 2025-12-04T09:20:35.1877345Z * [new branch] gh/andrewor14/35/orig -> origin/gh/andrewor14/35/orig 2025-12-04T09:20:35.1879899Z * [new branch] gh/andrewor14/50/base -> origin/gh/andrewor14/50/base 2025-12-04T09:20:35.1881690Z * [new branch] gh/andrewor14/50/head -> origin/gh/andrewor14/50/head 2025-12-04T09:20:35.1883538Z * [new branch] gh/andrewor14/50/orig -> origin/gh/andrewor14/50/orig 2025-12-04T09:20:35.1886446Z * [new branch] gh/andyanwang/30/base -> origin/gh/andyanwang/30/base 2025-12-04T09:20:35.1888341Z * [new branch] gh/andyanwang/30/orig -> origin/gh/andyanwang/30/orig 2025-12-04T09:20:35.1890718Z * [new branch] gh/andyanwang/31/base -> origin/gh/andyanwang/31/base 2025-12-04T09:20:35.1892700Z * [new branch] gh/andyanwang/31/orig -> origin/gh/andyanwang/31/orig 2025-12-04T09:20:35.1895084Z * [new branch] gh/andyanwang/39/base -> origin/gh/andyanwang/39/base 2025-12-04T09:20:35.1896895Z * [new branch] gh/andyanwang/39/head -> origin/gh/andyanwang/39/head 2025-12-04T09:20:35.1898715Z * [new branch] gh/andyanwang/39/orig -> origin/gh/andyanwang/39/orig 2025-12-04T09:20:35.1901410Z * [new branch] gh/andyanwang/42/base -> origin/gh/andyanwang/42/base 2025-12-04T09:20:35.1903014Z * [new branch] gh/andyanwang/42/head -> origin/gh/andyanwang/42/head 2025-12-04T09:20:35.1904807Z * [new branch] gh/andyanwang/42/orig -> origin/gh/andyanwang/42/orig 2025-12-04T09:20:35.1907236Z * [new branch] gh/andyanwang/45/base -> origin/gh/andyanwang/45/base 2025-12-04T09:20:35.1909025Z * [new branch] gh/andyanwang/45/head -> origin/gh/andyanwang/45/head 2025-12-04T09:20:35.1910838Z * [new branch] gh/andyanwang/45/orig -> origin/gh/andyanwang/45/orig 2025-12-04T09:20:35.1913766Z * [new branch] gh/angelayi/107/base -> origin/gh/angelayi/107/base 2025-12-04T09:20:35.1915513Z * [new branch] gh/angelayi/107/head -> origin/gh/angelayi/107/head 2025-12-04T09:20:35.1917859Z * [new branch] gh/angelayi/114/base -> origin/gh/angelayi/114/base 2025-12-04T09:20:35.1919646Z * [new branch] gh/angelayi/114/head -> origin/gh/angelayi/114/head 2025-12-04T09:20:35.1921434Z * [new branch] gh/angelayi/114/orig -> origin/gh/angelayi/114/orig 2025-12-04T09:20:35.1923769Z * [new branch] gh/angelayi/116/base -> origin/gh/angelayi/116/base 2025-12-04T09:20:35.1925540Z * [new branch] gh/angelayi/116/head -> origin/gh/angelayi/116/head 2025-12-04T09:20:35.1927722Z * [new branch] gh/angelayi/116/orig -> origin/gh/angelayi/116/orig 2025-12-04T09:20:35.1930187Z * [new branch] gh/angelayi/122/base -> origin/gh/angelayi/122/base 2025-12-04T09:20:35.1932108Z * [new branch] gh/angelayi/122/head -> origin/gh/angelayi/122/head 2025-12-04T09:20:35.1933850Z * [new branch] gh/angelayi/122/orig -> origin/gh/angelayi/122/orig 2025-12-04T09:20:35.1936459Z * [new branch] gh/angelayi/124/base -> origin/gh/angelayi/124/base 2025-12-04T09:20:35.1938376Z * [new branch] gh/angelayi/124/head -> origin/gh/angelayi/124/head 2025-12-04T09:20:35.1939864Z * [new branch] gh/angelayi/124/orig -> origin/gh/angelayi/124/orig 2025-12-04T09:20:35.1942355Z * [new branch] gh/angelayi/128/base -> origin/gh/angelayi/128/base 2025-12-04T09:20:35.1944076Z * [new branch] gh/angelayi/128/head -> origin/gh/angelayi/128/head 2025-12-04T09:20:35.1945905Z * [new branch] gh/angelayi/128/orig -> origin/gh/angelayi/128/orig 2025-12-04T09:20:35.1948330Z * [new branch] gh/angelayi/131/base -> origin/gh/angelayi/131/base 2025-12-04T09:20:35.1950145Z * [new branch] gh/angelayi/131/head -> origin/gh/angelayi/131/head 2025-12-04T09:20:35.1951827Z * [new branch] gh/angelayi/131/orig -> origin/gh/angelayi/131/orig 2025-12-04T09:20:35.1954954Z * [new branch] gh/angelayi/132/base -> origin/gh/angelayi/132/base 2025-12-04T09:20:35.1956552Z * [new branch] gh/angelayi/132/head -> origin/gh/angelayi/132/head 2025-12-04T09:20:35.1958417Z * [new branch] gh/angelayi/132/orig -> origin/gh/angelayi/132/orig 2025-12-04T09:20:35.1960731Z * [new branch] gh/angelayi/133/base -> origin/gh/angelayi/133/base 2025-12-04T09:20:35.1962517Z * [new branch] gh/angelayi/133/head -> origin/gh/angelayi/133/head 2025-12-04T09:20:35.1964317Z * [new branch] gh/angelayi/133/orig -> origin/gh/angelayi/133/orig 2025-12-04T09:20:35.1966863Z * [new branch] gh/angelayi/134/base -> origin/gh/angelayi/134/base 2025-12-04T09:20:35.1968724Z * [new branch] gh/angelayi/134/head -> origin/gh/angelayi/134/head 2025-12-04T09:20:35.1970611Z * [new branch] gh/angelayi/134/orig -> origin/gh/angelayi/134/orig 2025-12-04T09:20:35.1973308Z * [new branch] gh/angelayi/135/base -> origin/gh/angelayi/135/base 2025-12-04T09:20:35.1975154Z * [new branch] gh/angelayi/135/head -> origin/gh/angelayi/135/head 2025-12-04T09:20:35.1976867Z * [new branch] gh/angelayi/135/orig -> origin/gh/angelayi/135/orig 2025-12-04T09:20:35.1979409Z * [new branch] gh/angelayi/136/base -> origin/gh/angelayi/136/base 2025-12-04T09:20:35.1981172Z * [new branch] gh/angelayi/136/head -> origin/gh/angelayi/136/head 2025-12-04T09:20:35.1982789Z * [new branch] gh/angelayi/136/orig -> origin/gh/angelayi/136/orig 2025-12-04T09:20:35.1985196Z * [new branch] gh/angelayi/137/base -> origin/gh/angelayi/137/base 2025-12-04T09:20:35.1986869Z * [new branch] gh/angelayi/137/head -> origin/gh/angelayi/137/head 2025-12-04T09:20:35.1988737Z * [new branch] gh/angelayi/137/orig -> origin/gh/angelayi/137/orig 2025-12-04T09:20:35.1991095Z * [new branch] gh/angelayi/138/base -> origin/gh/angelayi/138/base 2025-12-04T09:20:35.1992876Z * [new branch] gh/angelayi/138/head -> origin/gh/angelayi/138/head 2025-12-04T09:20:35.1994700Z * [new branch] gh/angelayi/138/orig -> origin/gh/angelayi/138/orig 2025-12-04T09:20:35.1997096Z * [new branch] gh/angelayi/139/base -> origin/gh/angelayi/139/base 2025-12-04T09:20:35.1998957Z * [new branch] gh/angelayi/139/head -> origin/gh/angelayi/139/head 2025-12-04T09:20:35.2000693Z * [new branch] gh/angelayi/139/orig -> origin/gh/angelayi/139/orig 2025-12-04T09:20:35.2003195Z * [new branch] gh/angelayi/140/base -> origin/gh/angelayi/140/base 2025-12-04T09:20:35.2004940Z * [new branch] gh/angelayi/140/head -> origin/gh/angelayi/140/head 2025-12-04T09:20:35.2006735Z * [new branch] gh/angelayi/140/orig -> origin/gh/angelayi/140/orig 2025-12-04T09:20:35.2009769Z * [new branch] gh/angelayi/141/base -> origin/gh/angelayi/141/base 2025-12-04T09:20:35.2011421Z * [new branch] gh/angelayi/141/head -> origin/gh/angelayi/141/head 2025-12-04T09:20:35.2013038Z * [new branch] gh/angelayi/141/orig -> origin/gh/angelayi/141/orig 2025-12-04T09:20:35.2015488Z * [new branch] gh/angelayi/142/base -> origin/gh/angelayi/142/base 2025-12-04T09:20:35.2017306Z * [new branch] gh/angelayi/142/head -> origin/gh/angelayi/142/head 2025-12-04T09:20:35.2019102Z * [new branch] gh/angelayi/142/orig -> origin/gh/angelayi/142/orig 2025-12-04T09:20:35.2021669Z * [new branch] gh/angelayi/143/base -> origin/gh/angelayi/143/base 2025-12-04T09:20:35.2023393Z * [new branch] gh/angelayi/143/head -> origin/gh/angelayi/143/head 2025-12-04T09:20:35.2025109Z * [new branch] gh/angelayi/143/orig -> origin/gh/angelayi/143/orig 2025-12-04T09:20:35.2028076Z * [new branch] gh/angelayi/144/base -> origin/gh/angelayi/144/base 2025-12-04T09:20:35.2029944Z * [new branch] gh/angelayi/144/head -> origin/gh/angelayi/144/head 2025-12-04T09:20:35.2031853Z * [new branch] gh/angelayi/144/orig -> origin/gh/angelayi/144/orig 2025-12-04T09:20:35.2034831Z * [new branch] gh/anijain2305/753/base -> origin/gh/anijain2305/753/base 2025-12-04T09:20:35.2036573Z * [new branch] gh/anijain2305/753/head -> origin/gh/anijain2305/753/head 2025-12-04T09:20:35.2038346Z * [new branch] gh/anijain2305/753/orig -> origin/gh/anijain2305/753/orig 2025-12-04T09:20:35.2040815Z * [new branch] gh/anijain2305/810/base -> origin/gh/anijain2305/810/base 2025-12-04T09:20:35.2042670Z * [new branch] gh/anijain2305/810/head -> origin/gh/anijain2305/810/head 2025-12-04T09:20:35.2044400Z * [new branch] gh/anijain2305/810/orig -> origin/gh/anijain2305/810/orig 2025-12-04T09:20:35.2046808Z * [new branch] gh/anijain2305/854/base -> origin/gh/anijain2305/854/base 2025-12-04T09:20:35.2048564Z * [new branch] gh/anijain2305/854/head -> origin/gh/anijain2305/854/head 2025-12-04T09:20:35.2050372Z * [new branch] gh/anijain2305/854/orig -> origin/gh/anijain2305/854/orig 2025-12-04T09:20:35.2052839Z * [new branch] gh/anijain2305/864/base -> origin/gh/anijain2305/864/base 2025-12-04T09:20:35.2054594Z * [new branch] gh/anijain2305/864/head -> origin/gh/anijain2305/864/head 2025-12-04T09:20:35.2056340Z * [new branch] gh/anijain2305/864/orig -> origin/gh/anijain2305/864/orig 2025-12-04T09:20:35.2058735Z * [new branch] gh/anijain2305/870/base -> origin/gh/anijain2305/870/base 2025-12-04T09:20:35.2060651Z * [new branch] gh/anijain2305/870/head -> origin/gh/anijain2305/870/head 2025-12-04T09:20:35.2062331Z * [new branch] gh/anijain2305/870/orig -> origin/gh/anijain2305/870/orig 2025-12-04T09:20:35.2064810Z * [new branch] gh/anijain2305/873/base -> origin/gh/anijain2305/873/base 2025-12-04T09:20:35.2066612Z * [new branch] gh/anijain2305/873/head -> origin/gh/anijain2305/873/head 2025-12-04T09:20:35.2068327Z * [new branch] gh/anijain2305/873/orig -> origin/gh/anijain2305/873/orig 2025-12-04T09:20:35.2070787Z * [new branch] gh/anijain2305/894/base -> origin/gh/anijain2305/894/base 2025-12-04T09:20:35.2072796Z * [new branch] gh/anijain2305/894/head -> origin/gh/anijain2305/894/head 2025-12-04T09:20:35.2074490Z * [new branch] gh/anijain2305/894/orig -> origin/gh/anijain2305/894/orig 2025-12-04T09:20:35.2076941Z * [new branch] gh/anijain2305/895/base -> origin/gh/anijain2305/895/base 2025-12-04T09:20:35.2078677Z * [new branch] gh/anijain2305/895/head -> origin/gh/anijain2305/895/head 2025-12-04T09:20:35.2080606Z * [new branch] gh/anijain2305/895/orig -> origin/gh/anijain2305/895/orig 2025-12-04T09:20:35.2082873Z * [new branch] gh/anijain2305/910/base -> origin/gh/anijain2305/910/base 2025-12-04T09:20:35.2084678Z * [new branch] gh/anijain2305/910/head -> origin/gh/anijain2305/910/head 2025-12-04T09:20:35.2086386Z * [new branch] gh/anijain2305/910/orig -> origin/gh/anijain2305/910/orig 2025-12-04T09:20:35.2088794Z * [new branch] gh/anijain2305/919/base -> origin/gh/anijain2305/919/base 2025-12-04T09:20:35.2090667Z * [new branch] gh/anijain2305/919/head -> origin/gh/anijain2305/919/head 2025-12-04T09:20:35.2092453Z * [new branch] gh/anijain2305/919/orig -> origin/gh/anijain2305/919/orig 2025-12-04T09:20:35.2094840Z * [new branch] gh/anijain2305/922/base -> origin/gh/anijain2305/922/base 2025-12-04T09:20:35.2096626Z * [new branch] gh/anijain2305/922/head -> origin/gh/anijain2305/922/head 2025-12-04T09:20:35.2098463Z * [new branch] gh/anijain2305/922/orig -> origin/gh/anijain2305/922/orig 2025-12-04T09:20:35.2100979Z * [new branch] gh/anijain2305/932/base -> origin/gh/anijain2305/932/base 2025-12-04T09:20:35.2102843Z * [new branch] gh/anijain2305/932/head -> origin/gh/anijain2305/932/head 2025-12-04T09:20:35.2104677Z * [new branch] gh/anijain2305/932/orig -> origin/gh/anijain2305/932/orig 2025-12-04T09:20:35.2107032Z * [new branch] gh/anijain2305/940/base -> origin/gh/anijain2305/940/base 2025-12-04T09:20:35.2108789Z * [new branch] gh/anijain2305/940/head -> origin/gh/anijain2305/940/head 2025-12-04T09:20:35.2110543Z * [new branch] gh/anijain2305/940/orig -> origin/gh/anijain2305/940/orig 2025-12-04T09:20:35.2113008Z * [new branch] gh/anijain2305/941/base -> origin/gh/anijain2305/941/base 2025-12-04T09:20:35.2114819Z * [new branch] gh/anijain2305/941/head -> origin/gh/anijain2305/941/head 2025-12-04T09:20:35.2116603Z * [new branch] gh/anijain2305/941/orig -> origin/gh/anijain2305/941/orig 2025-12-04T09:20:35.2118964Z * [new branch] gh/anijain2305/942/base -> origin/gh/anijain2305/942/base 2025-12-04T09:20:35.2120762Z * [new branch] gh/anijain2305/942/head -> origin/gh/anijain2305/942/head 2025-12-04T09:20:35.2122581Z * [new branch] gh/anijain2305/942/orig -> origin/gh/anijain2305/942/orig 2025-12-04T09:20:35.2125054Z * [new branch] gh/anijain2305/943/base -> origin/gh/anijain2305/943/base 2025-12-04T09:20:35.2126860Z * [new branch] gh/anijain2305/943/head -> origin/gh/anijain2305/943/head 2025-12-04T09:20:35.2128579Z * [new branch] gh/anijain2305/943/orig -> origin/gh/anijain2305/943/orig 2025-12-04T09:20:35.2131489Z * [new branch] gh/anijain2305/944/base -> origin/gh/anijain2305/944/base 2025-12-04T09:20:35.2133296Z * [new branch] gh/anijain2305/944/head -> origin/gh/anijain2305/944/head 2025-12-04T09:20:35.2135406Z * [new branch] gh/anijain2305/944/orig -> origin/gh/anijain2305/944/orig 2025-12-04T09:20:35.2137883Z * [new branch] gh/anijain2305/945/base -> origin/gh/anijain2305/945/base 2025-12-04T09:20:35.2139839Z * [new branch] gh/anijain2305/945/head -> origin/gh/anijain2305/945/head 2025-12-04T09:20:35.2141601Z * [new branch] gh/anijain2305/945/orig -> origin/gh/anijain2305/945/orig 2025-12-04T09:20:35.2143974Z * [new branch] gh/anijain2305/946/base -> origin/gh/anijain2305/946/base 2025-12-04T09:20:35.2145695Z * [new branch] gh/anijain2305/946/head -> origin/gh/anijain2305/946/head 2025-12-04T09:20:35.2147689Z * [new branch] gh/anijain2305/946/orig -> origin/gh/anijain2305/946/orig 2025-12-04T09:20:35.2149991Z * [new branch] gh/anijain2305/947/base -> origin/gh/anijain2305/947/base 2025-12-04T09:20:35.2151677Z * [new branch] gh/anijain2305/947/head -> origin/gh/anijain2305/947/head 2025-12-04T09:20:35.2153387Z * [new branch] gh/anijain2305/947/orig -> origin/gh/anijain2305/947/orig 2025-12-04T09:20:35.2155890Z * [new branch] gh/anijain2305/948/base -> origin/gh/anijain2305/948/base 2025-12-04T09:20:35.2157583Z * [new branch] gh/anijain2305/948/head -> origin/gh/anijain2305/948/head 2025-12-04T09:20:35.2159311Z * [new branch] gh/anijain2305/948/orig -> origin/gh/anijain2305/948/orig 2025-12-04T09:20:35.2161798Z * [new branch] gh/anijain2305/949/base -> origin/gh/anijain2305/949/base 2025-12-04T09:20:35.2163846Z * [new branch] gh/anijain2305/949/head -> origin/gh/anijain2305/949/head 2025-12-04T09:20:35.2165625Z * [new branch] gh/anijain2305/949/orig -> origin/gh/anijain2305/949/orig 2025-12-04T09:20:35.2168130Z * [new branch] gh/anijain2305/950/base -> origin/gh/anijain2305/950/base 2025-12-04T09:20:35.2169949Z * [new branch] gh/anijain2305/950/head -> origin/gh/anijain2305/950/head 2025-12-04T09:20:35.2173242Z * [new branch] gh/anijain2305/950/orig -> origin/gh/anijain2305/950/orig 2025-12-04T09:20:35.2175890Z * [new branch] gh/anijain2305/951/base -> origin/gh/anijain2305/951/base 2025-12-04T09:20:35.2177614Z * [new branch] gh/anijain2305/951/head -> origin/gh/anijain2305/951/head 2025-12-04T09:20:35.2179848Z * [new branch] gh/anijain2305/951/orig -> origin/gh/anijain2305/951/orig 2025-12-04T09:20:35.2182204Z * [new branch] gh/anijain2305/952/base -> origin/gh/anijain2305/952/base 2025-12-04T09:20:35.2183899Z * [new branch] gh/anijain2305/952/head -> origin/gh/anijain2305/952/head 2025-12-04T09:20:35.2185640Z * [new branch] gh/anijain2305/952/orig -> origin/gh/anijain2305/952/orig 2025-12-04T09:20:35.2188035Z * [new branch] gh/anijain2305/953/base -> origin/gh/anijain2305/953/base 2025-12-04T09:20:35.2189954Z * [new branch] gh/anijain2305/953/head -> origin/gh/anijain2305/953/head 2025-12-04T09:20:35.2191691Z * [new branch] gh/anijain2305/953/orig -> origin/gh/anijain2305/953/orig 2025-12-04T09:20:35.2194219Z * [new branch] gh/anijain2305/954/base -> origin/gh/anijain2305/954/base 2025-12-04T09:20:35.2196051Z * [new branch] gh/anijain2305/954/head -> origin/gh/anijain2305/954/head 2025-12-04T09:20:35.2197763Z * [new branch] gh/anijain2305/954/orig -> origin/gh/anijain2305/954/orig 2025-12-04T09:20:35.2200286Z * [new branch] gh/anijain2305/955/base -> origin/gh/anijain2305/955/base 2025-12-04T09:20:35.2202107Z * [new branch] gh/anijain2305/955/head -> origin/gh/anijain2305/955/head 2025-12-04T09:20:35.2203813Z * [new branch] gh/anijain2305/955/orig -> origin/gh/anijain2305/955/orig 2025-12-04T09:20:35.2206343Z * [new branch] gh/anijain2305/956/base -> origin/gh/anijain2305/956/base 2025-12-04T09:20:35.2208105Z * [new branch] gh/anijain2305/956/head -> origin/gh/anijain2305/956/head 2025-12-04T09:20:35.2209838Z * [new branch] gh/anijain2305/956/orig -> origin/gh/anijain2305/956/orig 2025-12-04T09:20:35.2212342Z * [new branch] gh/anijain2305/957/base -> origin/gh/anijain2305/957/base 2025-12-04T09:20:35.2214229Z * [new branch] gh/anijain2305/957/head -> origin/gh/anijain2305/957/head 2025-12-04T09:20:35.2215967Z * [new branch] gh/anijain2305/957/orig -> origin/gh/anijain2305/957/orig 2025-12-04T09:20:35.2218399Z * [new branch] gh/anijain2305/958/base -> origin/gh/anijain2305/958/base 2025-12-04T09:20:35.2220530Z * [new branch] gh/anijain2305/958/head -> origin/gh/anijain2305/958/head 2025-12-04T09:20:35.2222131Z * [new branch] gh/anijain2305/958/orig -> origin/gh/anijain2305/958/orig 2025-12-04T09:20:35.2224617Z * [new branch] gh/anijain2305/959/base -> origin/gh/anijain2305/959/base 2025-12-04T09:20:35.2226299Z * [new branch] gh/anijain2305/959/head -> origin/gh/anijain2305/959/head 2025-12-04T09:20:35.2228050Z * [new branch] gh/anijain2305/959/orig -> origin/gh/anijain2305/959/orig 2025-12-04T09:20:35.2230570Z * [new branch] gh/anijain2305/960/base -> origin/gh/anijain2305/960/base 2025-12-04T09:20:35.2232396Z * [new branch] gh/anijain2305/960/head -> origin/gh/anijain2305/960/head 2025-12-04T09:20:35.2234118Z * [new branch] gh/anijain2305/960/orig -> origin/gh/anijain2305/960/orig 2025-12-04T09:20:35.2236601Z * [new branch] gh/anijain2305/961/base -> origin/gh/anijain2305/961/base 2025-12-04T09:20:35.2238513Z * [new branch] gh/anijain2305/961/head -> origin/gh/anijain2305/961/head 2025-12-04T09:20:35.2240231Z * [new branch] gh/anijain2305/961/orig -> origin/gh/anijain2305/961/orig 2025-12-04T09:20:35.2242692Z * [new branch] gh/anijain2305/962/base -> origin/gh/anijain2305/962/base 2025-12-04T09:20:35.2244444Z * [new branch] gh/anijain2305/962/head -> origin/gh/anijain2305/962/head 2025-12-04T09:20:35.2246181Z * [new branch] gh/anijain2305/962/orig -> origin/gh/anijain2305/962/orig 2025-12-04T09:20:35.2248941Z * [new branch] gh/anijain2305/963/base -> origin/gh/anijain2305/963/base 2025-12-04T09:20:35.2250803Z * [new branch] gh/anijain2305/963/head -> origin/gh/anijain2305/963/head 2025-12-04T09:20:35.2252776Z * [new branch] gh/anijain2305/963/orig -> origin/gh/anijain2305/963/orig 2025-12-04T09:20:35.2255221Z * [new branch] gh/anijain2305/964/base -> origin/gh/anijain2305/964/base 2025-12-04T09:20:35.2256944Z * [new branch] gh/anijain2305/964/head -> origin/gh/anijain2305/964/head 2025-12-04T09:20:35.2258708Z * [new branch] gh/anijain2305/964/orig -> origin/gh/anijain2305/964/orig 2025-12-04T09:20:35.2261344Z * [new branch] gh/anijain2305/965/base -> origin/gh/anijain2305/965/base 2025-12-04T09:20:35.2263080Z * [new branch] gh/anijain2305/965/head -> origin/gh/anijain2305/965/head 2025-12-04T09:20:35.2264891Z * [new branch] gh/anijain2305/965/orig -> origin/gh/anijain2305/965/orig 2025-12-04T09:20:35.2267202Z * [new branch] gh/anijain2305/966/base -> origin/gh/anijain2305/966/base 2025-12-04T09:20:35.2269007Z * [new branch] gh/anijain2305/966/head -> origin/gh/anijain2305/966/head 2025-12-04T09:20:35.2270723Z * [new branch] gh/anijain2305/966/orig -> origin/gh/anijain2305/966/orig 2025-12-04T09:20:35.2273597Z * [new branch] gh/anijain2305/967/base -> origin/gh/anijain2305/967/base 2025-12-04T09:20:35.2275317Z * [new branch] gh/anijain2305/967/head -> origin/gh/anijain2305/967/head 2025-12-04T09:20:35.2277173Z * [new branch] gh/anijain2305/967/orig -> origin/gh/anijain2305/967/orig 2025-12-04T09:20:35.2279530Z * [new branch] gh/anijain2305/968/base -> origin/gh/anijain2305/968/base 2025-12-04T09:20:35.2281263Z * [new branch] gh/anijain2305/968/head -> origin/gh/anijain2305/968/head 2025-12-04T09:20:35.2283063Z * [new branch] gh/anijain2305/968/orig -> origin/gh/anijain2305/968/orig 2025-12-04T09:20:35.2285534Z * [new branch] gh/anijain2305/969/base -> origin/gh/anijain2305/969/base 2025-12-04T09:20:35.2287366Z * [new branch] gh/anijain2305/969/head -> origin/gh/anijain2305/969/head 2025-12-04T09:20:35.2289604Z * [new branch] gh/anijain2305/969/orig -> origin/gh/anijain2305/969/orig 2025-12-04T09:20:35.2291828Z * [new branch] gh/anijain2305/970/base -> origin/gh/anijain2305/970/base 2025-12-04T09:20:35.2293613Z * [new branch] gh/anijain2305/970/head -> origin/gh/anijain2305/970/head 2025-12-04T09:20:35.2295398Z * [new branch] gh/anijain2305/970/orig -> origin/gh/anijain2305/970/orig 2025-12-04T09:20:35.2298317Z * [new branch] gh/anjali411/216/base -> origin/gh/anjali411/216/base 2025-12-04T09:20:35.2300244Z * [new branch] gh/anjali411/216/head -> origin/gh/anjali411/216/head 2025-12-04T09:20:35.2301758Z * [new branch] gh/anjali411/216/orig -> origin/gh/anjali411/216/orig 2025-12-04T09:20:35.2304881Z * [new branch] gh/anshul-si/1/base -> origin/gh/anshul-si/1/base 2025-12-04T09:20:35.2306616Z * [new branch] gh/anshul-si/1/head -> origin/gh/anshul-si/1/head 2025-12-04T09:20:35.2308903Z * [new branch] gh/anshul-si/2/base -> origin/gh/anshul-si/2/base 2025-12-04T09:20:35.2310592Z * [new branch] gh/anshul-si/2/head -> origin/gh/anshul-si/2/head 2025-12-04T09:20:35.2312994Z * [new branch] gh/anshul-si/3/base -> origin/gh/anshul-si/3/base 2025-12-04T09:20:35.2314689Z * [new branch] gh/anshul-si/3/head -> origin/gh/anshul-si/3/head 2025-12-04T09:20:35.2316996Z * [new branch] gh/anshul-si/4/base -> origin/gh/anshul-si/4/base 2025-12-04T09:20:35.2318753Z * [new branch] gh/anshul-si/4/head -> origin/gh/anshul-si/4/head 2025-12-04T09:20:35.2320949Z * [new branch] gh/anshul-si/5/base -> origin/gh/anshul-si/5/base 2025-12-04T09:20:35.2322566Z * [new branch] gh/anshul-si/5/head -> origin/gh/anshul-si/5/head 2025-12-04T09:20:35.2325067Z * [new branch] gh/anshul-si/53/base -> origin/gh/anshul-si/53/base 2025-12-04T09:20:35.2326848Z * [new branch] gh/anshul-si/53/head -> origin/gh/anshul-si/53/head 2025-12-04T09:20:35.2329301Z * [new branch] gh/anshul-si/58/base -> origin/gh/anshul-si/58/base 2025-12-04T09:20:35.2331109Z * [new branch] gh/anshul-si/58/head -> origin/gh/anshul-si/58/head 2025-12-04T09:20:35.2333250Z * [new branch] gh/anshul-si/66/base -> origin/gh/anshul-si/66/base 2025-12-04T09:20:35.2335046Z * [new branch] gh/anshul-si/66/head -> origin/gh/anshul-si/66/head 2025-12-04T09:20:35.2336753Z * [new branch] gh/anshul-si/66/orig -> origin/gh/anshul-si/66/orig 2025-12-04T09:20:35.2338997Z * [new branch] gh/anshul-si/67/base -> origin/gh/anshul-si/67/base 2025-12-04T09:20:35.2340913Z * [new branch] gh/anshul-si/67/head -> origin/gh/anshul-si/67/head 2025-12-04T09:20:35.2342641Z * [new branch] gh/anshul-si/67/orig -> origin/gh/anshul-si/67/orig 2025-12-04T09:20:35.2345189Z * [new branch] gh/anshul-si/68/base -> origin/gh/anshul-si/68/base 2025-12-04T09:20:35.2346902Z * [new branch] gh/anshul-si/68/head -> origin/gh/anshul-si/68/head 2025-12-04T09:20:35.2348612Z * [new branch] gh/anshul-si/68/orig -> origin/gh/anshul-si/68/orig 2025-12-04T09:20:35.2351134Z * [new branch] gh/anshul-si/69/base -> origin/gh/anshul-si/69/base 2025-12-04T09:20:35.2352966Z * [new branch] gh/anshul-si/69/head -> origin/gh/anshul-si/69/head 2025-12-04T09:20:35.2354893Z * [new branch] gh/anshul-si/69/orig -> origin/gh/anshul-si/69/orig 2025-12-04T09:20:35.2357236Z * [new branch] gh/anshul-si/70/base -> origin/gh/anshul-si/70/base 2025-12-04T09:20:35.2359072Z * [new branch] gh/anshul-si/70/head -> origin/gh/anshul-si/70/head 2025-12-04T09:20:35.2360882Z * [new branch] gh/anshul-si/70/orig -> origin/gh/anshul-si/70/orig 2025-12-04T09:20:35.2363234Z * [new branch] gh/anshul-si/71/base -> origin/gh/anshul-si/71/base 2025-12-04T09:20:35.2364986Z * [new branch] gh/anshul-si/71/head -> origin/gh/anshul-si/71/head 2025-12-04T09:20:35.2366720Z * [new branch] gh/anshul-si/71/orig -> origin/gh/anshul-si/71/orig 2025-12-04T09:20:35.2369171Z * [new branch] gh/anshul-si/72/base -> origin/gh/anshul-si/72/base 2025-12-04T09:20:35.2370913Z * [new branch] gh/anshul-si/72/head -> origin/gh/anshul-si/72/head 2025-12-04T09:20:35.2373184Z * [new branch] gh/anshul-si/72/orig -> origin/gh/anshul-si/72/orig 2025-12-04T09:20:35.2375429Z * [new branch] gh/anshul-si/73/base -> origin/gh/anshul-si/73/base 2025-12-04T09:20:35.2377178Z * [new branch] gh/anshul-si/73/head -> origin/gh/anshul-si/73/head 2025-12-04T09:20:35.2379094Z * [new branch] gh/anshul-si/73/orig -> origin/gh/anshul-si/73/orig 2025-12-04T09:20:35.2382291Z * [new branch] gh/aorenste/132/base -> origin/gh/aorenste/132/base 2025-12-04T09:20:35.2383952Z * [new branch] gh/aorenste/132/head -> origin/gh/aorenste/132/head 2025-12-04T09:20:35.2386492Z * [new branch] gh/aorenste/134/base -> origin/gh/aorenste/134/base 2025-12-04T09:20:35.2388336Z * [new branch] gh/aorenste/134/head -> origin/gh/aorenste/134/head 2025-12-04T09:20:35.2390117Z * [new branch] gh/aorenste/134/orig -> origin/gh/aorenste/134/orig 2025-12-04T09:20:35.2392680Z * [new branch] gh/aorenste/139/base -> origin/gh/aorenste/139/base 2025-12-04T09:20:35.2394356Z * [new branch] gh/aorenste/139/head -> origin/gh/aorenste/139/head 2025-12-04T09:20:35.2396168Z * [new branch] gh/aorenste/139/orig -> origin/gh/aorenste/139/orig 2025-12-04T09:20:35.2398480Z * [new branch] gh/aorenste/141/base -> origin/gh/aorenste/141/base 2025-12-04T09:20:35.2400243Z * [new branch] gh/aorenste/141/head -> origin/gh/aorenste/141/head 2025-12-04T09:20:35.2402975Z * [new branch] gh/aorenste/145/base -> origin/gh/aorenste/145/base 2025-12-04T09:20:35.2404808Z * [new branch] gh/aorenste/145/head -> origin/gh/aorenste/145/head 2025-12-04T09:20:35.2406691Z * [new branch] gh/aorenste/145/orig -> origin/gh/aorenste/145/orig 2025-12-04T09:20:35.2409168Z * [new branch] gh/aorenste/146/base -> origin/gh/aorenste/146/base 2025-12-04T09:20:35.2411011Z * [new branch] gh/aorenste/146/head -> origin/gh/aorenste/146/head 2025-12-04T09:20:35.2412697Z * [new branch] gh/aorenste/146/orig -> origin/gh/aorenste/146/orig 2025-12-04T09:20:35.2415170Z * [new branch] gh/aorenste/147/base -> origin/gh/aorenste/147/base 2025-12-04T09:20:35.2416995Z * [new branch] gh/aorenste/147/head -> origin/gh/aorenste/147/head 2025-12-04T09:20:35.2418773Z * [new branch] gh/aorenste/147/orig -> origin/gh/aorenste/147/orig 2025-12-04T09:20:35.2421309Z * [new branch] gh/aorenste/148/base -> origin/gh/aorenste/148/base 2025-12-04T09:20:35.2423060Z * [new branch] gh/aorenste/148/head -> origin/gh/aorenste/148/head 2025-12-04T09:20:35.2424877Z * [new branch] gh/aorenste/148/orig -> origin/gh/aorenste/148/orig 2025-12-04T09:20:35.2427292Z * [new branch] gh/aorenste/149/base -> origin/gh/aorenste/149/base 2025-12-04T09:20:35.2429028Z * [new branch] gh/aorenste/149/head -> origin/gh/aorenste/149/head 2025-12-04T09:20:35.2430780Z * [new branch] gh/aorenste/149/orig -> origin/gh/aorenste/149/orig 2025-12-04T09:20:35.2433370Z * [new branch] gh/aorenste/150/base -> origin/gh/aorenste/150/base 2025-12-04T09:20:35.2435113Z * [new branch] gh/aorenste/150/head -> origin/gh/aorenste/150/head 2025-12-04T09:20:35.2436778Z * [new branch] gh/aorenste/150/orig -> origin/gh/aorenste/150/orig 2025-12-04T09:20:35.2439083Z * [new branch] gh/aorenste/151/base -> origin/gh/aorenste/151/base 2025-12-04T09:20:35.2440922Z * [new branch] gh/aorenste/151/head -> origin/gh/aorenste/151/head 2025-12-04T09:20:35.2442606Z * [new branch] gh/aorenste/151/orig -> origin/gh/aorenste/151/orig 2025-12-04T09:20:35.2445025Z * [new branch] gh/aorenste/152/base -> origin/gh/aorenste/152/base 2025-12-04T09:20:35.2446756Z * [new branch] gh/aorenste/152/head -> origin/gh/aorenste/152/head 2025-12-04T09:20:35.2448508Z * [new branch] gh/aorenste/152/orig -> origin/gh/aorenste/152/orig 2025-12-04T09:20:35.2450847Z * [new branch] gh/aorenste/153/base -> origin/gh/aorenste/153/base 2025-12-04T09:20:35.2452710Z * [new branch] gh/aorenste/153/head -> origin/gh/aorenste/153/head 2025-12-04T09:20:35.2454727Z * [new branch] gh/aorenste/153/orig -> origin/gh/aorenste/153/orig 2025-12-04T09:20:35.2456715Z * [new branch] gh/aorenste/154/base -> origin/gh/aorenste/154/base 2025-12-04T09:20:35.2458353Z * [new branch] gh/aorenste/154/head -> origin/gh/aorenste/154/head 2025-12-04T09:20:35.2460265Z * [new branch] gh/aorenste/154/orig -> origin/gh/aorenste/154/orig 2025-12-04T09:20:35.2462517Z * [new branch] gh/aorenste/155/base -> origin/gh/aorenste/155/base 2025-12-04T09:20:35.2464365Z * [new branch] gh/aorenste/155/head -> origin/gh/aorenste/155/head 2025-12-04T09:20:35.2466630Z * [new branch] gh/aorenste/155/orig -> origin/gh/aorenste/155/orig 2025-12-04T09:20:35.2468512Z * [new branch] gh/aorenste/156/base -> origin/gh/aorenste/156/base 2025-12-04T09:20:35.2470241Z * [new branch] gh/aorenste/156/head -> origin/gh/aorenste/156/head 2025-12-04T09:20:35.2472043Z * [new branch] gh/aorenste/156/orig -> origin/gh/aorenste/156/orig 2025-12-04T09:20:35.2475339Z * [new branch] gh/aorenste/157/base -> origin/gh/aorenste/157/base 2025-12-04T09:20:35.2476769Z * [new branch] gh/aorenste/157/head -> origin/gh/aorenste/157/head 2025-12-04T09:20:35.2478471Z * [new branch] gh/aorenste/157/orig -> origin/gh/aorenste/157/orig 2025-12-04T09:20:35.2481035Z * [new branch] gh/aorenste/158/base -> origin/gh/aorenste/158/base 2025-12-04T09:20:35.2482612Z * [new branch] gh/aorenste/158/head -> origin/gh/aorenste/158/head 2025-12-04T09:20:35.2484253Z * [new branch] gh/aorenste/158/orig -> origin/gh/aorenste/158/orig 2025-12-04T09:20:35.2487005Z * [new branch] gh/aorenste/159/base -> origin/gh/aorenste/159/base 2025-12-04T09:20:35.2488314Z * [new branch] gh/aorenste/159/head -> origin/gh/aorenste/159/head 2025-12-04T09:20:35.2490051Z * [new branch] gh/aorenste/159/orig -> origin/gh/aorenste/159/orig 2025-12-04T09:20:35.2493309Z * [new branch] gh/avikchaudhuri/1/base -> origin/gh/avikchaudhuri/1/base 2025-12-04T09:20:35.2494822Z * [new branch] gh/avikchaudhuri/1/head -> origin/gh/avikchaudhuri/1/head 2025-12-04T09:20:35.2497352Z * [new branch] gh/avikchaudhuri/2/base -> origin/gh/avikchaudhuri/2/base 2025-12-04T09:20:35.2498799Z * [new branch] gh/avikchaudhuri/2/head -> origin/gh/avikchaudhuri/2/head 2025-12-04T09:20:35.2500656Z * [new branch] gh/avikchaudhuri/2/orig -> origin/gh/avikchaudhuri/2/orig 2025-12-04T09:20:35.2504283Z * [new branch] gh/bdhirsh/666/base -> origin/gh/bdhirsh/666/base 2025-12-04T09:20:35.2505543Z * [new branch] gh/bdhirsh/666/head -> origin/gh/bdhirsh/666/head 2025-12-04T09:20:35.2507245Z * [new branch] gh/bdhirsh/666/orig -> origin/gh/bdhirsh/666/orig 2025-12-04T09:20:35.2509865Z * [new branch] gh/bdhirsh/668/base -> origin/gh/bdhirsh/668/base 2025-12-04T09:20:35.2511333Z * [new branch] gh/bdhirsh/668/head -> origin/gh/bdhirsh/668/head 2025-12-04T09:20:35.2513334Z * [new branch] gh/bdhirsh/668/orig -> origin/gh/bdhirsh/668/orig 2025-12-04T09:20:35.2515948Z * [new branch] gh/bdhirsh/669/base -> origin/gh/bdhirsh/669/base 2025-12-04T09:20:35.2517555Z * [new branch] gh/bdhirsh/669/head -> origin/gh/bdhirsh/669/head 2025-12-04T09:20:35.2519148Z * [new branch] gh/bdhirsh/669/orig -> origin/gh/bdhirsh/669/orig 2025-12-04T09:20:35.2522020Z * [new branch] gh/bdhirsh/670/base -> origin/gh/bdhirsh/670/base 2025-12-04T09:20:35.2523541Z * [new branch] gh/bdhirsh/670/head -> origin/gh/bdhirsh/670/head 2025-12-04T09:20:35.2525293Z * [new branch] gh/bdhirsh/670/orig -> origin/gh/bdhirsh/670/orig 2025-12-04T09:20:35.2528049Z * [new branch] gh/bdhirsh/672/base -> origin/gh/bdhirsh/672/base 2025-12-04T09:20:35.2529499Z * [new branch] gh/bdhirsh/672/head -> origin/gh/bdhirsh/672/head 2025-12-04T09:20:35.2531226Z * [new branch] gh/bdhirsh/672/orig -> origin/gh/bdhirsh/672/orig 2025-12-04T09:20:35.2534072Z * [new branch] gh/bdhirsh/675/base -> origin/gh/bdhirsh/675/base 2025-12-04T09:20:35.2535651Z * [new branch] gh/bdhirsh/675/head -> origin/gh/bdhirsh/675/head 2025-12-04T09:20:35.2537423Z * [new branch] gh/bdhirsh/675/orig -> origin/gh/bdhirsh/675/orig 2025-12-04T09:20:35.2540238Z * [new branch] gh/bdhirsh/676/base -> origin/gh/bdhirsh/676/base 2025-12-04T09:20:35.2541815Z * [new branch] gh/bdhirsh/676/head -> origin/gh/bdhirsh/676/head 2025-12-04T09:20:35.2543612Z * [new branch] gh/bdhirsh/676/orig -> origin/gh/bdhirsh/676/orig 2025-12-04T09:20:35.2546215Z * [new branch] gh/bdhirsh/677/base -> origin/gh/bdhirsh/677/base 2025-12-04T09:20:35.2548030Z * [new branch] gh/bdhirsh/677/head -> origin/gh/bdhirsh/677/head 2025-12-04T09:20:35.2549820Z * [new branch] gh/bdhirsh/677/orig -> origin/gh/bdhirsh/677/orig 2025-12-04T09:20:35.2552563Z * [new branch] gh/bdhirsh/678/base -> origin/gh/bdhirsh/678/base 2025-12-04T09:20:35.2554189Z * [new branch] gh/bdhirsh/678/head -> origin/gh/bdhirsh/678/head 2025-12-04T09:20:35.2555967Z * [new branch] gh/bdhirsh/678/orig -> origin/gh/bdhirsh/678/orig 2025-12-04T09:20:35.2558723Z * [new branch] gh/bdhirsh/679/base -> origin/gh/bdhirsh/679/base 2025-12-04T09:20:35.2560212Z * [new branch] gh/bdhirsh/679/head -> origin/gh/bdhirsh/679/head 2025-12-04T09:20:35.2561977Z * [new branch] gh/bdhirsh/679/orig -> origin/gh/bdhirsh/679/orig 2025-12-04T09:20:35.2564611Z * [new branch] gh/bdhirsh/680/base -> origin/gh/bdhirsh/680/base 2025-12-04T09:20:35.2566138Z * [new branch] gh/bdhirsh/680/head -> origin/gh/bdhirsh/680/head 2025-12-04T09:20:35.2568270Z * [new branch] gh/bdhirsh/680/orig -> origin/gh/bdhirsh/680/orig 2025-12-04T09:20:35.2570431Z * [new branch] gh/bdhirsh/681/base -> origin/gh/bdhirsh/681/base 2025-12-04T09:20:35.2573835Z * [new branch] gh/bdhirsh/681/head -> origin/gh/bdhirsh/681/head 2025-12-04T09:20:35.2575403Z * [new branch] gh/bdhirsh/681/orig -> origin/gh/bdhirsh/681/orig 2025-12-04T09:20:35.2578434Z * [new branch] gh/benjaminglass1/101/base -> origin/gh/benjaminglass1/101/base 2025-12-04T09:20:35.2580043Z * [new branch] gh/benjaminglass1/101/head -> origin/gh/benjaminglass1/101/head 2025-12-04T09:20:35.2582093Z * [new branch] gh/benjaminglass1/101/orig -> origin/gh/benjaminglass1/101/orig 2025-12-04T09:20:35.2584732Z * [new branch] gh/benjaminglass1/102/base -> origin/gh/benjaminglass1/102/base 2025-12-04T09:20:35.2586261Z * [new branch] gh/benjaminglass1/102/head -> origin/gh/benjaminglass1/102/head 2025-12-04T09:20:35.2588040Z * [new branch] gh/benjaminglass1/102/orig -> origin/gh/benjaminglass1/102/orig 2025-12-04T09:20:35.2590671Z * [new branch] gh/benjaminglass1/106/base -> origin/gh/benjaminglass1/106/base 2025-12-04T09:20:35.2592241Z * [new branch] gh/benjaminglass1/106/head -> origin/gh/benjaminglass1/106/head 2025-12-04T09:20:35.2593967Z * [new branch] gh/benjaminglass1/106/orig -> origin/gh/benjaminglass1/106/orig 2025-12-04T09:20:35.2596656Z * [new branch] gh/benjaminglass1/107/base -> origin/gh/benjaminglass1/107/base 2025-12-04T09:20:35.2598132Z * [new branch] gh/benjaminglass1/107/head -> origin/gh/benjaminglass1/107/head 2025-12-04T09:20:35.2599910Z * [new branch] gh/benjaminglass1/107/orig -> origin/gh/benjaminglass1/107/orig 2025-12-04T09:20:35.2602546Z * [new branch] gh/benjaminglass1/108/base -> origin/gh/benjaminglass1/108/base 2025-12-04T09:20:35.2603910Z * [new branch] gh/benjaminglass1/108/head -> origin/gh/benjaminglass1/108/head 2025-12-04T09:20:35.2605658Z * [new branch] gh/benjaminglass1/108/orig -> origin/gh/benjaminglass1/108/orig 2025-12-04T09:20:35.2608306Z * [new branch] gh/benjaminglass1/109/base -> origin/gh/benjaminglass1/109/base 2025-12-04T09:20:35.2609730Z * [new branch] gh/benjaminglass1/109/head -> origin/gh/benjaminglass1/109/head 2025-12-04T09:20:35.2611505Z * [new branch] gh/benjaminglass1/109/orig -> origin/gh/benjaminglass1/109/orig 2025-12-04T09:20:35.2614173Z * [new branch] gh/benjaminglass1/97/base -> origin/gh/benjaminglass1/97/base 2025-12-04T09:20:35.2615739Z * [new branch] gh/benjaminglass1/97/head -> origin/gh/benjaminglass1/97/head 2025-12-04T09:20:35.2617930Z * [new branch] gh/benjaminglass1/97/orig -> origin/gh/benjaminglass1/97/orig 2025-12-04T09:20:35.2620832Z * [new branch] gh/bobrenjc93/570/base -> origin/gh/bobrenjc93/570/base 2025-12-04T09:20:35.2622330Z * [new branch] gh/bobrenjc93/570/head -> origin/gh/bobrenjc93/570/head 2025-12-04T09:20:35.2624010Z * [new branch] gh/bobrenjc93/570/orig -> origin/gh/bobrenjc93/570/orig 2025-12-04T09:20:35.2626614Z * [new branch] gh/bobrenjc93/604/base -> origin/gh/bobrenjc93/604/base 2025-12-04T09:20:35.2628107Z * [new branch] gh/bobrenjc93/604/head -> origin/gh/bobrenjc93/604/head 2025-12-04T09:20:35.2629879Z * [new branch] gh/bobrenjc93/604/orig -> origin/gh/bobrenjc93/604/orig 2025-12-04T09:20:35.2632743Z * [new branch] gh/bobrenjc93/638/base -> origin/gh/bobrenjc93/638/base 2025-12-04T09:20:35.2634208Z * [new branch] gh/bobrenjc93/638/head -> origin/gh/bobrenjc93/638/head 2025-12-04T09:20:35.2635917Z * [new branch] gh/bobrenjc93/638/orig -> origin/gh/bobrenjc93/638/orig 2025-12-04T09:20:35.2638372Z * [new branch] gh/bobrenjc93/653/base -> origin/gh/bobrenjc93/653/base 2025-12-04T09:20:35.2640053Z * [new branch] gh/bobrenjc93/653/head -> origin/gh/bobrenjc93/653/head 2025-12-04T09:20:35.2641770Z * [new branch] gh/bobrenjc93/653/orig -> origin/gh/bobrenjc93/653/orig 2025-12-04T09:20:35.2644412Z * [new branch] gh/bobrenjc93/654/base -> origin/gh/bobrenjc93/654/base 2025-12-04T09:20:35.2646043Z * [new branch] gh/bobrenjc93/654/head -> origin/gh/bobrenjc93/654/head 2025-12-04T09:20:35.2648239Z * [new branch] gh/bobrenjc93/654/orig -> origin/gh/bobrenjc93/654/orig 2025-12-04T09:20:35.2650594Z * [new branch] gh/bobrenjc93/657/base -> origin/gh/bobrenjc93/657/base 2025-12-04T09:20:35.2652290Z * [new branch] gh/bobrenjc93/657/head -> origin/gh/bobrenjc93/657/head 2025-12-04T09:20:35.2653998Z * [new branch] gh/bobrenjc93/657/orig -> origin/gh/bobrenjc93/657/orig 2025-12-04T09:20:35.2656364Z * [new branch] gh/bobrenjc93/672/base -> origin/gh/bobrenjc93/672/base 2025-12-04T09:20:35.2658039Z * [new branch] gh/bobrenjc93/672/head -> origin/gh/bobrenjc93/672/head 2025-12-04T09:20:35.2659967Z * [new branch] gh/bobrenjc93/672/orig -> origin/gh/bobrenjc93/672/orig 2025-12-04T09:20:35.2662451Z * [new branch] gh/bobrenjc93/679/base -> origin/gh/bobrenjc93/679/base 2025-12-04T09:20:35.2664399Z * [new branch] gh/bobrenjc93/679/head -> origin/gh/bobrenjc93/679/head 2025-12-04T09:20:35.2666135Z * [new branch] gh/bobrenjc93/679/orig -> origin/gh/bobrenjc93/679/orig 2025-12-04T09:20:35.2668537Z * [new branch] gh/bobrenjc93/680/base -> origin/gh/bobrenjc93/680/base 2025-12-04T09:20:35.2670301Z * [new branch] gh/bobrenjc93/680/head -> origin/gh/bobrenjc93/680/head 2025-12-04T09:20:35.2672387Z * [new branch] gh/bobrenjc93/680/orig -> origin/gh/bobrenjc93/680/orig 2025-12-04T09:20:35.2674589Z * [new branch] gh/bobrenjc93/681/base -> origin/gh/bobrenjc93/681/base 2025-12-04T09:20:35.2676319Z * [new branch] gh/bobrenjc93/681/head -> origin/gh/bobrenjc93/681/head 2025-12-04T09:20:35.2678131Z * [new branch] gh/bobrenjc93/681/orig -> origin/gh/bobrenjc93/681/orig 2025-12-04T09:20:35.2680616Z * [new branch] gh/bobrenjc93/682/base -> origin/gh/bobrenjc93/682/base 2025-12-04T09:20:35.2682178Z * [new branch] gh/bobrenjc93/682/head -> origin/gh/bobrenjc93/682/head 2025-12-04T09:20:35.2683933Z * [new branch] gh/bobrenjc93/682/orig -> origin/gh/bobrenjc93/682/orig 2025-12-04T09:20:35.2686321Z * [new branch] gh/bobrenjc93/683/base -> origin/gh/bobrenjc93/683/base 2025-12-04T09:20:35.2688094Z * [new branch] gh/bobrenjc93/683/head -> origin/gh/bobrenjc93/683/head 2025-12-04T09:20:35.2689853Z * [new branch] gh/bobrenjc93/683/orig -> origin/gh/bobrenjc93/683/orig 2025-12-04T09:20:35.2692235Z * [new branch] gh/bobrenjc93/684/base -> origin/gh/bobrenjc93/684/base 2025-12-04T09:20:35.2694193Z * [new branch] gh/bobrenjc93/684/head -> origin/gh/bobrenjc93/684/head 2025-12-04T09:20:35.2696062Z * [new branch] gh/bobrenjc93/684/orig -> origin/gh/bobrenjc93/684/orig 2025-12-04T09:20:35.2698271Z * [new branch] gh/bobrenjc93/685/base -> origin/gh/bobrenjc93/685/base 2025-12-04T09:20:35.2700405Z * [new branch] gh/bobrenjc93/685/head -> origin/gh/bobrenjc93/685/head 2025-12-04T09:20:35.2702324Z * [new branch] gh/bobrenjc93/685/orig -> origin/gh/bobrenjc93/685/orig 2025-12-04T09:20:35.2704921Z * [new branch] gh/bobrenjc93/686/base -> origin/gh/bobrenjc93/686/base 2025-12-04T09:20:35.2707667Z * [new branch] gh/bobrenjc93/686/head -> origin/gh/bobrenjc93/686/head 2025-12-04T09:20:35.2708493Z * [new branch] gh/bobrenjc93/686/orig -> origin/gh/bobrenjc93/686/orig 2025-12-04T09:20:35.2710631Z * [new branch] gh/bobrenjc93/687/base -> origin/gh/bobrenjc93/687/base 2025-12-04T09:20:35.2712787Z * [new branch] gh/bobrenjc93/687/head -> origin/gh/bobrenjc93/687/head 2025-12-04T09:20:35.2714435Z * [new branch] gh/bobrenjc93/687/orig -> origin/gh/bobrenjc93/687/orig 2025-12-04T09:20:35.2717304Z * [new branch] gh/bobrenjc93/688/base -> origin/gh/bobrenjc93/688/base 2025-12-04T09:20:35.2719024Z * [new branch] gh/bobrenjc93/688/head -> origin/gh/bobrenjc93/688/head 2025-12-04T09:20:35.2720842Z * [new branch] gh/bobrenjc93/688/orig -> origin/gh/bobrenjc93/688/orig 2025-12-04T09:20:35.2723138Z * [new branch] gh/bobrenjc93/689/base -> origin/gh/bobrenjc93/689/base 2025-12-04T09:20:35.2724946Z * [new branch] gh/bobrenjc93/689/head -> origin/gh/bobrenjc93/689/head 2025-12-04T09:20:35.2726785Z * [new branch] gh/bobrenjc93/689/orig -> origin/gh/bobrenjc93/689/orig 2025-12-04T09:20:35.2729021Z * [new branch] gh/bobrenjc93/690/base -> origin/gh/bobrenjc93/690/base 2025-12-04T09:20:35.2730765Z * [new branch] gh/bobrenjc93/690/head -> origin/gh/bobrenjc93/690/head 2025-12-04T09:20:35.2732533Z * [new branch] gh/bobrenjc93/690/orig -> origin/gh/bobrenjc93/690/orig 2025-12-04T09:20:35.2735606Z * [new branch] gh/bobrenjc93/691/base -> origin/gh/bobrenjc93/691/base 2025-12-04T09:20:35.2737544Z * [new branch] gh/bobrenjc93/691/head -> origin/gh/bobrenjc93/691/head 2025-12-04T09:20:35.2739694Z * [new branch] gh/bobrenjc93/691/orig -> origin/gh/bobrenjc93/691/orig 2025-12-04T09:20:35.2742696Z * [new branch] gh/bobrenjc93/692/base -> origin/gh/bobrenjc93/692/base 2025-12-04T09:20:35.2744462Z * [new branch] gh/bobrenjc93/692/head -> origin/gh/bobrenjc93/692/head 2025-12-04T09:20:35.2746224Z * [new branch] gh/bobrenjc93/692/orig -> origin/gh/bobrenjc93/692/orig 2025-12-04T09:20:35.2748513Z * [new branch] gh/bobrenjc93/693/base -> origin/gh/bobrenjc93/693/base 2025-12-04T09:20:35.2750308Z * [new branch] gh/bobrenjc93/693/head -> origin/gh/bobrenjc93/693/head 2025-12-04T09:20:35.2752055Z * [new branch] gh/bobrenjc93/693/orig -> origin/gh/bobrenjc93/693/orig 2025-12-04T09:20:35.2754491Z * [new branch] gh/bobrenjc93/694/base -> origin/gh/bobrenjc93/694/base 2025-12-04T09:20:35.2756359Z * [new branch] gh/bobrenjc93/694/head -> origin/gh/bobrenjc93/694/head 2025-12-04T09:20:35.2758105Z * [new branch] gh/bobrenjc93/694/orig -> origin/gh/bobrenjc93/694/orig 2025-12-04T09:20:35.2760568Z * [new branch] gh/bobrenjc93/695/base -> origin/gh/bobrenjc93/695/base 2025-12-04T09:20:35.2762367Z * [new branch] gh/bobrenjc93/695/head -> origin/gh/bobrenjc93/695/head 2025-12-04T09:20:35.2764115Z * [new branch] gh/bobrenjc93/695/orig -> origin/gh/bobrenjc93/695/orig 2025-12-04T09:20:35.2767030Z * [new branch] gh/c00w/23/base -> origin/gh/c00w/23/base 2025-12-04T09:20:35.2768790Z * [new branch] gh/c00w/23/head -> origin/gh/c00w/23/head 2025-12-04T09:20:35.2771275Z * [new branch] gh/c00w/53/base -> origin/gh/c00w/53/base 2025-12-04T09:20:35.2773212Z * [new branch] gh/c00w/53/head -> origin/gh/c00w/53/head 2025-12-04T09:20:35.2774932Z * [new branch] gh/c00w/53/orig -> origin/gh/c00w/53/orig 2025-12-04T09:20:35.2777137Z * [new branch] gh/c00w/54/base -> origin/gh/c00w/54/base 2025-12-04T09:20:35.2778951Z * [new branch] gh/c00w/54/head -> origin/gh/c00w/54/head 2025-12-04T09:20:35.2780894Z * [new branch] gh/c00w/54/orig -> origin/gh/c00w/54/orig 2025-12-04T09:20:35.2783267Z * [new branch] gh/c00w/56/base -> origin/gh/c00w/56/base 2025-12-04T09:20:35.2785159Z * [new branch] gh/c00w/56/head -> origin/gh/c00w/56/head 2025-12-04T09:20:35.2786769Z * [new branch] gh/c00w/56/orig -> origin/gh/c00w/56/orig 2025-12-04T09:20:35.2789043Z * [new branch] gh/c00w/57/base -> origin/gh/c00w/57/base 2025-12-04T09:20:35.2791091Z * [new branch] gh/c00w/57/head -> origin/gh/c00w/57/head 2025-12-04T09:20:35.2792915Z * [new branch] gh/c00w/57/orig -> origin/gh/c00w/57/orig 2025-12-04T09:20:35.2795193Z * [new branch] gh/c00w/58/base -> origin/gh/c00w/58/base 2025-12-04T09:20:35.2796954Z * [new branch] gh/c00w/58/head -> origin/gh/c00w/58/head 2025-12-04T09:20:35.2798689Z * [new branch] gh/c00w/58/orig -> origin/gh/c00w/58/orig 2025-12-04T09:20:35.2801585Z * [new branch] gh/clee2000/1/base -> origin/gh/clee2000/1/base 2025-12-04T09:20:35.2803450Z * [new branch] gh/clee2000/1/head -> origin/gh/clee2000/1/head 2025-12-04T09:20:35.2805182Z * [new branch] gh/clee2000/1/orig -> origin/gh/clee2000/1/orig 2025-12-04T09:20:35.2808217Z * [new branch] gh/coconutruben/1/base -> origin/gh/coconutruben/1/base 2025-12-04T09:20:35.2810058Z * [new branch] gh/coconutruben/1/head -> origin/gh/coconutruben/1/head 2025-12-04T09:20:35.2812653Z * [new branch] gh/coconutruben/55/base -> origin/gh/coconutruben/55/base 2025-12-04T09:20:35.2814374Z * [new branch] gh/coconutruben/55/head -> origin/gh/coconutruben/55/head 2025-12-04T09:20:35.2816152Z * [new branch] gh/coconutruben/55/orig -> origin/gh/coconutruben/55/orig 2025-12-04T09:20:35.2818629Z * [new branch] gh/coconutruben/57/base -> origin/gh/coconutruben/57/base 2025-12-04T09:20:35.2820690Z * [new branch] gh/coconutruben/57/head -> origin/gh/coconutruben/57/head 2025-12-04T09:20:35.2822357Z * [new branch] gh/coconutruben/57/orig -> origin/gh/coconutruben/57/orig 2025-12-04T09:20:35.2824822Z * [new branch] gh/coconutruben/70/base -> origin/gh/coconutruben/70/base 2025-12-04T09:20:35.2826633Z * [new branch] gh/coconutruben/70/head -> origin/gh/coconutruben/70/head 2025-12-04T09:20:35.2828450Z * [new branch] gh/coconutruben/70/orig -> origin/gh/coconutruben/70/orig 2025-12-04T09:20:35.2830716Z * [new branch] gh/coconutruben/71/base -> origin/gh/coconutruben/71/base 2025-12-04T09:20:35.2832621Z * [new branch] gh/coconutruben/71/head -> origin/gh/coconutruben/71/head 2025-12-04T09:20:35.2834586Z * [new branch] gh/coconutruben/71/orig -> origin/gh/coconutruben/71/orig 2025-12-04T09:20:35.2836609Z * [new branch] gh/coconutruben/72/base -> origin/gh/coconutruben/72/base 2025-12-04T09:20:35.2838370Z * [new branch] gh/coconutruben/72/head -> origin/gh/coconutruben/72/head 2025-12-04T09:20:35.2840120Z * [new branch] gh/coconutruben/72/orig -> origin/gh/coconutruben/72/orig 2025-12-04T09:20:35.2842264Z * [new branch] gh/coconutruben/73/base -> origin/gh/coconutruben/73/base 2025-12-04T09:20:35.2844068Z * [new branch] gh/coconutruben/73/head -> origin/gh/coconutruben/73/head 2025-12-04T09:20:35.2845820Z * [new branch] gh/coconutruben/73/orig -> origin/gh/coconutruben/73/orig 2025-12-04T09:20:35.2848346Z * [new branch] gh/coconutruben/74/base -> origin/gh/coconutruben/74/base 2025-12-04T09:20:35.2850163Z * [new branch] gh/coconutruben/74/head -> origin/gh/coconutruben/74/head 2025-12-04T09:20:35.2852042Z * [new branch] gh/coconutruben/74/orig -> origin/gh/coconutruben/74/orig 2025-12-04T09:20:35.2854660Z * [new branch] gh/coconutruben/79/base -> origin/gh/coconutruben/79/base 2025-12-04T09:20:35.2856559Z * [new branch] gh/coconutruben/79/head -> origin/gh/coconutruben/79/head 2025-12-04T09:20:35.2858250Z * [new branch] gh/coconutruben/79/orig -> origin/gh/coconutruben/79/orig 2025-12-04T09:20:35.2860778Z * [new branch] gh/coconutruben/80/base -> origin/gh/coconutruben/80/base 2025-12-04T09:20:35.2862572Z * [new branch] gh/coconutruben/80/head -> origin/gh/coconutruben/80/head 2025-12-04T09:20:35.2875538Z * [new branch] gh/coconutruben/80/orig -> origin/gh/coconutruben/80/orig 2025-12-04T09:20:35.2876021Z * [new branch] gh/coconutruben/82/base -> origin/gh/coconutruben/82/base 2025-12-04T09:20:35.2876647Z * [new branch] gh/coconutruben/82/head -> origin/gh/coconutruben/82/head 2025-12-04T09:20:35.2877052Z * [new branch] gh/coconutruben/82/orig -> origin/gh/coconutruben/82/orig 2025-12-04T09:20:35.2877470Z * [new branch] gh/coconutruben/83/base -> origin/gh/coconutruben/83/base 2025-12-04T09:20:35.2877880Z * [new branch] gh/coconutruben/83/head -> origin/gh/coconutruben/83/head 2025-12-04T09:20:35.2878282Z * [new branch] gh/coconutruben/83/orig -> origin/gh/coconutruben/83/orig 2025-12-04T09:20:35.2879056Z * [new branch] gh/coconutruben/84/base -> origin/gh/coconutruben/84/base 2025-12-04T09:20:35.2880975Z * [new branch] gh/coconutruben/84/head -> origin/gh/coconutruben/84/head 2025-12-04T09:20:35.2882545Z * [new branch] gh/coconutruben/84/orig -> origin/gh/coconutruben/84/orig 2025-12-04T09:20:35.2884885Z * [new branch] gh/coconutruben/85/base -> origin/gh/coconutruben/85/base 2025-12-04T09:20:35.2886810Z * [new branch] gh/coconutruben/85/head -> origin/gh/coconutruben/85/head 2025-12-04T09:20:35.2888571Z * [new branch] gh/coconutruben/85/orig -> origin/gh/coconutruben/85/orig 2025-12-04T09:20:35.2890964Z * [new branch] gh/coconutruben/86/base -> origin/gh/coconutruben/86/base 2025-12-04T09:20:35.2892827Z * [new branch] gh/coconutruben/86/head -> origin/gh/coconutruben/86/head 2025-12-04T09:20:35.2894574Z * [new branch] gh/coconutruben/86/orig -> origin/gh/coconutruben/86/orig 2025-12-04T09:20:35.2897481Z * [new branch] gh/colinchan15/1/base -> origin/gh/colinchan15/1/base 2025-12-04T09:20:35.2899329Z * [new branch] gh/colinchan15/1/head -> origin/gh/colinchan15/1/head 2025-12-04T09:20:35.2901642Z * [new branch] gh/colinchan15/2/base -> origin/gh/colinchan15/2/base 2025-12-04T09:20:35.2903371Z * [new branch] gh/colinchan15/2/head -> origin/gh/colinchan15/2/head 2025-12-04T09:20:35.2905611Z * [new branch] gh/colinchan15/3/base -> origin/gh/colinchan15/3/base 2025-12-04T09:20:35.2907261Z * [new branch] gh/colinchan15/3/head -> origin/gh/colinchan15/3/head 2025-12-04T09:20:35.2909531Z * [new branch] gh/colinchan15/6/base -> origin/gh/colinchan15/6/base 2025-12-04T09:20:35.2911304Z * [new branch] gh/colinchan15/6/head -> origin/gh/colinchan15/6/head 2025-12-04T09:20:35.2914188Z * [new branch] gh/d4l3k/1/base -> origin/gh/d4l3k/1/base 2025-12-04T09:20:35.2915979Z * [new branch] gh/d4l3k/1/head -> origin/gh/d4l3k/1/head 2025-12-04T09:20:35.2918259Z * [new branch] gh/d4l3k/2/base -> origin/gh/d4l3k/2/base 2025-12-04T09:20:35.2919999Z * [new branch] gh/d4l3k/2/head -> origin/gh/d4l3k/2/head 2025-12-04T09:20:35.2921857Z * [new branch] gh/d4l3k/2/orig -> origin/gh/d4l3k/2/orig 2025-12-04T09:20:35.2924197Z * [new branch] gh/d4l3k/3/base -> origin/gh/d4l3k/3/base 2025-12-04T09:20:35.2925906Z * [new branch] gh/d4l3k/3/head -> origin/gh/d4l3k/3/head 2025-12-04T09:20:35.2927800Z * [new branch] gh/d4l3k/3/orig -> origin/gh/d4l3k/3/orig 2025-12-04T09:20:35.2930078Z * [new branch] gh/d4l3k/4/base -> origin/gh/d4l3k/4/base 2025-12-04T09:20:35.2931807Z * [new branch] gh/d4l3k/4/head -> origin/gh/d4l3k/4/head 2025-12-04T09:20:35.2933539Z * [new branch] gh/d4l3k/4/orig -> origin/gh/d4l3k/4/orig 2025-12-04T09:20:35.2935840Z * [new branch] gh/d4l3k/5/base -> origin/gh/d4l3k/5/base 2025-12-04T09:20:35.2937580Z * [new branch] gh/d4l3k/5/orig -> origin/gh/d4l3k/5/orig 2025-12-04T09:20:35.2940699Z * [new branch] gh/davidberard98/392/base -> origin/gh/davidberard98/392/base 2025-12-04T09:20:35.2942400Z * [new branch] gh/davidberard98/392/head -> origin/gh/davidberard98/392/head 2025-12-04T09:20:35.2944217Z * [new branch] gh/davidberard98/392/orig -> origin/gh/davidberard98/392/orig 2025-12-04T09:20:35.2946692Z * [new branch] gh/davidberard98/399/base -> origin/gh/davidberard98/399/base 2025-12-04T09:20:35.2948481Z * [new branch] gh/davidberard98/399/head -> origin/gh/davidberard98/399/head 2025-12-04T09:20:35.2950346Z * [new branch] gh/davidberard98/399/orig -> origin/gh/davidberard98/399/orig 2025-12-04T09:20:35.2953104Z * [new branch] gh/desertfire/605/base -> origin/gh/desertfire/605/base 2025-12-04T09:20:35.2954816Z * [new branch] gh/desertfire/605/head -> origin/gh/desertfire/605/head 2025-12-04T09:20:35.2956632Z * [new branch] gh/desertfire/605/orig -> origin/gh/desertfire/605/orig 2025-12-04T09:20:35.2958945Z * [new branch] gh/desertfire/606/base -> origin/gh/desertfire/606/base 2025-12-04T09:20:35.2960655Z * [new branch] gh/desertfire/606/head -> origin/gh/desertfire/606/head 2025-12-04T09:20:35.2962536Z * [new branch] gh/desertfire/606/orig -> origin/gh/desertfire/606/orig 2025-12-04T09:20:35.2964914Z * [new branch] gh/desertfire/607/base -> origin/gh/desertfire/607/base 2025-12-04T09:20:35.2966710Z * [new branch] gh/desertfire/607/head -> origin/gh/desertfire/607/head 2025-12-04T09:20:35.2968483Z * [new branch] gh/desertfire/607/orig -> origin/gh/desertfire/607/orig 2025-12-04T09:20:35.2970796Z * [new branch] gh/desertfire/608/base -> origin/gh/desertfire/608/base 2025-12-04T09:20:35.2974478Z * [new branch] gh/desertfire/608/head -> origin/gh/desertfire/608/head 2025-12-04T09:20:35.2976259Z * [new branch] gh/desertfire/608/orig -> origin/gh/desertfire/608/orig 2025-12-04T09:20:35.2978568Z * [new branch] gh/desertfire/609/base -> origin/gh/desertfire/609/base 2025-12-04T09:20:35.2980521Z * [new branch] gh/desertfire/609/head -> origin/gh/desertfire/609/head 2025-12-04T09:20:35.2982185Z * [new branch] gh/desertfire/609/orig -> origin/gh/desertfire/609/orig 2025-12-04T09:20:35.2984704Z * [new branch] gh/desertfire/610/base -> origin/gh/desertfire/610/base 2025-12-04T09:20:35.2986546Z * [new branch] gh/desertfire/610/head -> origin/gh/desertfire/610/head 2025-12-04T09:20:35.2988265Z * [new branch] gh/desertfire/610/orig -> origin/gh/desertfire/610/orig 2025-12-04T09:20:35.2990578Z * [new branch] gh/desertfire/611/base -> origin/gh/desertfire/611/base 2025-12-04T09:20:35.2992382Z * [new branch] gh/desertfire/611/head -> origin/gh/desertfire/611/head 2025-12-04T09:20:35.2994205Z * [new branch] gh/desertfire/611/orig -> origin/gh/desertfire/611/orig 2025-12-04T09:20:35.2996649Z * [new branch] gh/desertfire/612/base -> origin/gh/desertfire/612/base 2025-12-04T09:20:35.2998575Z * [new branch] gh/desertfire/612/head -> origin/gh/desertfire/612/head 2025-12-04T09:20:35.3000207Z * [new branch] gh/desertfire/612/orig -> origin/gh/desertfire/612/orig 2025-12-04T09:20:35.3002872Z * [new branch] gh/desertfire/613/base -> origin/gh/desertfire/613/base 2025-12-04T09:20:35.3004693Z * [new branch] gh/desertfire/613/head -> origin/gh/desertfire/613/head 2025-12-04T09:20:35.3006519Z * [new branch] gh/desertfire/613/orig -> origin/gh/desertfire/613/orig 2025-12-04T09:20:35.3009037Z * [new branch] gh/desertfire/614/base -> origin/gh/desertfire/614/base 2025-12-04T09:20:35.3010827Z * [new branch] gh/desertfire/614/head -> origin/gh/desertfire/614/head 2025-12-04T09:20:35.3012639Z * [new branch] gh/desertfire/614/orig -> origin/gh/desertfire/614/orig 2025-12-04T09:20:35.3015110Z * [new branch] gh/desertfire/615/base -> origin/gh/desertfire/615/base 2025-12-04T09:20:35.3017101Z * [new branch] gh/desertfire/615/head -> origin/gh/desertfire/615/head 2025-12-04T09:20:35.3018820Z * [new branch] gh/desertfire/615/orig -> origin/gh/desertfire/615/orig 2025-12-04T09:20:35.3021178Z * [new branch] gh/desertfire/616/base -> origin/gh/desertfire/616/base 2025-12-04T09:20:35.3022931Z * [new branch] gh/desertfire/616/head -> origin/gh/desertfire/616/head 2025-12-04T09:20:35.3024651Z * [new branch] gh/desertfire/616/orig -> origin/gh/desertfire/616/orig 2025-12-04T09:20:35.3026854Z * [new branch] gh/desertfire/617/base -> origin/gh/desertfire/617/base 2025-12-04T09:20:35.3028701Z * [new branch] gh/desertfire/617/head -> origin/gh/desertfire/617/head 2025-12-04T09:20:35.3030426Z * [new branch] gh/desertfire/617/orig -> origin/gh/desertfire/617/orig 2025-12-04T09:20:35.3033300Z * [new branch] gh/dharakk/1/base -> origin/gh/dharakk/1/base 2025-12-04T09:20:35.3035579Z * [new branch] gh/dharakk/1/head -> origin/gh/dharakk/1/head 2025-12-04T09:20:35.3038630Z * [new branch] gh/drisspg/170/base -> origin/gh/drisspg/170/base 2025-12-04T09:20:35.3040294Z * [new branch] gh/drisspg/170/head -> origin/gh/drisspg/170/head 2025-12-04T09:20:35.3042126Z * [new branch] gh/drisspg/170/orig -> origin/gh/drisspg/170/orig 2025-12-04T09:20:35.3044480Z * [new branch] gh/drisspg/182/base -> origin/gh/drisspg/182/base 2025-12-04T09:20:35.3046242Z * [new branch] gh/drisspg/182/head -> origin/gh/drisspg/182/head 2025-12-04T09:20:35.3048517Z * [new branch] gh/drisspg/183/base -> origin/gh/drisspg/183/base 2025-12-04T09:20:35.3050137Z * [new branch] gh/drisspg/183/head -> origin/gh/drisspg/183/head 2025-12-04T09:20:35.3052344Z * [new branch] gh/drisspg/184/base -> origin/gh/drisspg/184/base 2025-12-04T09:20:35.3054034Z * [new branch] gh/drisspg/184/head -> origin/gh/drisspg/184/head 2025-12-04T09:20:35.3056411Z * [new branch] gh/drisspg/185/base -> origin/gh/drisspg/185/base 2025-12-04T09:20:35.3058155Z * [new branch] gh/drisspg/185/head -> origin/gh/drisspg/185/head 2025-12-04T09:20:35.3060732Z * [new branch] gh/drisspg/194/base -> origin/gh/drisspg/194/base 2025-12-04T09:20:35.3062448Z * [new branch] gh/drisspg/194/head -> origin/gh/drisspg/194/head 2025-12-04T09:20:35.3064195Z * [new branch] gh/drisspg/194/orig -> origin/gh/drisspg/194/orig 2025-12-04T09:20:35.3066582Z * [new branch] gh/drisspg/200/base -> origin/gh/drisspg/200/base 2025-12-04T09:20:35.3068259Z * [new branch] gh/drisspg/200/head -> origin/gh/drisspg/200/head 2025-12-04T09:20:35.3070151Z * [new branch] gh/drisspg/200/orig -> origin/gh/drisspg/200/orig 2025-12-04T09:20:35.3072671Z * [new branch] gh/drisspg/218/base -> origin/gh/drisspg/218/base 2025-12-04T09:20:35.3074325Z * [new branch] gh/drisspg/218/head -> origin/gh/drisspg/218/head 2025-12-04T09:20:35.3076048Z * [new branch] gh/drisspg/218/orig -> origin/gh/drisspg/218/orig 2025-12-04T09:20:35.3078460Z * [new branch] gh/drisspg/219/base -> origin/gh/drisspg/219/base 2025-12-04T09:20:35.3080179Z * [new branch] gh/drisspg/219/head -> origin/gh/drisspg/219/head 2025-12-04T09:20:35.3081984Z * [new branch] gh/drisspg/219/orig -> origin/gh/drisspg/219/orig 2025-12-04T09:20:35.3084418Z * [new branch] gh/drisspg/220/base -> origin/gh/drisspg/220/base 2025-12-04T09:20:35.3086137Z * [new branch] gh/drisspg/220/head -> origin/gh/drisspg/220/head 2025-12-04T09:20:35.3087899Z * [new branch] gh/drisspg/220/orig -> origin/gh/drisspg/220/orig 2025-12-04T09:20:35.3090788Z * [new branch] gh/drisspg/221/base -> origin/gh/drisspg/221/base 2025-12-04T09:20:35.3092514Z * [new branch] gh/drisspg/221/head -> origin/gh/drisspg/221/head 2025-12-04T09:20:35.3094248Z * [new branch] gh/drisspg/221/orig -> origin/gh/drisspg/221/orig 2025-12-04T09:20:35.3096636Z * [new branch] gh/drisspg/222/base -> origin/gh/drisspg/222/base 2025-12-04T09:20:35.3098370Z * [new branch] gh/drisspg/222/head -> origin/gh/drisspg/222/head 2025-12-04T09:20:35.3100327Z * [new branch] gh/drisspg/222/orig -> origin/gh/drisspg/222/orig 2025-12-04T09:20:35.3102576Z * [new branch] gh/drisspg/223/base -> origin/gh/drisspg/223/base 2025-12-04T09:20:35.3104295Z * [new branch] gh/drisspg/223/head -> origin/gh/drisspg/223/head 2025-12-04T09:20:35.3106103Z * [new branch] gh/drisspg/223/orig -> origin/gh/drisspg/223/orig 2025-12-04T09:20:35.3108471Z * [new branch] gh/drisspg/224/base -> origin/gh/drisspg/224/base 2025-12-04T09:20:35.3110226Z * [new branch] gh/drisspg/224/head -> origin/gh/drisspg/224/head 2025-12-04T09:20:35.3111988Z * [new branch] gh/drisspg/224/orig -> origin/gh/drisspg/224/orig 2025-12-04T09:20:35.3114335Z * [new branch] gh/drisspg/225/base -> origin/gh/drisspg/225/base 2025-12-04T09:20:35.3116078Z * [new branch] gh/drisspg/225/head -> origin/gh/drisspg/225/head 2025-12-04T09:20:35.3117831Z * [new branch] gh/drisspg/225/orig -> origin/gh/drisspg/225/orig 2025-12-04T09:20:35.3120594Z * [new branch] gh/drisspg/226/base -> origin/gh/drisspg/226/base 2025-12-04T09:20:35.3122218Z * [new branch] gh/drisspg/226/head -> origin/gh/drisspg/226/head 2025-12-04T09:20:35.3123950Z * [new branch] gh/drisspg/226/orig -> origin/gh/drisspg/226/orig 2025-12-04T09:20:35.3126775Z * [new branch] gh/drisspg/227/base -> origin/gh/drisspg/227/base 2025-12-04T09:20:35.3128500Z * [new branch] gh/drisspg/227/head -> origin/gh/drisspg/227/head 2025-12-04T09:20:35.3130266Z * [new branch] gh/drisspg/227/orig -> origin/gh/drisspg/227/orig 2025-12-04T09:20:35.3132890Z * [new branch] gh/drisspg/228/base -> origin/gh/drisspg/228/base 2025-12-04T09:20:35.3134666Z * [new branch] gh/drisspg/228/head -> origin/gh/drisspg/228/head 2025-12-04T09:20:35.3136348Z * [new branch] gh/drisspg/228/orig -> origin/gh/drisspg/228/orig 2025-12-04T09:20:35.3138719Z * [new branch] gh/drisspg/229/base -> origin/gh/drisspg/229/base 2025-12-04T09:20:35.3140722Z * [new branch] gh/drisspg/229/head -> origin/gh/drisspg/229/head 2025-12-04T09:20:35.3142472Z * [new branch] gh/drisspg/229/orig -> origin/gh/drisspg/229/orig 2025-12-04T09:20:35.3144843Z * [new branch] gh/drisspg/230/base -> origin/gh/drisspg/230/base 2025-12-04T09:20:35.3146569Z * [new branch] gh/drisspg/230/head -> origin/gh/drisspg/230/head 2025-12-04T09:20:35.3148294Z * [new branch] gh/drisspg/230/orig -> origin/gh/drisspg/230/orig 2025-12-04T09:20:35.3151192Z * [new branch] gh/dsjohns2/1/base -> origin/gh/dsjohns2/1/base 2025-12-04T09:20:35.3153086Z * [new branch] gh/dsjohns2/1/head -> origin/gh/dsjohns2/1/head 2025-12-04T09:20:35.3156087Z * [new branch] gh/dzmitry-huba/1/base -> origin/gh/dzmitry-huba/1/base 2025-12-04T09:20:35.3157857Z * [new branch] gh/dzmitry-huba/1/head -> origin/gh/dzmitry-huba/1/head 2025-12-04T09:20:35.3160809Z * [new branch] gh/dzmitry-huba/12/base -> origin/gh/dzmitry-huba/12/base 2025-12-04T09:20:35.3162648Z * [new branch] gh/dzmitry-huba/12/head -> origin/gh/dzmitry-huba/12/head 2025-12-04T09:20:35.3164416Z * [new branch] gh/dzmitry-huba/12/orig -> origin/gh/dzmitry-huba/12/orig 2025-12-04T09:20:35.3166892Z * [new branch] gh/dzmitry-huba/13/base -> origin/gh/dzmitry-huba/13/base 2025-12-04T09:20:35.3168658Z * [new branch] gh/dzmitry-huba/13/head -> origin/gh/dzmitry-huba/13/head 2025-12-04T09:20:35.3170518Z * [new branch] gh/dzmitry-huba/13/orig -> origin/gh/dzmitry-huba/13/orig 2025-12-04T09:20:35.3172971Z * [new branch] gh/dzmitry-huba/14/base -> origin/gh/dzmitry-huba/14/base 2025-12-04T09:20:35.3174696Z * [new branch] gh/dzmitry-huba/14/head -> origin/gh/dzmitry-huba/14/head 2025-12-04T09:20:35.3176448Z * [new branch] gh/dzmitry-huba/14/orig -> origin/gh/dzmitry-huba/14/orig 2025-12-04T09:20:35.3178970Z * [new branch] gh/dzmitry-huba/15/base -> origin/gh/dzmitry-huba/15/base 2025-12-04T09:20:35.3180862Z * [new branch] gh/dzmitry-huba/15/head -> origin/gh/dzmitry-huba/15/head 2025-12-04T09:20:35.3182363Z * [new branch] gh/dzmitry-huba/15/orig -> origin/gh/dzmitry-huba/15/orig 2025-12-04T09:20:35.3184867Z * [new branch] gh/dzmitry-huba/16/base -> origin/gh/dzmitry-huba/16/base 2025-12-04T09:20:35.3186703Z * [new branch] gh/dzmitry-huba/16/head -> origin/gh/dzmitry-huba/16/head 2025-12-04T09:20:35.3188538Z * [new branch] gh/dzmitry-huba/16/orig -> origin/gh/dzmitry-huba/16/orig 2025-12-04T09:20:35.3190947Z * [new branch] gh/dzmitry-huba/17/base -> origin/gh/dzmitry-huba/17/base 2025-12-04T09:20:35.3192797Z * [new branch] gh/dzmitry-huba/17/head -> origin/gh/dzmitry-huba/17/head 2025-12-04T09:20:35.3194548Z * [new branch] gh/dzmitry-huba/17/orig -> origin/gh/dzmitry-huba/17/orig 2025-12-04T09:20:35.3196773Z * [new branch] gh/dzmitry-huba/2/base -> origin/gh/dzmitry-huba/2/base 2025-12-04T09:20:35.3198514Z * [new branch] gh/dzmitry-huba/2/head -> origin/gh/dzmitry-huba/2/head 2025-12-04T09:20:35.3200716Z * [new branch] gh/dzmitry-huba/3/base -> origin/gh/dzmitry-huba/3/base 2025-12-04T09:20:35.3202475Z * [new branch] gh/dzmitry-huba/3/head -> origin/gh/dzmitry-huba/3/head 2025-12-04T09:20:35.3205508Z * [new branch] gh/eellison/808/base -> origin/gh/eellison/808/base 2025-12-04T09:20:35.3207303Z * [new branch] gh/eellison/808/head -> origin/gh/eellison/808/head 2025-12-04T09:20:35.3209021Z * [new branch] gh/eellison/808/orig -> origin/gh/eellison/808/orig 2025-12-04T09:20:35.3211625Z * [new branch] gh/eellison/822/base -> origin/gh/eellison/822/base 2025-12-04T09:20:35.3213609Z * [new branch] gh/eellison/822/head -> origin/gh/eellison/822/head 2025-12-04T09:20:35.3215268Z * [new branch] gh/eellison/822/orig -> origin/gh/eellison/822/orig 2025-12-04T09:20:35.3217597Z * [new branch] gh/eellison/823/base -> origin/gh/eellison/823/base 2025-12-04T09:20:35.3219398Z * [new branch] gh/eellison/823/head -> origin/gh/eellison/823/head 2025-12-04T09:20:35.3221207Z * [new branch] gh/eellison/823/orig -> origin/gh/eellison/823/orig 2025-12-04T09:20:35.3223449Z * [new branch] gh/eellison/862/base -> origin/gh/eellison/862/base 2025-12-04T09:20:35.3225274Z * [new branch] gh/eellison/862/head -> origin/gh/eellison/862/head 2025-12-04T09:20:35.3226996Z * [new branch] gh/eellison/862/orig -> origin/gh/eellison/862/orig 2025-12-04T09:20:35.3229397Z * [new branch] gh/eellison/863/base -> origin/gh/eellison/863/base 2025-12-04T09:20:35.3231148Z * [new branch] gh/eellison/863/head -> origin/gh/eellison/863/head 2025-12-04T09:20:35.3232985Z * [new branch] gh/eellison/863/orig -> origin/gh/eellison/863/orig 2025-12-04T09:20:35.3235255Z * [new branch] gh/eellison/864/base -> origin/gh/eellison/864/base 2025-12-04T09:20:35.3237211Z * [new branch] gh/eellison/864/head -> origin/gh/eellison/864/head 2025-12-04T09:20:35.3238865Z * [new branch] gh/eellison/864/orig -> origin/gh/eellison/864/orig 2025-12-04T09:20:35.3241282Z * [new branch] gh/eellison/865/base -> origin/gh/eellison/865/base 2025-12-04T09:20:35.3243014Z * [new branch] gh/eellison/865/head -> origin/gh/eellison/865/head 2025-12-04T09:20:35.3244758Z * [new branch] gh/eellison/865/orig -> origin/gh/eellison/865/orig 2025-12-04T09:20:35.3247068Z * [new branch] gh/eellison/866/base -> origin/gh/eellison/866/base 2025-12-04T09:20:35.3248867Z * [new branch] gh/eellison/866/head -> origin/gh/eellison/866/head 2025-12-04T09:20:35.3250618Z * [new branch] gh/eellison/866/orig -> origin/gh/eellison/866/orig 2025-12-04T09:20:35.3253089Z * [new branch] gh/eellison/867/base -> origin/gh/eellison/867/base 2025-12-04T09:20:35.3254813Z * [new branch] gh/eellison/867/head -> origin/gh/eellison/867/head 2025-12-04T09:20:35.3256563Z * [new branch] gh/eellison/867/orig -> origin/gh/eellison/867/orig 2025-12-04T09:20:35.3259185Z * [new branch] gh/eellison/868/base -> origin/gh/eellison/868/base 2025-12-04T09:20:35.3261457Z * [new branch] gh/eellison/868/head -> origin/gh/eellison/868/head 2025-12-04T09:20:35.3262947Z * [new branch] gh/eellison/868/orig -> origin/gh/eellison/868/orig 2025-12-04T09:20:35.3265265Z * [new branch] gh/eellison/869/base -> origin/gh/eellison/869/base 2025-12-04T09:20:35.3267069Z * [new branch] gh/eellison/869/head -> origin/gh/eellison/869/head 2025-12-04T09:20:35.3268740Z * [new branch] gh/eellison/869/orig -> origin/gh/eellison/869/orig 2025-12-04T09:20:35.3271449Z * [new branch] gh/eellison/870/base -> origin/gh/eellison/870/base 2025-12-04T09:20:35.3273743Z * [new branch] gh/eellison/870/head -> origin/gh/eellison/870/head 2025-12-04T09:20:35.3275055Z * [new branch] gh/eellison/870/orig -> origin/gh/eellison/870/orig 2025-12-04T09:20:35.3277438Z * [new branch] gh/eellison/871/base -> origin/gh/eellison/871/base 2025-12-04T09:20:35.3279108Z * [new branch] gh/eellison/871/head -> origin/gh/eellison/871/head 2025-12-04T09:20:35.3280866Z * [new branch] gh/eellison/871/orig -> origin/gh/eellison/871/orig 2025-12-04T09:20:35.3283247Z * [new branch] gh/eellison/872/base -> origin/gh/eellison/872/base 2025-12-04T09:20:35.3284917Z * [new branch] gh/eellison/872/head -> origin/gh/eellison/872/head 2025-12-04T09:20:35.3286668Z * [new branch] gh/eellison/872/orig -> origin/gh/eellison/872/orig 2025-12-04T09:20:35.3289178Z * [new branch] gh/eellison/873/base -> origin/gh/eellison/873/base 2025-12-04T09:20:35.3290892Z * [new branch] gh/eellison/873/head -> origin/gh/eellison/873/head 2025-12-04T09:20:35.3292661Z * [new branch] gh/eellison/873/orig -> origin/gh/eellison/873/orig 2025-12-04T09:20:35.3295054Z * [new branch] gh/eellison/874/base -> origin/gh/eellison/874/base 2025-12-04T09:20:35.3296855Z * [new branch] gh/eellison/874/head -> origin/gh/eellison/874/head 2025-12-04T09:20:35.3298598Z * [new branch] gh/eellison/874/orig -> origin/gh/eellison/874/orig 2025-12-04T09:20:35.3301780Z * [new branch] gh/eellison/875/base -> origin/gh/eellison/875/base 2025-12-04T09:20:35.3303566Z * [new branch] gh/eellison/875/head -> origin/gh/eellison/875/head 2025-12-04T09:20:35.3305344Z * [new branch] gh/eellison/875/orig -> origin/gh/eellison/875/orig 2025-12-04T09:20:35.3307745Z * [new branch] gh/eellison/876/base -> origin/gh/eellison/876/base 2025-12-04T09:20:35.3309565Z * [new branch] gh/eellison/876/head -> origin/gh/eellison/876/head 2025-12-04T09:20:35.3311385Z * [new branch] gh/eellison/876/orig -> origin/gh/eellison/876/orig 2025-12-04T09:20:35.3313767Z * [new branch] gh/eellison/877/base -> origin/gh/eellison/877/base 2025-12-04T09:20:35.3315515Z * [new branch] gh/eellison/877/head -> origin/gh/eellison/877/head 2025-12-04T09:20:35.3317252Z * [new branch] gh/eellison/877/orig -> origin/gh/eellison/877/orig 2025-12-04T09:20:35.3319791Z * [new branch] gh/eellison/878/base -> origin/gh/eellison/878/base 2025-12-04T09:20:35.3321477Z * [new branch] gh/eellison/878/head -> origin/gh/eellison/878/head 2025-12-04T09:20:35.3323188Z * [new branch] gh/eellison/878/orig -> origin/gh/eellison/878/orig 2025-12-04T09:20:35.3325677Z * [new branch] gh/eellison/879/base -> origin/gh/eellison/879/base 2025-12-04T09:20:35.3327416Z * [new branch] gh/eellison/879/head -> origin/gh/eellison/879/head 2025-12-04T09:20:35.3329213Z * [new branch] gh/eellison/879/orig -> origin/gh/eellison/879/orig 2025-12-04T09:20:35.3331517Z * [new branch] gh/eellison/880/base -> origin/gh/eellison/880/base 2025-12-04T09:20:35.3333366Z * [new branch] gh/eellison/880/head -> origin/gh/eellison/880/head 2025-12-04T09:20:35.3335086Z * [new branch] gh/eellison/880/orig -> origin/gh/eellison/880/orig 2025-12-04T09:20:35.3337573Z * [new branch] gh/eellison/881/base -> origin/gh/eellison/881/base 2025-12-04T09:20:35.3339423Z * [new branch] gh/eellison/881/head -> origin/gh/eellison/881/head 2025-12-04T09:20:35.3341208Z * [new branch] gh/eellison/881/orig -> origin/gh/eellison/881/orig 2025-12-04T09:20:35.3343578Z * [new branch] gh/eellison/882/base -> origin/gh/eellison/882/base 2025-12-04T09:20:35.3345422Z * [new branch] gh/eellison/882/head -> origin/gh/eellison/882/head 2025-12-04T09:20:35.3347294Z * [new branch] gh/eellison/882/orig -> origin/gh/eellison/882/orig 2025-12-04T09:20:35.3349658Z * [new branch] gh/eellison/883/base -> origin/gh/eellison/883/base 2025-12-04T09:20:35.3351416Z * [new branch] gh/eellison/883/head -> origin/gh/eellison/883/head 2025-12-04T09:20:35.3353197Z * [new branch] gh/eellison/883/orig -> origin/gh/eellison/883/orig 2025-12-04T09:20:35.3355442Z * [new branch] gh/eellison/884/base -> origin/gh/eellison/884/base 2025-12-04T09:20:35.3357161Z * [new branch] gh/eellison/884/head -> origin/gh/eellison/884/head 2025-12-04T09:20:35.3358853Z * [new branch] gh/eellison/884/orig -> origin/gh/eellison/884/orig 2025-12-04T09:20:35.3362383Z * [new branch] gh/etaf/147/base -> origin/gh/etaf/147/base 2025-12-04T09:20:35.3364089Z * [new branch] gh/etaf/147/head -> origin/gh/etaf/147/head 2025-12-04T09:20:35.3366581Z * [new branch] gh/etaf/154/base -> origin/gh/etaf/154/base 2025-12-04T09:20:35.3368481Z * [new branch] gh/etaf/154/head -> origin/gh/etaf/154/head 2025-12-04T09:20:35.3370175Z * [new branch] gh/etaf/154/orig -> origin/gh/etaf/154/orig 2025-12-04T09:20:35.3374403Z * [new branch] gh/etaf/156/base -> origin/gh/etaf/156/base 2025-12-04T09:20:35.3376127Z * [new branch] gh/etaf/156/head -> origin/gh/etaf/156/head 2025-12-04T09:20:35.3377889Z * [new branch] gh/etaf/156/orig -> origin/gh/etaf/156/orig 2025-12-04T09:20:35.3380575Z * [new branch] gh/etaf/157/base -> origin/gh/etaf/157/base 2025-12-04T09:20:35.3382375Z * [new branch] gh/etaf/157/head -> origin/gh/etaf/157/head 2025-12-04T09:20:35.3384116Z * [new branch] gh/etaf/157/orig -> origin/gh/etaf/157/orig 2025-12-04T09:20:35.3386498Z * [new branch] gh/etaf/158/base -> origin/gh/etaf/158/base 2025-12-04T09:20:35.3388741Z * [new branch] gh/etaf/158/head -> origin/gh/etaf/158/head 2025-12-04T09:20:35.3390506Z * [new branch] gh/etaf/158/orig -> origin/gh/etaf/158/orig 2025-12-04T09:20:35.3392933Z * [new branch] gh/etaf/159/base -> origin/gh/etaf/159/base 2025-12-04T09:20:35.3394864Z * [new branch] gh/etaf/159/head -> origin/gh/etaf/159/head 2025-12-04T09:20:35.3396581Z * [new branch] gh/etaf/159/orig -> origin/gh/etaf/159/orig 2025-12-04T09:20:35.3399113Z * [new branch] gh/etaf/160/base -> origin/gh/etaf/160/base 2025-12-04T09:20:35.3400890Z * [new branch] gh/etaf/160/head -> origin/gh/etaf/160/head 2025-12-04T09:20:35.3402690Z * [new branch] gh/etaf/160/orig -> origin/gh/etaf/160/orig 2025-12-04T09:20:35.3405140Z * [new branch] gh/etaf/161/base -> origin/gh/etaf/161/base 2025-12-04T09:20:35.3406959Z * [new branch] gh/etaf/161/head -> origin/gh/etaf/161/head 2025-12-04T09:20:35.3408706Z * [new branch] gh/etaf/161/orig -> origin/gh/etaf/161/orig 2025-12-04T09:20:35.3411103Z * [new branch] gh/etaf/166/base -> origin/gh/etaf/166/base 2025-12-04T09:20:35.3413012Z * [new branch] gh/etaf/166/head -> origin/gh/etaf/166/head 2025-12-04T09:20:35.3414765Z * [new branch] gh/etaf/166/orig -> origin/gh/etaf/166/orig 2025-12-04T09:20:35.3417047Z * [new branch] gh/etaf/167/base -> origin/gh/etaf/167/base 2025-12-04T09:20:35.3418994Z * [new branch] gh/etaf/167/head -> origin/gh/etaf/167/head 2025-12-04T09:20:35.3420934Z * [new branch] gh/etaf/167/orig -> origin/gh/etaf/167/orig 2025-12-04T09:20:35.3423291Z * [new branch] gh/etaf/168/base -> origin/gh/etaf/168/base 2025-12-04T09:20:35.3425086Z * [new branch] gh/etaf/168/head -> origin/gh/etaf/168/head 2025-12-04T09:20:35.3426842Z * [new branch] gh/etaf/168/orig -> origin/gh/etaf/168/orig 2025-12-04T09:20:35.3429381Z * [new branch] gh/etaf/172/base -> origin/gh/etaf/172/base 2025-12-04T09:20:35.3431052Z * [new branch] gh/etaf/172/head -> origin/gh/etaf/172/head 2025-12-04T09:20:35.3432831Z * [new branch] gh/etaf/172/orig -> origin/gh/etaf/172/orig 2025-12-04T09:20:35.3435324Z * [new branch] gh/etaf/173/base -> origin/gh/etaf/173/base 2025-12-04T09:20:35.3437183Z * [new branch] gh/etaf/173/head -> origin/gh/etaf/173/head 2025-12-04T09:20:35.3438952Z * [new branch] gh/etaf/173/orig -> origin/gh/etaf/173/orig 2025-12-04T09:20:35.3441385Z * [new branch] gh/etaf/174/base -> origin/gh/etaf/174/base 2025-12-04T09:20:35.3443278Z * [new branch] gh/etaf/174/head -> origin/gh/etaf/174/head 2025-12-04T09:20:35.3445646Z * [new branch] gh/etaf/175/base -> origin/gh/etaf/175/base 2025-12-04T09:20:35.3447491Z * [new branch] gh/etaf/175/head -> origin/gh/etaf/175/head 2025-12-04T09:20:35.3449116Z * [new branch] gh/etaf/175/orig -> origin/gh/etaf/175/orig 2025-12-04T09:20:35.3451497Z * [new branch] gh/etaf/176/base -> origin/gh/etaf/176/base 2025-12-04T09:20:35.3453309Z * [new branch] gh/etaf/176/head -> origin/gh/etaf/176/head 2025-12-04T09:20:35.3455077Z * [new branch] gh/etaf/176/orig -> origin/gh/etaf/176/orig 2025-12-04T09:20:35.3457945Z * [new branch] gh/etaf/177/base -> origin/gh/etaf/177/base 2025-12-04T09:20:35.3460014Z * [new branch] gh/etaf/177/head -> origin/gh/etaf/177/head 2025-12-04T09:20:35.3461794Z * [new branch] gh/etaf/177/orig -> origin/gh/etaf/177/orig 2025-12-04T09:20:35.3464281Z * [new branch] gh/etaf/178/base -> origin/gh/etaf/178/base 2025-12-04T09:20:35.3466159Z * [new branch] gh/etaf/178/head -> origin/gh/etaf/178/head 2025-12-04T09:20:35.3467951Z * [new branch] gh/etaf/178/orig -> origin/gh/etaf/178/orig 2025-12-04T09:20:35.3470420Z * [new branch] gh/etaf/179/base -> origin/gh/etaf/179/base 2025-12-04T09:20:35.3472411Z * [new branch] gh/etaf/179/head -> origin/gh/etaf/179/head 2025-12-04T09:20:35.3474130Z * [new branch] gh/etaf/179/orig -> origin/gh/etaf/179/orig 2025-12-04T09:20:35.3476480Z * [new branch] gh/etaf/180/base -> origin/gh/etaf/180/base 2025-12-04T09:20:35.3478211Z * [new branch] gh/etaf/180/head -> origin/gh/etaf/180/head 2025-12-04T09:20:35.3479953Z * [new branch] gh/etaf/180/orig -> origin/gh/etaf/180/orig 2025-12-04T09:20:35.3483416Z * [new branch] gh/exclamaforte/1/base -> origin/gh/exclamaforte/1/base 2025-12-04T09:20:35.3485024Z * [new branch] gh/exclamaforte/1/head -> origin/gh/exclamaforte/1/head 2025-12-04T09:20:35.3487282Z * [new branch] gh/exclamaforte/2/base -> origin/gh/exclamaforte/2/base 2025-12-04T09:20:35.3488989Z * [new branch] gh/exclamaforte/2/head -> origin/gh/exclamaforte/2/head 2025-12-04T09:20:35.3491438Z * [new branch] gh/exclamaforte/3/base -> origin/gh/exclamaforte/3/base 2025-12-04T09:20:35.3493187Z * [new branch] gh/exclamaforte/3/head -> origin/gh/exclamaforte/3/head 2025-12-04T09:20:35.3495984Z * [new branch] gh/exclamaforte/4/base -> origin/gh/exclamaforte/4/base 2025-12-04T09:20:35.3497790Z * [new branch] gh/exclamaforte/4/head -> origin/gh/exclamaforte/4/head 2025-12-04T09:20:35.3500917Z * [new branch] gh/ezyang/2374/base -> origin/gh/ezyang/2374/base 2025-12-04T09:20:35.3502658Z * [new branch] gh/ezyang/2374/head -> origin/gh/ezyang/2374/head 2025-12-04T09:20:35.3504624Z * [new branch] gh/ezyang/2374/orig -> origin/gh/ezyang/2374/orig 2025-12-04T09:20:35.3506780Z * [new branch] gh/ezyang/2973/base -> origin/gh/ezyang/2973/base 2025-12-04T09:20:35.3508468Z * [new branch] gh/ezyang/2973/head -> origin/gh/ezyang/2973/head 2025-12-04T09:20:35.3510271Z * [new branch] gh/ezyang/2973/orig -> origin/gh/ezyang/2973/orig 2025-12-04T09:20:35.3512733Z * [new branch] gh/ezyang/2974/base -> origin/gh/ezyang/2974/base 2025-12-04T09:20:35.3514443Z * [new branch] gh/ezyang/2974/head -> origin/gh/ezyang/2974/head 2025-12-04T09:20:35.3516186Z * [new branch] gh/ezyang/2974/orig -> origin/gh/ezyang/2974/orig 2025-12-04T09:20:35.3518539Z * [new branch] gh/ezyang/3131/base -> origin/gh/ezyang/3131/base 2025-12-04T09:20:35.3520280Z * [new branch] gh/ezyang/3131/head -> origin/gh/ezyang/3131/head 2025-12-04T09:20:35.3522027Z * [new branch] gh/ezyang/3131/orig -> origin/gh/ezyang/3131/orig 2025-12-04T09:20:35.3524465Z * [new branch] gh/ezyang/3139/base -> origin/gh/ezyang/3139/base 2025-12-04T09:20:35.3526250Z * [new branch] gh/ezyang/3139/head -> origin/gh/ezyang/3139/head 2025-12-04T09:20:35.3527993Z * [new branch] gh/ezyang/3139/orig -> origin/gh/ezyang/3139/orig 2025-12-04T09:20:35.3530309Z * [new branch] gh/ezyang/3140/base -> origin/gh/ezyang/3140/base 2025-12-04T09:20:35.3532067Z * [new branch] gh/ezyang/3140/head -> origin/gh/ezyang/3140/head 2025-12-04T09:20:35.3533838Z * [new branch] gh/ezyang/3140/orig -> origin/gh/ezyang/3140/orig 2025-12-04T09:20:35.3536308Z * [new branch] gh/ezyang/3143/base -> origin/gh/ezyang/3143/base 2025-12-04T09:20:35.3538036Z * [new branch] gh/ezyang/3143/head -> origin/gh/ezyang/3143/head 2025-12-04T09:20:35.3539978Z * [new branch] gh/ezyang/3143/orig -> origin/gh/ezyang/3143/orig 2025-12-04T09:20:35.3542272Z * [new branch] gh/ezyang/3144/base -> origin/gh/ezyang/3144/base 2025-12-04T09:20:35.3544005Z * [new branch] gh/ezyang/3144/head -> origin/gh/ezyang/3144/head 2025-12-04T09:20:35.3545750Z * [new branch] gh/ezyang/3144/orig -> origin/gh/ezyang/3144/orig 2025-12-04T09:20:35.3548107Z * [new branch] gh/ezyang/3167/base -> origin/gh/ezyang/3167/base 2025-12-04T09:20:35.3549844Z * [new branch] gh/ezyang/3167/head -> origin/gh/ezyang/3167/head 2025-12-04T09:20:35.3551602Z * [new branch] gh/ezyang/3167/orig -> origin/gh/ezyang/3167/orig 2025-12-04T09:20:35.3553940Z * [new branch] gh/ezyang/3173/base -> origin/gh/ezyang/3173/base 2025-12-04T09:20:35.3555663Z * [new branch] gh/ezyang/3173/head -> origin/gh/ezyang/3173/head 2025-12-04T09:20:35.3557447Z * [new branch] gh/ezyang/3173/orig -> origin/gh/ezyang/3173/orig 2025-12-04T09:20:35.3559887Z * [new branch] gh/ezyang/3175/base -> origin/gh/ezyang/3175/base 2025-12-04T09:20:35.3561747Z * [new branch] gh/ezyang/3175/head -> origin/gh/ezyang/3175/head 2025-12-04T09:20:35.3563464Z * [new branch] gh/ezyang/3175/orig -> origin/gh/ezyang/3175/orig 2025-12-04T09:20:35.3565785Z * [new branch] gh/ezyang/3182/base -> origin/gh/ezyang/3182/base 2025-12-04T09:20:35.3567525Z * [new branch] gh/ezyang/3182/head -> origin/gh/ezyang/3182/head 2025-12-04T09:20:35.3569273Z * [new branch] gh/ezyang/3182/orig -> origin/gh/ezyang/3182/orig 2025-12-04T09:20:35.3571782Z * [new branch] gh/ezyang/3185/base -> origin/gh/ezyang/3185/base 2025-12-04T09:20:35.3573812Z * [new branch] gh/ezyang/3185/head -> origin/gh/ezyang/3185/head 2025-12-04T09:20:35.3575331Z * [new branch] gh/ezyang/3185/orig -> origin/gh/ezyang/3185/orig 2025-12-04T09:20:35.3577589Z * [new branch] gh/ezyang/3189/base -> origin/gh/ezyang/3189/base 2025-12-04T09:20:35.3579387Z * [new branch] gh/ezyang/3189/head -> origin/gh/ezyang/3189/head 2025-12-04T09:20:35.3581153Z * [new branch] gh/ezyang/3189/orig -> origin/gh/ezyang/3189/orig 2025-12-04T09:20:35.3583613Z * [new branch] gh/ezyang/3191/base -> origin/gh/ezyang/3191/base 2025-12-04T09:20:35.3585391Z * [new branch] gh/ezyang/3191/head -> origin/gh/ezyang/3191/head 2025-12-04T09:20:35.3587240Z * [new branch] gh/ezyang/3191/orig -> origin/gh/ezyang/3191/orig 2025-12-04T09:20:35.3589967Z * [new branch] gh/ezyang/3192/base -> origin/gh/ezyang/3192/base 2025-12-04T09:20:35.3591728Z * [new branch] gh/ezyang/3192/head -> origin/gh/ezyang/3192/head 2025-12-04T09:20:35.3593621Z * [new branch] gh/ezyang/3192/orig -> origin/gh/ezyang/3192/orig 2025-12-04T09:20:35.3596008Z * [new branch] gh/ezyang/3193/base -> origin/gh/ezyang/3193/base 2025-12-04T09:20:35.3597736Z * [new branch] gh/ezyang/3193/head -> origin/gh/ezyang/3193/head 2025-12-04T09:20:35.3599521Z * [new branch] gh/ezyang/3193/orig -> origin/gh/ezyang/3193/orig 2025-12-04T09:20:35.3601830Z * [new branch] gh/ezyang/3194/base -> origin/gh/ezyang/3194/base 2025-12-04T09:20:35.3603508Z * [new branch] gh/ezyang/3194/head -> origin/gh/ezyang/3194/head 2025-12-04T09:20:35.3605285Z * [new branch] gh/ezyang/3194/orig -> origin/gh/ezyang/3194/orig 2025-12-04T09:20:35.3607742Z * [new branch] gh/ezyang/3195/base -> origin/gh/ezyang/3195/base 2025-12-04T09:20:35.3609515Z * [new branch] gh/ezyang/3195/head -> origin/gh/ezyang/3195/head 2025-12-04T09:20:35.3611246Z * [new branch] gh/ezyang/3195/orig -> origin/gh/ezyang/3195/orig 2025-12-04T09:20:35.3613703Z * [new branch] gh/ezyang/3196/base -> origin/gh/ezyang/3196/base 2025-12-04T09:20:35.3615514Z * [new branch] gh/ezyang/3196/head -> origin/gh/ezyang/3196/head 2025-12-04T09:20:35.3617349Z * [new branch] gh/ezyang/3196/orig -> origin/gh/ezyang/3196/orig 2025-12-04T09:20:35.3619844Z * [new branch] gh/ezyang/3197/base -> origin/gh/ezyang/3197/base 2025-12-04T09:20:35.3621563Z * [new branch] gh/ezyang/3197/head -> origin/gh/ezyang/3197/head 2025-12-04T09:20:35.3623294Z * [new branch] gh/ezyang/3197/orig -> origin/gh/ezyang/3197/orig 2025-12-04T09:20:35.3625720Z * [new branch] gh/ezyang/3198/base -> origin/gh/ezyang/3198/base 2025-12-04T09:20:35.3627493Z * [new branch] gh/ezyang/3198/head -> origin/gh/ezyang/3198/head 2025-12-04T09:20:35.3629447Z * [new branch] gh/ezyang/3198/orig -> origin/gh/ezyang/3198/orig 2025-12-04T09:20:35.3632349Z * [new branch] gh/ezyang/3199/base -> origin/gh/ezyang/3199/base 2025-12-04T09:20:35.3634215Z * [new branch] gh/ezyang/3199/head -> origin/gh/ezyang/3199/head 2025-12-04T09:20:35.3635983Z * [new branch] gh/ezyang/3199/orig -> origin/gh/ezyang/3199/orig 2025-12-04T09:20:35.3638417Z * [new branch] gh/ezyang/3200/base -> origin/gh/ezyang/3200/base 2025-12-04T09:20:35.3640181Z * [new branch] gh/ezyang/3200/head -> origin/gh/ezyang/3200/head 2025-12-04T09:20:35.3641915Z * [new branch] gh/ezyang/3200/orig -> origin/gh/ezyang/3200/orig 2025-12-04T09:20:35.3644668Z * [new branch] gh/ezyang/3201/base -> origin/gh/ezyang/3201/base 2025-12-04T09:20:35.3646247Z * [new branch] gh/ezyang/3201/head -> origin/gh/ezyang/3201/head 2025-12-04T09:20:35.3648012Z * [new branch] gh/ezyang/3201/orig -> origin/gh/ezyang/3201/orig 2025-12-04T09:20:35.3650443Z * [new branch] gh/ezyang/3202/base -> origin/gh/ezyang/3202/base 2025-12-04T09:20:35.3652112Z * [new branch] gh/ezyang/3202/head -> origin/gh/ezyang/3202/head 2025-12-04T09:20:35.3653825Z * [new branch] gh/ezyang/3202/orig -> origin/gh/ezyang/3202/orig 2025-12-04T09:20:35.3656315Z * [new branch] gh/ezyang/3203/base -> origin/gh/ezyang/3203/base 2025-12-04T09:20:35.3658140Z * [new branch] gh/ezyang/3203/head -> origin/gh/ezyang/3203/head 2025-12-04T09:20:35.3660095Z * [new branch] gh/ezyang/3203/orig -> origin/gh/ezyang/3203/orig 2025-12-04T09:20:35.3662351Z * [new branch] gh/ezyang/3204/base -> origin/gh/ezyang/3204/base 2025-12-04T09:20:35.3664128Z * [new branch] gh/ezyang/3204/head -> origin/gh/ezyang/3204/head 2025-12-04T09:20:35.3665802Z * [new branch] gh/ezyang/3204/orig -> origin/gh/ezyang/3204/orig 2025-12-04T09:20:35.3668211Z * [new branch] gh/ezyang/3205/base -> origin/gh/ezyang/3205/base 2025-12-04T09:20:35.3669907Z * [new branch] gh/ezyang/3205/head -> origin/gh/ezyang/3205/head 2025-12-04T09:20:35.3671738Z * [new branch] gh/ezyang/3205/orig -> origin/gh/ezyang/3205/orig 2025-12-04T09:20:35.3674813Z * [new branch] gh/ezyang/3206/base -> origin/gh/ezyang/3206/base 2025-12-04T09:20:35.3676524Z * [new branch] gh/ezyang/3206/head -> origin/gh/ezyang/3206/head 2025-12-04T09:20:35.3678290Z * [new branch] gh/ezyang/3206/orig -> origin/gh/ezyang/3206/orig 2025-12-04T09:20:35.3680768Z * [new branch] gh/ezyang/3207/base -> origin/gh/ezyang/3207/base 2025-12-04T09:20:35.3682520Z * [new branch] gh/ezyang/3207/head -> origin/gh/ezyang/3207/head 2025-12-04T09:20:35.3684345Z * [new branch] gh/ezyang/3207/orig -> origin/gh/ezyang/3207/orig 2025-12-04T09:20:35.3686717Z * [new branch] gh/ezyang/3208/base -> origin/gh/ezyang/3208/base 2025-12-04T09:20:35.3688471Z * [new branch] gh/ezyang/3208/head -> origin/gh/ezyang/3208/head 2025-12-04T09:20:35.3690216Z * [new branch] gh/ezyang/3208/orig -> origin/gh/ezyang/3208/orig 2025-12-04T09:20:35.3692531Z * [new branch] gh/ezyang/3209/base -> origin/gh/ezyang/3209/base 2025-12-04T09:20:35.3694802Z * [new branch] gh/ezyang/3209/head -> origin/gh/ezyang/3209/head 2025-12-04T09:20:35.3696567Z * [new branch] gh/ezyang/3209/orig -> origin/gh/ezyang/3209/orig 2025-12-04T09:20:35.3699511Z * [new branch] gh/fadara01/3/base -> origin/gh/fadara01/3/base 2025-12-04T09:20:35.3701255Z * [new branch] gh/fadara01/3/head -> origin/gh/fadara01/3/head 2025-12-04T09:20:35.3702862Z * [new branch] gh/fadara01/3/orig -> origin/gh/fadara01/3/orig 2025-12-04T09:20:35.3705382Z * [new branch] gh/fadara01/5/base -> origin/gh/fadara01/5/base 2025-12-04T09:20:35.3707107Z * [new branch] gh/fadara01/5/head -> origin/gh/fadara01/5/head 2025-12-04T09:20:35.3708829Z * [new branch] gh/fadara01/5/orig -> origin/gh/fadara01/5/orig 2025-12-04T09:20:35.3711240Z * [new branch] gh/fadara01/6/base -> origin/gh/fadara01/6/base 2025-12-04T09:20:35.3713031Z * [new branch] gh/fadara01/6/head -> origin/gh/fadara01/6/head 2025-12-04T09:20:35.3714705Z * [new branch] gh/fadara01/6/orig -> origin/gh/fadara01/6/orig 2025-12-04T09:20:35.3717199Z * [new branch] gh/fadara01/7/base -> origin/gh/fadara01/7/base 2025-12-04T09:20:35.3718859Z * [new branch] gh/fadara01/7/head -> origin/gh/fadara01/7/head 2025-12-04T09:20:35.3720689Z * [new branch] gh/fadara01/7/orig -> origin/gh/fadara01/7/orig 2025-12-04T09:20:35.3723020Z * [new branch] gh/fadara01/8/base -> origin/gh/fadara01/8/base 2025-12-04T09:20:35.3724773Z * [new branch] gh/fadara01/8/head -> origin/gh/fadara01/8/head 2025-12-04T09:20:35.3726551Z * [new branch] gh/fadara01/8/orig -> origin/gh/fadara01/8/orig 2025-12-04T09:20:35.3728955Z * [new branch] gh/fadara01/9/base -> origin/gh/fadara01/9/base 2025-12-04T09:20:35.3730659Z * [new branch] gh/fadara01/9/head -> origin/gh/fadara01/9/head 2025-12-04T09:20:35.3732435Z * [new branch] gh/fadara01/9/orig -> origin/gh/fadara01/9/orig 2025-12-04T09:20:35.3735285Z * [new branch] gh/fduwjj/182/base -> origin/gh/fduwjj/182/base 2025-12-04T09:20:35.3737023Z * [new branch] gh/fduwjj/182/head -> origin/gh/fduwjj/182/head 2025-12-04T09:20:35.3738777Z * [new branch] gh/fduwjj/182/orig -> origin/gh/fduwjj/182/orig 2025-12-04T09:20:35.3741372Z * [new branch] gh/fduwjj/211/base -> origin/gh/fduwjj/211/base 2025-12-04T09:20:35.3742974Z * [new branch] gh/fduwjj/211/head -> origin/gh/fduwjj/211/head 2025-12-04T09:20:35.3744691Z * [new branch] gh/fduwjj/211/orig -> origin/gh/fduwjj/211/orig 2025-12-04T09:20:35.3747043Z * [new branch] gh/fduwjj/212/base -> origin/gh/fduwjj/212/base 2025-12-04T09:20:35.3748833Z * [new branch] gh/fduwjj/212/head -> origin/gh/fduwjj/212/head 2025-12-04T09:20:35.3751082Z * [new branch] gh/fduwjj/212/orig -> origin/gh/fduwjj/212/orig 2025-12-04T09:20:35.3753552Z * [new branch] gh/fduwjj/213/base -> origin/gh/fduwjj/213/base 2025-12-04T09:20:35.3755277Z * [new branch] gh/fduwjj/213/head -> origin/gh/fduwjj/213/head 2025-12-04T09:20:35.3756981Z * [new branch] gh/fduwjj/213/orig -> origin/gh/fduwjj/213/orig 2025-12-04T09:20:35.3759432Z * [new branch] gh/fduwjj/226/base -> origin/gh/fduwjj/226/base 2025-12-04T09:20:35.3761124Z * [new branch] gh/fduwjj/226/head -> origin/gh/fduwjj/226/head 2025-12-04T09:20:35.3762750Z * [new branch] gh/fduwjj/226/orig -> origin/gh/fduwjj/226/orig 2025-12-04T09:20:35.3765344Z * [new branch] gh/fduwjj/229/base -> origin/gh/fduwjj/229/base 2025-12-04T09:20:35.3767023Z * [new branch] gh/fduwjj/229/head -> origin/gh/fduwjj/229/head 2025-12-04T09:20:35.3768805Z * [new branch] gh/fduwjj/229/orig -> origin/gh/fduwjj/229/orig 2025-12-04T09:20:35.3771338Z * [new branch] gh/fduwjj/233/base -> origin/gh/fduwjj/233/base 2025-12-04T09:20:35.3773360Z * [new branch] gh/fduwjj/233/head -> origin/gh/fduwjj/233/head 2025-12-04T09:20:35.3775196Z * [new branch] gh/fduwjj/233/orig -> origin/gh/fduwjj/233/orig 2025-12-04T09:20:35.3777540Z * [new branch] gh/fduwjj/234/base -> origin/gh/fduwjj/234/base 2025-12-04T09:20:35.3779333Z * [new branch] gh/fduwjj/234/head -> origin/gh/fduwjj/234/head 2025-12-04T09:20:35.3781136Z * [new branch] gh/fduwjj/234/orig -> origin/gh/fduwjj/234/orig 2025-12-04T09:20:35.3783900Z * [new branch] gh/fduwjj/235/base -> origin/gh/fduwjj/235/base 2025-12-04T09:20:35.3785639Z * [new branch] gh/fduwjj/235/head -> origin/gh/fduwjj/235/head 2025-12-04T09:20:35.3787406Z * [new branch] gh/fduwjj/235/orig -> origin/gh/fduwjj/235/orig 2025-12-04T09:20:35.3789852Z * [new branch] gh/fduwjj/236/base -> origin/gh/fduwjj/236/base 2025-12-04T09:20:35.3791432Z * [new branch] gh/fduwjj/236/head -> origin/gh/fduwjj/236/head 2025-12-04T09:20:35.3793118Z * [new branch] gh/fduwjj/236/orig -> origin/gh/fduwjj/236/orig 2025-12-04T09:20:35.3795370Z * [new branch] gh/fduwjj/237/base -> origin/gh/fduwjj/237/base 2025-12-04T09:20:35.3797072Z * [new branch] gh/fduwjj/237/head -> origin/gh/fduwjj/237/head 2025-12-04T09:20:35.3798877Z * [new branch] gh/fduwjj/237/orig -> origin/gh/fduwjj/237/orig 2025-12-04T09:20:35.3801253Z * [new branch] gh/fduwjj/238/base -> origin/gh/fduwjj/238/base 2025-12-04T09:20:35.3803120Z * [new branch] gh/fduwjj/238/head -> origin/gh/fduwjj/238/head 2025-12-04T09:20:35.3804831Z * [new branch] gh/fduwjj/238/orig -> origin/gh/fduwjj/238/orig 2025-12-04T09:20:35.3807236Z * [new branch] gh/fduwjj/239/base -> origin/gh/fduwjj/239/base 2025-12-04T09:20:35.3809020Z * [new branch] gh/fduwjj/239/head -> origin/gh/fduwjj/239/head 2025-12-04T09:20:35.3810756Z * [new branch] gh/fduwjj/239/orig -> origin/gh/fduwjj/239/orig 2025-12-04T09:20:35.3813578Z * [new branch] gh/fegin/332/base -> origin/gh/fegin/332/base 2025-12-04T09:20:35.3815327Z * [new branch] gh/fegin/332/head -> origin/gh/fegin/332/head 2025-12-04T09:20:35.3817150Z * [new branch] gh/fegin/332/orig -> origin/gh/fegin/332/orig 2025-12-04T09:20:35.3819585Z * [new branch] gh/fegin/333/base -> origin/gh/fegin/333/base 2025-12-04T09:20:35.3821455Z * [new branch] gh/fegin/333/head -> origin/gh/fegin/333/head 2025-12-04T09:20:35.3823229Z * [new branch] gh/fegin/333/orig -> origin/gh/fegin/333/orig 2025-12-04T09:20:35.3825562Z * [new branch] gh/fegin/334/base -> origin/gh/fegin/334/base 2025-12-04T09:20:35.3827297Z * [new branch] gh/fegin/334/head -> origin/gh/fegin/334/head 2025-12-04T09:20:35.3829146Z * [new branch] gh/fegin/334/orig -> origin/gh/fegin/334/orig 2025-12-04T09:20:35.3831503Z * [new branch] gh/fegin/335/base -> origin/gh/fegin/335/base 2025-12-04T09:20:35.3833258Z * [new branch] gh/fegin/335/head -> origin/gh/fegin/335/head 2025-12-04T09:20:35.3835061Z * [new branch] gh/fegin/335/orig -> origin/gh/fegin/335/orig 2025-12-04T09:20:35.3838038Z * [new branch] gh/fffrog/160/base -> origin/gh/fffrog/160/base 2025-12-04T09:20:35.3839793Z * [new branch] gh/fffrog/160/head -> origin/gh/fffrog/160/head 2025-12-04T09:20:35.3842566Z * [new branch] gh/fffrog/177/base -> origin/gh/fffrog/177/base 2025-12-04T09:20:35.3844386Z * [new branch] gh/fffrog/177/head -> origin/gh/fffrog/177/head 2025-12-04T09:20:35.3846252Z * [new branch] gh/fffrog/177/orig -> origin/gh/fffrog/177/orig 2025-12-04T09:20:35.3848519Z * [new branch] gh/fffrog/178/base -> origin/gh/fffrog/178/base 2025-12-04T09:20:35.3850228Z * [new branch] gh/fffrog/178/head -> origin/gh/fffrog/178/head 2025-12-04T09:20:35.3852025Z * [new branch] gh/fffrog/178/orig -> origin/gh/fffrog/178/orig 2025-12-04T09:20:35.3854359Z * [new branch] gh/fffrog/181/base -> origin/gh/fffrog/181/base 2025-12-04T09:20:35.3856121Z * [new branch] gh/fffrog/181/head -> origin/gh/fffrog/181/head 2025-12-04T09:20:35.3857949Z * [new branch] gh/fffrog/181/orig -> origin/gh/fffrog/181/orig 2025-12-04T09:20:35.3860575Z * [new branch] gh/fffrog/183/base -> origin/gh/fffrog/183/base 2025-12-04T09:20:35.3862157Z * [new branch] gh/fffrog/183/head -> origin/gh/fffrog/183/head 2025-12-04T09:20:35.3863849Z * [new branch] gh/fffrog/183/orig -> origin/gh/fffrog/183/orig 2025-12-04T09:20:35.3867158Z * [new branch] gh/fxdawnn/10/base -> origin/gh/fxdawnn/10/base 2025-12-04T09:20:35.3868523Z * [new branch] gh/fxdawnn/10/head -> origin/gh/fxdawnn/10/head 2025-12-04T09:20:35.3870290Z * [new branch] gh/fxdawnn/10/orig -> origin/gh/fxdawnn/10/orig 2025-12-04T09:20:35.3874597Z * [new branch] gh/fxdawnn/11/base -> origin/gh/fxdawnn/11/base 2025-12-04T09:20:35.3876107Z * [new branch] gh/fxdawnn/11/head -> origin/gh/fxdawnn/11/head 2025-12-04T09:20:35.3877798Z * [new branch] gh/fxdawnn/11/orig -> origin/gh/fxdawnn/11/orig 2025-12-04T09:20:35.3880129Z * [new branch] gh/fxdawnn/12/base -> origin/gh/fxdawnn/12/base 2025-12-04T09:20:35.3882148Z * [new branch] gh/fxdawnn/12/head -> origin/gh/fxdawnn/12/head 2025-12-04T09:20:35.3883876Z * [new branch] gh/fxdawnn/12/orig -> origin/gh/fxdawnn/12/orig 2025-12-04T09:20:35.3886196Z * [new branch] gh/fxdawnn/13/base -> origin/gh/fxdawnn/13/base 2025-12-04T09:20:35.3888024Z * [new branch] gh/fxdawnn/13/head -> origin/gh/fxdawnn/13/head 2025-12-04T09:20:35.3889763Z * [new branch] gh/fxdawnn/13/orig -> origin/gh/fxdawnn/13/orig 2025-12-04T09:20:35.3892219Z * [new branch] gh/fxdawnn/14/base -> origin/gh/fxdawnn/14/base 2025-12-04T09:20:35.3893907Z * [new branch] gh/fxdawnn/14/head -> origin/gh/fxdawnn/14/head 2025-12-04T09:20:35.3895594Z * [new branch] gh/fxdawnn/14/orig -> origin/gh/fxdawnn/14/orig 2025-12-04T09:20:35.3897941Z * [new branch] gh/fxdawnn/15/base -> origin/gh/fxdawnn/15/base 2025-12-04T09:20:35.3899849Z * [new branch] gh/fxdawnn/15/head -> origin/gh/fxdawnn/15/head 2025-12-04T09:20:35.3901635Z * [new branch] gh/fxdawnn/15/orig -> origin/gh/fxdawnn/15/orig 2025-12-04T09:20:35.3904420Z * [new branch] gh/fxdawnn/6/base -> origin/gh/fxdawnn/6/base 2025-12-04T09:20:35.3906172Z * [new branch] gh/fxdawnn/6/head -> origin/gh/fxdawnn/6/head 2025-12-04T09:20:35.3907953Z * [new branch] gh/fxdawnn/6/orig -> origin/gh/fxdawnn/6/orig 2025-12-04T09:20:35.3911190Z * [new branch] gh/fxdawnn/7/base -> origin/gh/fxdawnn/7/base 2025-12-04T09:20:35.3912763Z * [new branch] gh/fxdawnn/7/head -> origin/gh/fxdawnn/7/head 2025-12-04T09:20:35.3914442Z * [new branch] gh/fxdawnn/7/orig -> origin/gh/fxdawnn/7/orig 2025-12-04T09:20:35.3916922Z * [new branch] gh/fxdawnn/9/base -> origin/gh/fxdawnn/9/base 2025-12-04T09:20:35.3918590Z * [new branch] gh/fxdawnn/9/head -> origin/gh/fxdawnn/9/head 2025-12-04T09:20:35.3920309Z * [new branch] gh/fxdawnn/9/orig -> origin/gh/fxdawnn/9/orig 2025-12-04T09:20:35.3923091Z * [new branch] gh/galv/1/base -> origin/gh/galv/1/base 2025-12-04T09:20:35.3924827Z * [new branch] gh/galv/1/head -> origin/gh/galv/1/head 2025-12-04T09:20:35.3926642Z * [new branch] gh/galv/1/orig -> origin/gh/galv/1/orig 2025-12-04T09:20:35.3929063Z * [new branch] gh/galv/2/base -> origin/gh/galv/2/base 2025-12-04T09:20:35.3930723Z * [new branch] gh/galv/2/head -> origin/gh/galv/2/head 2025-12-04T09:20:35.3932509Z * [new branch] gh/galv/2/orig -> origin/gh/galv/2/orig 2025-12-04T09:20:35.3934961Z * [new branch] gh/galv/3/base -> origin/gh/galv/3/base 2025-12-04T09:20:35.3936597Z * [new branch] gh/galv/3/head -> origin/gh/galv/3/head 2025-12-04T09:20:35.3938402Z * [new branch] gh/galv/3/orig -> origin/gh/galv/3/orig 2025-12-04T09:20:35.3941988Z * [new branch] gh/guangyey/134/base -> origin/gh/guangyey/134/base 2025-12-04T09:20:35.3943782Z * [new branch] gh/guangyey/134/head -> origin/gh/guangyey/134/head 2025-12-04T09:20:35.3945456Z * [new branch] gh/guangyey/134/orig -> origin/gh/guangyey/134/orig 2025-12-04T09:20:35.3947765Z * [new branch] gh/guangyey/163/base -> origin/gh/guangyey/163/base 2025-12-04T09:20:35.3949513Z * [new branch] gh/guangyey/163/head -> origin/gh/guangyey/163/head 2025-12-04T09:20:35.3951317Z * [new branch] gh/guangyey/163/orig -> origin/gh/guangyey/163/orig 2025-12-04T09:20:35.3953570Z * [new branch] gh/guangyey/168/base -> origin/gh/guangyey/168/base 2025-12-04T09:20:35.3955287Z * [new branch] gh/guangyey/168/head -> origin/gh/guangyey/168/head 2025-12-04T09:20:35.3957081Z * [new branch] gh/guangyey/168/orig -> origin/gh/guangyey/168/orig 2025-12-04T09:20:35.3959358Z * [new branch] gh/guangyey/169/base -> origin/gh/guangyey/169/base 2025-12-04T09:20:35.3961140Z * [new branch] gh/guangyey/169/head -> origin/gh/guangyey/169/head 2025-12-04T09:20:35.3962933Z * [new branch] gh/guangyey/169/orig -> origin/gh/guangyey/169/orig 2025-12-04T09:20:35.3965296Z * [new branch] gh/guangyey/170/base -> origin/gh/guangyey/170/base 2025-12-04T09:20:35.3967031Z * [new branch] gh/guangyey/170/head -> origin/gh/guangyey/170/head 2025-12-04T09:20:35.3968772Z * [new branch] gh/guangyey/170/orig -> origin/gh/guangyey/170/orig 2025-12-04T09:20:35.3971304Z * [new branch] gh/guangyey/171/base -> origin/gh/guangyey/171/base 2025-12-04T09:20:35.3973215Z * [new branch] gh/guangyey/171/head -> origin/gh/guangyey/171/head 2025-12-04T09:20:35.3974951Z * [new branch] gh/guangyey/171/orig -> origin/gh/guangyey/171/orig 2025-12-04T09:20:35.3977309Z * [new branch] gh/guangyey/178/base -> origin/gh/guangyey/178/base 2025-12-04T09:20:35.3979025Z * [new branch] gh/guangyey/178/head -> origin/gh/guangyey/178/head 2025-12-04T09:20:35.3980925Z * [new branch] gh/guangyey/178/orig -> origin/gh/guangyey/178/orig 2025-12-04T09:20:35.3983155Z * [new branch] gh/guangyey/182/base -> origin/gh/guangyey/182/base 2025-12-04T09:20:35.3984935Z * [new branch] gh/guangyey/182/head -> origin/gh/guangyey/182/head 2025-12-04T09:20:35.3986685Z * [new branch] gh/guangyey/182/orig -> origin/gh/guangyey/182/orig 2025-12-04T09:20:35.3988931Z * [new branch] gh/guangyey/183/base -> origin/gh/guangyey/183/base 2025-12-04T09:20:35.3990672Z * [new branch] gh/guangyey/183/head -> origin/gh/guangyey/183/head 2025-12-04T09:20:35.3992544Z * [new branch] gh/guangyey/183/orig -> origin/gh/guangyey/183/orig 2025-12-04T09:20:35.3994905Z * [new branch] gh/guangyey/185/base -> origin/gh/guangyey/185/base 2025-12-04T09:20:35.3996635Z * [new branch] gh/guangyey/185/head -> origin/gh/guangyey/185/head 2025-12-04T09:20:35.3998367Z * [new branch] gh/guangyey/185/orig -> origin/gh/guangyey/185/orig 2025-12-04T09:20:35.4000730Z * [new branch] gh/guangyey/186/base -> origin/gh/guangyey/186/base 2025-12-04T09:20:35.4002491Z * [new branch] gh/guangyey/186/head -> origin/gh/guangyey/186/head 2025-12-04T09:20:35.4004423Z * [new branch] gh/guangyey/186/orig -> origin/gh/guangyey/186/orig 2025-12-04T09:20:35.4006672Z * [new branch] gh/guangyey/187/base -> origin/gh/guangyey/187/base 2025-12-04T09:20:35.4008394Z * [new branch] gh/guangyey/187/head -> origin/gh/guangyey/187/head 2025-12-04T09:20:35.4010198Z * [new branch] gh/guangyey/187/orig -> origin/gh/guangyey/187/orig 2025-12-04T09:20:35.4012458Z * [new branch] gh/guangyey/188/base -> origin/gh/guangyey/188/base 2025-12-04T09:20:35.4014168Z * [new branch] gh/guangyey/188/head -> origin/gh/guangyey/188/head 2025-12-04T09:20:35.4015900Z * [new branch] gh/guangyey/188/orig -> origin/gh/guangyey/188/orig 2025-12-04T09:20:35.4018304Z * [new branch] gh/guangyey/190/base -> origin/gh/guangyey/190/base 2025-12-04T09:20:35.4020066Z * [new branch] gh/guangyey/190/head -> origin/gh/guangyey/190/head 2025-12-04T09:20:35.4021725Z * [new branch] gh/guangyey/190/orig -> origin/gh/guangyey/190/orig 2025-12-04T09:20:35.4024059Z * [new branch] gh/guangyey/208/base -> origin/gh/guangyey/208/base 2025-12-04T09:20:35.4025799Z * [new branch] gh/guangyey/208/head -> origin/gh/guangyey/208/head 2025-12-04T09:20:35.4027531Z * [new branch] gh/guangyey/208/orig -> origin/gh/guangyey/208/orig 2025-12-04T09:20:35.4029910Z * [new branch] gh/guangyey/228/base -> origin/gh/guangyey/228/base 2025-12-04T09:20:35.4031775Z * [new branch] gh/guangyey/228/head -> origin/gh/guangyey/228/head 2025-12-04T09:20:35.4033557Z * [new branch] gh/guangyey/228/orig -> origin/gh/guangyey/228/orig 2025-12-04T09:20:35.4036384Z * [new branch] gh/guangyey/230/base -> origin/gh/guangyey/230/base 2025-12-04T09:20:35.4038134Z * [new branch] gh/guangyey/230/head -> origin/gh/guangyey/230/head 2025-12-04T09:20:35.4039870Z * [new branch] gh/guangyey/230/orig -> origin/gh/guangyey/230/orig 2025-12-04T09:20:35.4042231Z * [new branch] gh/guangyey/231/base -> origin/gh/guangyey/231/base 2025-12-04T09:20:35.4044245Z * [new branch] gh/guangyey/231/head -> origin/gh/guangyey/231/head 2025-12-04T09:20:35.4046013Z * [new branch] gh/guangyey/231/orig -> origin/gh/guangyey/231/orig 2025-12-04T09:20:35.4048350Z * [new branch] gh/guangyey/232/base -> origin/gh/guangyey/232/base 2025-12-04T09:20:35.4050070Z * [new branch] gh/guangyey/232/head -> origin/gh/guangyey/232/head 2025-12-04T09:20:35.4051873Z * [new branch] gh/guangyey/232/orig -> origin/gh/guangyey/232/orig 2025-12-04T09:20:35.4054338Z * [new branch] gh/guangyey/233/base -> origin/gh/guangyey/233/base 2025-12-04T09:20:35.4056036Z * [new branch] gh/guangyey/233/head -> origin/gh/guangyey/233/head 2025-12-04T09:20:35.4057900Z * [new branch] gh/guangyey/233/orig -> origin/gh/guangyey/233/orig 2025-12-04T09:20:35.4060502Z * [new branch] gh/guangyey/234/base -> origin/gh/guangyey/234/base 2025-12-04T09:20:35.4062111Z * [new branch] gh/guangyey/234/head -> origin/gh/guangyey/234/head 2025-12-04T09:20:35.4063829Z * [new branch] gh/guangyey/234/orig -> origin/gh/guangyey/234/orig 2025-12-04T09:20:35.4066281Z * [new branch] gh/guangyey/235/base -> origin/gh/guangyey/235/base 2025-12-04T09:20:35.4067992Z * [new branch] gh/guangyey/235/head -> origin/gh/guangyey/235/head 2025-12-04T09:20:35.4069686Z * [new branch] gh/guangyey/235/orig -> origin/gh/guangyey/235/orig 2025-12-04T09:20:35.4072219Z * [new branch] gh/guangyey/236/base -> origin/gh/guangyey/236/base 2025-12-04T09:20:35.4074398Z * [new branch] gh/guangyey/236/head -> origin/gh/guangyey/236/head 2025-12-04T09:20:35.4075972Z * [new branch] gh/guangyey/236/orig -> origin/gh/guangyey/236/orig 2025-12-04T09:20:35.4078339Z * [new branch] gh/guangyey/237/base -> origin/gh/guangyey/237/base 2025-12-04T09:20:35.4080083Z * [new branch] gh/guangyey/237/head -> origin/gh/guangyey/237/head 2025-12-04T09:20:35.4081872Z * [new branch] gh/guangyey/237/orig -> origin/gh/guangyey/237/orig 2025-12-04T09:20:35.4084392Z * [new branch] gh/guangyey/238/base -> origin/gh/guangyey/238/base 2025-12-04T09:20:35.4086171Z * [new branch] gh/guangyey/238/head -> origin/gh/guangyey/238/head 2025-12-04T09:20:35.4088584Z * [new branch] gh/guangyey/239/base -> origin/gh/guangyey/239/base 2025-12-04T09:20:35.4090319Z * [new branch] gh/guangyey/239/head -> origin/gh/guangyey/239/head 2025-12-04T09:20:35.4091960Z * [new branch] gh/guangyey/239/orig -> origin/gh/guangyey/239/orig 2025-12-04T09:20:35.4094411Z * [new branch] gh/guangyey/240/base -> origin/gh/guangyey/240/base 2025-12-04T09:20:35.4096123Z * [new branch] gh/guangyey/240/head -> origin/gh/guangyey/240/head 2025-12-04T09:20:35.4097860Z * [new branch] gh/guangyey/240/orig -> origin/gh/guangyey/240/orig 2025-12-04T09:20:35.4100845Z * [new branch] gh/guangyey/241/base -> origin/gh/guangyey/241/base 2025-12-04T09:20:35.4102541Z * [new branch] gh/guangyey/241/head -> origin/gh/guangyey/241/head 2025-12-04T09:20:35.4104223Z * [new branch] gh/guangyey/241/orig -> origin/gh/guangyey/241/orig 2025-12-04T09:20:35.4106712Z * [new branch] gh/guangyey/242/base -> origin/gh/guangyey/242/base 2025-12-04T09:20:35.4108495Z * [new branch] gh/guangyey/242/head -> origin/gh/guangyey/242/head 2025-12-04T09:20:35.4110169Z * [new branch] gh/guangyey/242/orig -> origin/gh/guangyey/242/orig 2025-12-04T09:20:35.4112620Z * [new branch] gh/guangyey/243/base -> origin/gh/guangyey/243/base 2025-12-04T09:20:35.4114327Z * [new branch] gh/guangyey/243/head -> origin/gh/guangyey/243/head 2025-12-04T09:20:35.4116055Z * [new branch] gh/guangyey/243/orig -> origin/gh/guangyey/243/orig 2025-12-04T09:20:35.4118488Z * [new branch] gh/guangyey/244/base -> origin/gh/guangyey/244/base 2025-12-04T09:20:35.4120323Z * [new branch] gh/guangyey/244/head -> origin/gh/guangyey/244/head 2025-12-04T09:20:35.4122178Z * [new branch] gh/guangyey/244/orig -> origin/gh/guangyey/244/orig 2025-12-04T09:20:35.4124498Z * [new branch] gh/guangyey/245/base -> origin/gh/guangyey/245/base 2025-12-04T09:20:35.4126243Z * [new branch] gh/guangyey/245/head -> origin/gh/guangyey/245/head 2025-12-04T09:20:35.4127943Z * [new branch] gh/guangyey/245/orig -> origin/gh/guangyey/245/orig 2025-12-04T09:20:35.4130390Z * [new branch] gh/guangyey/246/base -> origin/gh/guangyey/246/base 2025-12-04T09:20:35.4132115Z * [new branch] gh/guangyey/246/head -> origin/gh/guangyey/246/head 2025-12-04T09:20:35.4133951Z * [new branch] gh/guangyey/246/orig -> origin/gh/guangyey/246/orig 2025-12-04T09:20:35.4136412Z * [new branch] gh/guangyey/247/base -> origin/gh/guangyey/247/base 2025-12-04T09:20:35.4138169Z * [new branch] gh/guangyey/247/head -> origin/gh/guangyey/247/head 2025-12-04T09:20:35.4140026Z * [new branch] gh/guangyey/247/orig -> origin/gh/guangyey/247/orig 2025-12-04T09:20:35.4142819Z * [new branch] gh/guangyey/248/base -> origin/gh/guangyey/248/base 2025-12-04T09:20:35.4144684Z * [new branch] gh/guangyey/248/head -> origin/gh/guangyey/248/head 2025-12-04T09:20:35.4146336Z * [new branch] gh/guangyey/248/orig -> origin/gh/guangyey/248/orig 2025-12-04T09:20:35.4148735Z * [new branch] gh/guangyey/249/base -> origin/gh/guangyey/249/base 2025-12-04T09:20:35.4150768Z * [new branch] gh/guangyey/249/head -> origin/gh/guangyey/249/head 2025-12-04T09:20:35.4152549Z * [new branch] gh/guangyey/249/orig -> origin/gh/guangyey/249/orig 2025-12-04T09:20:35.4155043Z * [new branch] gh/guangyey/250/base -> origin/gh/guangyey/250/base 2025-12-04T09:20:35.4156778Z * [new branch] gh/guangyey/250/head -> origin/gh/guangyey/250/head 2025-12-04T09:20:35.4158593Z * [new branch] gh/guangyey/250/orig -> origin/gh/guangyey/250/orig 2025-12-04T09:20:35.4160863Z * [new branch] gh/guangyey/251/base -> origin/gh/guangyey/251/base 2025-12-04T09:20:35.4162605Z * [new branch] gh/guangyey/251/head -> origin/gh/guangyey/251/head 2025-12-04T09:20:35.4164370Z * [new branch] gh/guangyey/251/orig -> origin/gh/guangyey/251/orig 2025-12-04T09:20:35.4166751Z * [new branch] gh/guangyey/252/base -> origin/gh/guangyey/252/base 2025-12-04T09:20:35.4168559Z * [new branch] gh/guangyey/252/head -> origin/gh/guangyey/252/head 2025-12-04T09:20:35.4170221Z * [new branch] gh/guangyey/252/orig -> origin/gh/guangyey/252/orig 2025-12-04T09:20:35.4174155Z * [new branch] gh/guangyey/253/base -> origin/gh/guangyey/253/base 2025-12-04T09:20:35.4175926Z * [new branch] gh/guangyey/253/head -> origin/gh/guangyey/253/head 2025-12-04T09:20:35.4177581Z * [new branch] gh/guangyey/253/orig -> origin/gh/guangyey/253/orig 2025-12-04T09:20:35.4180172Z * [new branch] gh/guangyey/254/base -> origin/gh/guangyey/254/base 2025-12-04T09:20:35.4181971Z * [new branch] gh/guangyey/254/head -> origin/gh/guangyey/254/head 2025-12-04T09:20:35.4183719Z * [new branch] gh/guangyey/254/orig -> origin/gh/guangyey/254/orig 2025-12-04T09:20:35.4186069Z * [new branch] gh/guangyey/255/base -> origin/gh/guangyey/255/base 2025-12-04T09:20:35.4187786Z * [new branch] gh/guangyey/255/head -> origin/gh/guangyey/255/head 2025-12-04T09:20:35.4189458Z * [new branch] gh/guangyey/255/orig -> origin/gh/guangyey/255/orig 2025-12-04T09:20:35.4192542Z * [new branch] gh/guilhermeleobas/107/base -> origin/gh/guilhermeleobas/107/base 2025-12-04T09:20:35.4194290Z * [new branch] gh/guilhermeleobas/107/head -> origin/gh/guilhermeleobas/107/head 2025-12-04T09:20:35.4196026Z * [new branch] gh/guilhermeleobas/107/orig -> origin/gh/guilhermeleobas/107/orig 2025-12-04T09:20:35.4198441Z * [new branch] gh/guilhermeleobas/108/base -> origin/gh/guilhermeleobas/108/base 2025-12-04T09:20:35.4200071Z * [new branch] gh/guilhermeleobas/108/head -> origin/gh/guilhermeleobas/108/head 2025-12-04T09:20:35.4201781Z * [new branch] gh/guilhermeleobas/108/orig -> origin/gh/guilhermeleobas/108/orig 2025-12-04T09:20:35.4204208Z * [new branch] gh/guilhermeleobas/150/base -> origin/gh/guilhermeleobas/150/base 2025-12-04T09:20:35.4206933Z * [new branch] gh/guilhermeleobas/150/head -> origin/gh/guilhermeleobas/150/head 2025-12-04T09:20:35.4208744Z * [new branch] gh/guilhermeleobas/150/orig -> origin/gh/guilhermeleobas/150/orig 2025-12-04T09:20:35.4211016Z * [new branch] gh/guilhermeleobas/168/base -> origin/gh/guilhermeleobas/168/base 2025-12-04T09:20:35.4212674Z * [new branch] gh/guilhermeleobas/168/head -> origin/gh/guilhermeleobas/168/head 2025-12-04T09:20:35.4214695Z * [new branch] gh/guilhermeleobas/168/orig -> origin/gh/guilhermeleobas/168/orig 2025-12-04T09:20:35.4216947Z * [new branch] gh/guilhermeleobas/169/base -> origin/gh/guilhermeleobas/169/base 2025-12-04T09:20:35.4218576Z * [new branch] gh/guilhermeleobas/169/head -> origin/gh/guilhermeleobas/169/head 2025-12-04T09:20:35.4220428Z * [new branch] gh/guilhermeleobas/169/orig -> origin/gh/guilhermeleobas/169/orig 2025-12-04T09:20:35.4222820Z * [new branch] gh/guilhermeleobas/170/base -> origin/gh/guilhermeleobas/170/base 2025-12-04T09:20:35.4224606Z * [new branch] gh/guilhermeleobas/170/head -> origin/gh/guilhermeleobas/170/head 2025-12-04T09:20:35.4226433Z * [new branch] gh/guilhermeleobas/170/orig -> origin/gh/guilhermeleobas/170/orig 2025-12-04T09:20:35.4228803Z * [new branch] gh/guilhermeleobas/171/base -> origin/gh/guilhermeleobas/171/base 2025-12-04T09:20:35.4230516Z * [new branch] gh/guilhermeleobas/171/head -> origin/gh/guilhermeleobas/171/head 2025-12-04T09:20:35.4232311Z * [new branch] gh/guilhermeleobas/171/orig -> origin/gh/guilhermeleobas/171/orig 2025-12-04T09:20:35.4234656Z * [new branch] gh/guilhermeleobas/173/base -> origin/gh/guilhermeleobas/173/base 2025-12-04T09:20:35.4236362Z * [new branch] gh/guilhermeleobas/173/head -> origin/gh/guilhermeleobas/173/head 2025-12-04T09:20:35.4238054Z * [new branch] gh/guilhermeleobas/173/orig -> origin/gh/guilhermeleobas/173/orig 2025-12-04T09:20:35.4240436Z * [new branch] gh/guilhermeleobas/193/base -> origin/gh/guilhermeleobas/193/base 2025-12-04T09:20:35.4242220Z * [new branch] gh/guilhermeleobas/193/head -> origin/gh/guilhermeleobas/193/head 2025-12-04T09:20:35.4244136Z * [new branch] gh/guilhermeleobas/193/orig -> origin/gh/guilhermeleobas/193/orig 2025-12-04T09:20:35.4246432Z * [new branch] gh/guilhermeleobas/204/base -> origin/gh/guilhermeleobas/204/base 2025-12-04T09:20:35.4248217Z * [new branch] gh/guilhermeleobas/204/head -> origin/gh/guilhermeleobas/204/head 2025-12-04T09:20:35.4250013Z * [new branch] gh/guilhermeleobas/204/orig -> origin/gh/guilhermeleobas/204/orig 2025-12-04T09:20:35.4252373Z * [new branch] gh/guilhermeleobas/211/base -> origin/gh/guilhermeleobas/211/base 2025-12-04T09:20:35.4254236Z * [new branch] gh/guilhermeleobas/211/head -> origin/gh/guilhermeleobas/211/head 2025-12-04T09:20:35.4255966Z * [new branch] gh/guilhermeleobas/211/orig -> origin/gh/guilhermeleobas/211/orig 2025-12-04T09:20:35.4258263Z * [new branch] gh/guilhermeleobas/226/base -> origin/gh/guilhermeleobas/226/base 2025-12-04T09:20:35.4260101Z * [new branch] gh/guilhermeleobas/226/head -> origin/gh/guilhermeleobas/226/head 2025-12-04T09:20:35.4261779Z * [new branch] gh/guilhermeleobas/226/orig -> origin/gh/guilhermeleobas/226/orig 2025-12-04T09:20:35.4264176Z * [new branch] gh/guilhermeleobas/236/base -> origin/gh/guilhermeleobas/236/base 2025-12-04T09:20:35.4265931Z * [new branch] gh/guilhermeleobas/236/head -> origin/gh/guilhermeleobas/236/head 2025-12-04T09:20:35.4267709Z * [new branch] gh/guilhermeleobas/236/orig -> origin/gh/guilhermeleobas/236/orig 2025-12-04T09:20:35.4270002Z * [new branch] gh/guilhermeleobas/247/base -> origin/gh/guilhermeleobas/247/base 2025-12-04T09:20:35.4271894Z * [new branch] gh/guilhermeleobas/247/head -> origin/gh/guilhermeleobas/247/head 2025-12-04T09:20:35.4273711Z * [new branch] gh/guilhermeleobas/247/orig -> origin/gh/guilhermeleobas/247/orig 2025-12-04T09:20:35.4276162Z * [new branch] gh/guilhermeleobas/248/base -> origin/gh/guilhermeleobas/248/base 2025-12-04T09:20:35.4277879Z * [new branch] gh/guilhermeleobas/248/head -> origin/gh/guilhermeleobas/248/head 2025-12-04T09:20:35.4279667Z * [new branch] gh/guilhermeleobas/248/orig -> origin/gh/guilhermeleobas/248/orig 2025-12-04T09:20:35.4284071Z * [new branch] gh/guilhermeleobas/250/base -> origin/gh/guilhermeleobas/250/base 2025-12-04T09:20:35.4284771Z * [new branch] gh/guilhermeleobas/250/head -> origin/gh/guilhermeleobas/250/head 2025-12-04T09:20:35.4285218Z * [new branch] gh/guilhermeleobas/250/orig -> origin/gh/guilhermeleobas/250/orig 2025-12-04T09:20:35.4288328Z * [new branch] gh/guilhermeleobas/253/base -> origin/gh/guilhermeleobas/253/base 2025-12-04T09:20:35.4290046Z * [new branch] gh/guilhermeleobas/253/head -> origin/gh/guilhermeleobas/253/head 2025-12-04T09:20:35.4291816Z * [new branch] gh/guilhermeleobas/253/orig -> origin/gh/guilhermeleobas/253/orig 2025-12-04T09:20:35.4294256Z * [new branch] gh/guilhermeleobas/254/base -> origin/gh/guilhermeleobas/254/base 2025-12-04T09:20:35.4296081Z * [new branch] gh/guilhermeleobas/254/head -> origin/gh/guilhermeleobas/254/head 2025-12-04T09:20:35.4297879Z * [new branch] gh/guilhermeleobas/254/orig -> origin/gh/guilhermeleobas/254/orig 2025-12-04T09:20:35.4300517Z * [new branch] gh/guilhermeleobas/255/base -> origin/gh/guilhermeleobas/255/base 2025-12-04T09:20:35.4302350Z * [new branch] gh/guilhermeleobas/255/head -> origin/gh/guilhermeleobas/255/head 2025-12-04T09:20:35.4304129Z * [new branch] gh/guilhermeleobas/255/orig -> origin/gh/guilhermeleobas/255/orig 2025-12-04T09:20:35.4306818Z * [new branch] gh/guilhermeleobas/256/base -> origin/gh/guilhermeleobas/256/base 2025-12-04T09:20:35.4308257Z * [new branch] gh/guilhermeleobas/256/head -> origin/gh/guilhermeleobas/256/head 2025-12-04T09:20:35.4309989Z * [new branch] gh/guilhermeleobas/256/orig -> origin/gh/guilhermeleobas/256/orig 2025-12-04T09:20:35.4312299Z * [new branch] gh/guilhermeleobas/257/base -> origin/gh/guilhermeleobas/257/base 2025-12-04T09:20:35.4314017Z * [new branch] gh/guilhermeleobas/257/head -> origin/gh/guilhermeleobas/257/head 2025-12-04T09:20:35.4315858Z * [new branch] gh/guilhermeleobas/257/orig -> origin/gh/guilhermeleobas/257/orig 2025-12-04T09:20:35.4318272Z * [new branch] gh/guilhermeleobas/258/base -> origin/gh/guilhermeleobas/258/base 2025-12-04T09:20:35.4320017Z * [new branch] gh/guilhermeleobas/258/head -> origin/gh/guilhermeleobas/258/head 2025-12-04T09:20:35.4322050Z * [new branch] gh/guilhermeleobas/258/orig -> origin/gh/guilhermeleobas/258/orig 2025-12-04T09:20:35.4324421Z * [new branch] gh/guilhermeleobas/259/base -> origin/gh/guilhermeleobas/259/base 2025-12-04T09:20:35.4326196Z * [new branch] gh/guilhermeleobas/259/head -> origin/gh/guilhermeleobas/259/head 2025-12-04T09:20:35.4327908Z * [new branch] gh/guilhermeleobas/259/orig -> origin/gh/guilhermeleobas/259/orig 2025-12-04T09:20:35.4330333Z * [new branch] gh/guilhermeleobas/260/base -> origin/gh/guilhermeleobas/260/base 2025-12-04T09:20:35.4331984Z * [new branch] gh/guilhermeleobas/260/head -> origin/gh/guilhermeleobas/260/head 2025-12-04T09:20:35.4333733Z * [new branch] gh/guilhermeleobas/260/orig -> origin/gh/guilhermeleobas/260/orig 2025-12-04T09:20:35.4336111Z * [new branch] gh/guilhermeleobas/261/base -> origin/gh/guilhermeleobas/261/base 2025-12-04T09:20:35.4337934Z * [new branch] gh/guilhermeleobas/261/head -> origin/gh/guilhermeleobas/261/head 2025-12-04T09:20:35.4339762Z * [new branch] gh/guilhermeleobas/261/orig -> origin/gh/guilhermeleobas/261/orig 2025-12-04T09:20:35.4342175Z * [new branch] gh/guilhermeleobas/262/base -> origin/gh/guilhermeleobas/262/base 2025-12-04T09:20:35.4344359Z * [new branch] gh/guilhermeleobas/262/head -> origin/gh/guilhermeleobas/262/head 2025-12-04T09:20:35.4345756Z * [new branch] gh/guilhermeleobas/262/orig -> origin/gh/guilhermeleobas/262/orig 2025-12-04T09:20:35.4348255Z * [new branch] gh/guilhermeleobas/263/base -> origin/gh/guilhermeleobas/263/base 2025-12-04T09:20:35.4349907Z * [new branch] gh/guilhermeleobas/263/head -> origin/gh/guilhermeleobas/263/head 2025-12-04T09:20:35.4351594Z * [new branch] gh/guilhermeleobas/263/orig -> origin/gh/guilhermeleobas/263/orig 2025-12-04T09:20:35.4354093Z * [new branch] gh/guilhermeleobas/264/base -> origin/gh/guilhermeleobas/264/base 2025-12-04T09:20:35.4355824Z * [new branch] gh/guilhermeleobas/264/head -> origin/gh/guilhermeleobas/264/head 2025-12-04T09:20:35.4357588Z * [new branch] gh/guilhermeleobas/264/orig -> origin/gh/guilhermeleobas/264/orig 2025-12-04T09:20:35.4360135Z * [new branch] gh/guilhermeleobas/265/base -> origin/gh/guilhermeleobas/265/base 2025-12-04T09:20:35.4361915Z * [new branch] gh/guilhermeleobas/265/head -> origin/gh/guilhermeleobas/265/head 2025-12-04T09:20:35.4363694Z * [new branch] gh/guilhermeleobas/265/orig -> origin/gh/guilhermeleobas/265/orig 2025-12-04T09:20:35.4366137Z * [new branch] gh/guilhermeleobas/266/base -> origin/gh/guilhermeleobas/266/base 2025-12-04T09:20:35.4367888Z * [new branch] gh/guilhermeleobas/266/head -> origin/gh/guilhermeleobas/266/head 2025-12-04T09:20:35.4369676Z * [new branch] gh/guilhermeleobas/266/orig -> origin/gh/guilhermeleobas/266/orig 2025-12-04T09:20:35.4372189Z * [new branch] gh/guilhermeleobas/267/base -> origin/gh/guilhermeleobas/267/base 2025-12-04T09:20:35.4374046Z * [new branch] gh/guilhermeleobas/267/head -> origin/gh/guilhermeleobas/267/head 2025-12-04T09:20:35.4375791Z * [new branch] gh/guilhermeleobas/267/orig -> origin/gh/guilhermeleobas/267/orig 2025-12-04T09:20:35.4378678Z * [new branch] gh/hameerabbasi/1/base -> origin/gh/hameerabbasi/1/base 2025-12-04T09:20:35.4380642Z * [new branch] gh/hameerabbasi/1/head -> origin/gh/hameerabbasi/1/head 2025-12-04T09:20:35.4383211Z * [new branch] gh/hameerabbasi/2/base -> origin/gh/hameerabbasi/2/base 2025-12-04T09:20:35.4384941Z * [new branch] gh/hameerabbasi/2/head -> origin/gh/hameerabbasi/2/head 2025-12-04T09:20:35.4386711Z * [new branch] gh/hameerabbasi/2/orig -> origin/gh/hameerabbasi/2/orig 2025-12-04T09:20:35.4389078Z * [new branch] gh/hameerabbasi/3/base -> origin/gh/hameerabbasi/3/base 2025-12-04T09:20:35.4390790Z * [new branch] gh/hameerabbasi/3/head -> origin/gh/hameerabbasi/3/head 2025-12-04T09:20:35.4392780Z * [new branch] gh/hameerabbasi/3/orig -> origin/gh/hameerabbasi/3/orig 2025-12-04T09:20:35.4395103Z * [new branch] gh/hameerabbasi/4/base -> origin/gh/hameerabbasi/4/base 2025-12-04T09:20:35.4396848Z * [new branch] gh/hameerabbasi/4/head -> origin/gh/hameerabbasi/4/head 2025-12-04T09:20:35.4398499Z * [new branch] gh/hameerabbasi/4/orig -> origin/gh/hameerabbasi/4/orig 2025-12-04T09:20:35.4401428Z * [new branch] gh/huydhn/1/next -> origin/gh/huydhn/1/next 2025-12-04T09:20:35.4404261Z * [new branch] gh/huydhn/2/next -> origin/gh/huydhn/2/next 2025-12-04T09:20:35.4406607Z * [new branch] gh/huydhn/3/next -> origin/gh/huydhn/3/next 2025-12-04T09:20:35.4408918Z * [new branch] gh/huydhn/4/next -> origin/gh/huydhn/4/next 2025-12-04T09:20:35.4411224Z * [new branch] gh/huydhn/5/next -> origin/gh/huydhn/5/next 2025-12-04T09:20:35.4413570Z * [new branch] gh/huydhn/6/next -> origin/gh/huydhn/6/next 2025-12-04T09:20:35.4416327Z * [new branch] gh/int3/97/base -> origin/gh/int3/97/base 2025-12-04T09:20:35.4418040Z * [new branch] gh/int3/97/head -> origin/gh/int3/97/head 2025-12-04T09:20:35.4421241Z * [new branch] gh/isuruf/101/base -> origin/gh/isuruf/101/base 2025-12-04T09:20:35.4422890Z * [new branch] gh/isuruf/101/head -> origin/gh/isuruf/101/head 2025-12-04T09:20:35.4425181Z * [new branch] gh/isuruf/146/base -> origin/gh/isuruf/146/base 2025-12-04T09:20:35.4426940Z * [new branch] gh/isuruf/146/head -> origin/gh/isuruf/146/head 2025-12-04T09:20:35.4428725Z * [new branch] gh/isuruf/146/orig -> origin/gh/isuruf/146/orig 2025-12-04T09:20:35.4431075Z * [new branch] gh/isuruf/158/base -> origin/gh/isuruf/158/base 2025-12-04T09:20:35.4432900Z * [new branch] gh/isuruf/158/head -> origin/gh/isuruf/158/head 2025-12-04T09:20:35.4435603Z * [new branch] gh/isuruf/159/base -> origin/gh/isuruf/159/base 2025-12-04T09:20:35.4437341Z * [new branch] gh/isuruf/159/head -> origin/gh/isuruf/159/head 2025-12-04T09:20:35.4439744Z * [new branch] gh/isuruf/160/base -> origin/gh/isuruf/160/base 2025-12-04T09:20:35.4441519Z * [new branch] gh/isuruf/160/head -> origin/gh/isuruf/160/head 2025-12-04T09:20:35.4443289Z * [new branch] gh/isuruf/160/orig -> origin/gh/isuruf/160/orig 2025-12-04T09:20:35.4445646Z * [new branch] gh/isuruf/81/base -> origin/gh/isuruf/81/base 2025-12-04T09:20:35.4447317Z * [new branch] gh/isuruf/81/head -> origin/gh/isuruf/81/head 2025-12-04T09:20:35.4449050Z * [new branch] gh/isuruf/81/orig -> origin/gh/isuruf/81/orig 2025-12-04T09:20:35.4451839Z * [new branch] gh/jamesjwu/176/base -> origin/gh/jamesjwu/176/base 2025-12-04T09:20:35.4453660Z * [new branch] gh/jamesjwu/176/head -> origin/gh/jamesjwu/176/head 2025-12-04T09:20:35.4455454Z * [new branch] gh/jamesjwu/176/orig -> origin/gh/jamesjwu/176/orig 2025-12-04T09:20:35.4457760Z * [new branch] gh/jamesjwu/187/base -> origin/gh/jamesjwu/187/base 2025-12-04T09:20:35.4459594Z * [new branch] gh/jamesjwu/187/head -> origin/gh/jamesjwu/187/head 2025-12-04T09:20:35.4461493Z * [new branch] gh/jamesjwu/187/orig -> origin/gh/jamesjwu/187/orig 2025-12-04T09:20:35.4463831Z * [new branch] gh/jamesjwu/196/base -> origin/gh/jamesjwu/196/base 2025-12-04T09:20:35.4465559Z * [new branch] gh/jamesjwu/196/head -> origin/gh/jamesjwu/196/head 2025-12-04T09:20:35.4467413Z * [new branch] gh/jamesjwu/196/orig -> origin/gh/jamesjwu/196/orig 2025-12-04T09:20:35.4469756Z * [new branch] gh/jamesjwu/198/base -> origin/gh/jamesjwu/198/base 2025-12-04T09:20:35.4471430Z * [new branch] gh/jamesjwu/198/head -> origin/gh/jamesjwu/198/head 2025-12-04T09:20:35.4473452Z * [new branch] gh/jamesjwu/198/orig -> origin/gh/jamesjwu/198/orig 2025-12-04T09:20:35.4475896Z * [new branch] gh/jamesjwu/207/base -> origin/gh/jamesjwu/207/base 2025-12-04T09:20:35.4477718Z * [new branch] gh/jamesjwu/207/head -> origin/gh/jamesjwu/207/head 2025-12-04T09:20:35.4479515Z * [new branch] gh/jamesjwu/207/orig -> origin/gh/jamesjwu/207/orig 2025-12-04T09:20:35.4481887Z * [new branch] gh/jamesjwu/208/base -> origin/gh/jamesjwu/208/base 2025-12-04T09:20:35.4483772Z * [new branch] gh/jamesjwu/208/head -> origin/gh/jamesjwu/208/head 2025-12-04T09:20:35.4485453Z * [new branch] gh/jamesjwu/208/orig -> origin/gh/jamesjwu/208/orig 2025-12-04T09:20:35.4487855Z * [new branch] gh/jamesjwu/52/base -> origin/gh/jamesjwu/52/base 2025-12-04T09:20:35.4489608Z * [new branch] gh/jamesjwu/52/head -> origin/gh/jamesjwu/52/head 2025-12-04T09:20:35.4491971Z * [new branch] gh/jamesjwu/53/base -> origin/gh/jamesjwu/53/base 2025-12-04T09:20:35.4493518Z * [new branch] gh/jamesjwu/53/head -> origin/gh/jamesjwu/53/head 2025-12-04T09:20:35.4495695Z * [new branch] gh/jamesjwu/54/base -> origin/gh/jamesjwu/54/base 2025-12-04T09:20:35.4497476Z * [new branch] gh/jamesjwu/54/head -> origin/gh/jamesjwu/54/head 2025-12-04T09:20:35.4499771Z * [new branch] gh/jamesjwu/55/base -> origin/gh/jamesjwu/55/base 2025-12-04T09:20:35.4501488Z * [new branch] gh/jamesjwu/55/head -> origin/gh/jamesjwu/55/head 2025-12-04T09:20:35.4503734Z * [new branch] gh/jamesjwu/56/base -> origin/gh/jamesjwu/56/base 2025-12-04T09:20:35.4505486Z * [new branch] gh/jamesjwu/56/head -> origin/gh/jamesjwu/56/head 2025-12-04T09:20:35.4507622Z * [new branch] gh/jamesjwu/57/base -> origin/gh/jamesjwu/57/base 2025-12-04T09:20:35.4509279Z * [new branch] gh/jamesjwu/57/head -> origin/gh/jamesjwu/57/head 2025-12-04T09:20:35.4511493Z * [new branch] gh/jamesjwu/58/base -> origin/gh/jamesjwu/58/base 2025-12-04T09:20:35.4513217Z * [new branch] gh/jamesjwu/58/head -> origin/gh/jamesjwu/58/head 2025-12-04T09:20:35.4516075Z * [new branch] gh/jamesjwu/59/base -> origin/gh/jamesjwu/59/base 2025-12-04T09:20:35.4517873Z * [new branch] gh/jamesjwu/59/head -> origin/gh/jamesjwu/59/head 2025-12-04T09:20:35.4520110Z * [new branch] gh/jamesjwu/60/base -> origin/gh/jamesjwu/60/base 2025-12-04T09:20:35.4521962Z * [new branch] gh/jamesjwu/60/head -> origin/gh/jamesjwu/60/head 2025-12-04T09:20:35.4524227Z * [new branch] gh/jamesjwu/61/base -> origin/gh/jamesjwu/61/base 2025-12-04T09:20:35.4525929Z * [new branch] gh/jamesjwu/61/head -> origin/gh/jamesjwu/61/head 2025-12-04T09:20:35.4528214Z * [new branch] gh/jamesjwu/62/base -> origin/gh/jamesjwu/62/base 2025-12-04T09:20:35.4529946Z * [new branch] gh/jamesjwu/62/head -> origin/gh/jamesjwu/62/head 2025-12-04T09:20:35.4531953Z * [new branch] gh/jamesjwu/63/base -> origin/gh/jamesjwu/63/base 2025-12-04T09:20:35.4533781Z * [new branch] gh/jamesjwu/63/head -> origin/gh/jamesjwu/63/head 2025-12-04T09:20:35.4536681Z * [new branch] gh/jamesjwu/64/base -> origin/gh/jamesjwu/64/base 2025-12-04T09:20:35.4538408Z * [new branch] gh/jamesjwu/64/head -> origin/gh/jamesjwu/64/head 2025-12-04T09:20:35.4540828Z * [new branch] gh/jamesjwu/65/base -> origin/gh/jamesjwu/65/base 2025-12-04T09:20:35.4542517Z * [new branch] gh/jamesjwu/65/head -> origin/gh/jamesjwu/65/head 2025-12-04T09:20:35.4545481Z * [new branch] gh/janeyx99/165/base -> origin/gh/janeyx99/165/base 2025-12-04T09:20:35.4547282Z * [new branch] gh/janeyx99/165/head -> origin/gh/janeyx99/165/head 2025-12-04T09:20:35.4549036Z * [new branch] gh/janeyx99/165/orig -> origin/gh/janeyx99/165/orig 2025-12-04T09:20:35.4551244Z * [new branch] gh/janeyx99/201/base -> origin/gh/janeyx99/201/base 2025-12-04T09:20:35.4553051Z * [new branch] gh/janeyx99/201/head -> origin/gh/janeyx99/201/head 2025-12-04T09:20:35.4554802Z * [new branch] gh/janeyx99/201/orig -> origin/gh/janeyx99/201/orig 2025-12-04T09:20:35.4557292Z * [new branch] gh/janeyx99/225/base -> origin/gh/janeyx99/225/base 2025-12-04T09:20:35.4559122Z * [new branch] gh/janeyx99/225/head -> origin/gh/janeyx99/225/head 2025-12-04T09:20:35.4560839Z * [new branch] gh/janeyx99/225/orig -> origin/gh/janeyx99/225/orig 2025-12-04T09:20:35.4563163Z * [new branch] gh/janeyx99/299/base -> origin/gh/janeyx99/299/base 2025-12-04T09:20:35.4565044Z * [new branch] gh/janeyx99/299/head -> origin/gh/janeyx99/299/head 2025-12-04T09:20:35.4566672Z * [new branch] gh/janeyx99/299/orig -> origin/gh/janeyx99/299/orig 2025-12-04T09:20:35.4569370Z * [new branch] gh/janeyx99/302/base -> origin/gh/janeyx99/302/base 2025-12-04T09:20:35.4571257Z * [new branch] gh/janeyx99/302/head -> origin/gh/janeyx99/302/head 2025-12-04T09:20:35.4575244Z * [new branch] gh/janeyx99/303/base -> origin/gh/janeyx99/303/base 2025-12-04T09:20:35.4576884Z * [new branch] gh/janeyx99/303/head -> origin/gh/janeyx99/303/head 2025-12-04T09:20:35.4579304Z * [new branch] gh/janeyx99/305/base -> origin/gh/janeyx99/305/base 2025-12-04T09:20:35.4581066Z * [new branch] gh/janeyx99/305/head -> origin/gh/janeyx99/305/head 2025-12-04T09:20:35.4583374Z * [new branch] gh/janeyx99/306/base -> origin/gh/janeyx99/306/base 2025-12-04T09:20:35.4585091Z * [new branch] gh/janeyx99/306/head -> origin/gh/janeyx99/306/head 2025-12-04T09:20:35.4587482Z * [new branch] gh/janeyx99/314/base -> origin/gh/janeyx99/314/base 2025-12-04T09:20:35.4589294Z * [new branch] gh/janeyx99/314/head -> origin/gh/janeyx99/314/head 2025-12-04T09:20:35.4591093Z * [new branch] gh/janeyx99/314/orig -> origin/gh/janeyx99/314/orig 2025-12-04T09:20:35.4593377Z * [new branch] gh/janeyx99/315/base -> origin/gh/janeyx99/315/base 2025-12-04T09:20:35.4595203Z * [new branch] gh/janeyx99/315/head -> origin/gh/janeyx99/315/head 2025-12-04T09:20:35.4596999Z * [new branch] gh/janeyx99/315/orig -> origin/gh/janeyx99/315/orig 2025-12-04T09:20:35.4599378Z * [new branch] gh/janeyx99/316/base -> origin/gh/janeyx99/316/base 2025-12-04T09:20:35.4601100Z * [new branch] gh/janeyx99/316/head -> origin/gh/janeyx99/316/head 2025-12-04T09:20:35.4602955Z * [new branch] gh/janeyx99/316/orig -> origin/gh/janeyx99/316/orig 2025-12-04T09:20:35.4605457Z * [new branch] gh/janeyx99/317/base -> origin/gh/janeyx99/317/base 2025-12-04T09:20:35.4607113Z * [new branch] gh/janeyx99/317/head -> origin/gh/janeyx99/317/head 2025-12-04T09:20:35.4608826Z * [new branch] gh/janeyx99/317/orig -> origin/gh/janeyx99/317/orig 2025-12-04T09:20:35.4611226Z * [new branch] gh/janeyx99/325/base -> origin/gh/janeyx99/325/base 2025-12-04T09:20:35.4613095Z * [new branch] gh/janeyx99/325/head -> origin/gh/janeyx99/325/head 2025-12-04T09:20:35.4614805Z * [new branch] gh/janeyx99/325/orig -> origin/gh/janeyx99/325/orig 2025-12-04T09:20:35.4617192Z * [new branch] gh/janeyx99/327/base -> origin/gh/janeyx99/327/base 2025-12-04T09:20:35.4618948Z * [new branch] gh/janeyx99/327/head -> origin/gh/janeyx99/327/head 2025-12-04T09:20:35.4620853Z * [new branch] gh/janeyx99/327/orig -> origin/gh/janeyx99/327/orig 2025-12-04T09:20:35.4623069Z * [new branch] gh/janeyx99/328/base -> origin/gh/janeyx99/328/base 2025-12-04T09:20:35.4624851Z * [new branch] gh/janeyx99/328/head -> origin/gh/janeyx99/328/head 2025-12-04T09:20:35.4626622Z * [new branch] gh/janeyx99/328/orig -> origin/gh/janeyx99/328/orig 2025-12-04T09:20:35.4628821Z * [new branch] gh/janeyx99/329/base -> origin/gh/janeyx99/329/base 2025-12-04T09:20:35.4630585Z * [new branch] gh/janeyx99/329/head -> origin/gh/janeyx99/329/head 2025-12-04T09:20:35.4632476Z * [new branch] gh/janeyx99/329/orig -> origin/gh/janeyx99/329/orig 2025-12-04T09:20:35.4635523Z * [new branch] gh/janeyx99/330/base -> origin/gh/janeyx99/330/base 2025-12-04T09:20:35.4637305Z * [new branch] gh/janeyx99/330/head -> origin/gh/janeyx99/330/head 2025-12-04T09:20:35.4638885Z * [new branch] gh/janeyx99/330/orig -> origin/gh/janeyx99/330/orig 2025-12-04T09:20:35.4641332Z * [new branch] gh/janeyx99/331/base -> origin/gh/janeyx99/331/base 2025-12-04T09:20:35.4643135Z * [new branch] gh/janeyx99/331/head -> origin/gh/janeyx99/331/head 2025-12-04T09:20:35.4644842Z * [new branch] gh/janeyx99/331/orig -> origin/gh/janeyx99/331/orig 2025-12-04T09:20:35.4647452Z * [new branch] gh/janeyx99/332/base -> origin/gh/janeyx99/332/base 2025-12-04T09:20:35.4649064Z * [new branch] gh/janeyx99/332/head -> origin/gh/janeyx99/332/head 2025-12-04T09:20:35.4650714Z * [new branch] gh/janeyx99/332/orig -> origin/gh/janeyx99/332/orig 2025-12-04T09:20:35.4653018Z * [new branch] gh/janeyx99/333/base -> origin/gh/janeyx99/333/base 2025-12-04T09:20:35.4654737Z * [new branch] gh/janeyx99/333/head -> origin/gh/janeyx99/333/head 2025-12-04T09:20:35.4656579Z * [new branch] gh/janeyx99/333/orig -> origin/gh/janeyx99/333/orig 2025-12-04T09:20:35.4659079Z * [new branch] gh/janeyx99/88/base -> origin/gh/janeyx99/88/base 2025-12-04T09:20:35.4660970Z * [new branch] gh/janeyx99/88/head -> origin/gh/janeyx99/88/head 2025-12-04T09:20:35.4662606Z * [new branch] gh/janeyx99/88/orig -> origin/gh/janeyx99/88/orig 2025-12-04T09:20:35.4665686Z * [new branch] gh/jansel/360/base -> origin/gh/jansel/360/base 2025-12-04T09:20:35.4667258Z * [new branch] gh/jansel/360/head -> origin/gh/jansel/360/head 2025-12-04T09:20:35.4669559Z * [new branch] gh/jansel/451/base -> origin/gh/jansel/451/base 2025-12-04T09:20:35.4671773Z * [new branch] gh/jansel/451/head -> origin/gh/jansel/451/head 2025-12-04T09:20:35.4673541Z * [new branch] gh/jansel/451/orig -> origin/gh/jansel/451/orig 2025-12-04T09:20:35.4675818Z * [new branch] gh/jansel/462/base -> origin/gh/jansel/462/base 2025-12-04T09:20:35.4677520Z * [new branch] gh/jansel/462/head -> origin/gh/jansel/462/head 2025-12-04T09:20:35.4679269Z * [new branch] gh/jansel/462/orig -> origin/gh/jansel/462/orig 2025-12-04T09:20:35.4681624Z * [new branch] gh/jansel/533/base -> origin/gh/jansel/533/base 2025-12-04T09:20:35.4683347Z * [new branch] gh/jansel/533/head -> origin/gh/jansel/533/head 2025-12-04T09:20:35.4685086Z * [new branch] gh/jansel/533/orig -> origin/gh/jansel/533/orig 2025-12-04T09:20:35.4687454Z * [new branch] gh/jansel/552/base -> origin/gh/jansel/552/base 2025-12-04T09:20:35.4689212Z * [new branch] gh/jansel/552/head -> origin/gh/jansel/552/head 2025-12-04T09:20:35.4690905Z * [new branch] gh/jansel/552/orig -> origin/gh/jansel/552/orig 2025-12-04T09:20:35.4693255Z * [new branch] gh/jansel/553/base -> origin/gh/jansel/553/base 2025-12-04T09:20:35.4694974Z * [new branch] gh/jansel/553/head -> origin/gh/jansel/553/head 2025-12-04T09:20:35.4696687Z * [new branch] gh/jansel/553/orig -> origin/gh/jansel/553/orig 2025-12-04T09:20:35.4698956Z * [new branch] gh/jansel/554/base -> origin/gh/jansel/554/base 2025-12-04T09:20:35.4700824Z * [new branch] gh/jansel/554/head -> origin/gh/jansel/554/head 2025-12-04T09:20:35.4702445Z * [new branch] gh/jansel/554/orig -> origin/gh/jansel/554/orig 2025-12-04T09:20:35.4704790Z * [new branch] gh/jansel/555/base -> origin/gh/jansel/555/base 2025-12-04T09:20:35.4706652Z * [new branch] gh/jansel/555/head -> origin/gh/jansel/555/head 2025-12-04T09:20:35.4708265Z * [new branch] gh/jansel/555/orig -> origin/gh/jansel/555/orig 2025-12-04T09:20:35.4710840Z * [new branch] gh/jansel/556/base -> origin/gh/jansel/556/base 2025-12-04T09:20:35.4712521Z * [new branch] gh/jansel/556/head -> origin/gh/jansel/556/head 2025-12-04T09:20:35.4714344Z * [new branch] gh/jansel/556/orig -> origin/gh/jansel/556/orig 2025-12-04T09:20:35.4716654Z * [new branch] gh/jansel/557/base -> origin/gh/jansel/557/base 2025-12-04T09:20:35.4718545Z * [new branch] gh/jansel/557/head -> origin/gh/jansel/557/head 2025-12-04T09:20:35.4720247Z * [new branch] gh/jansel/557/orig -> origin/gh/jansel/557/orig 2025-12-04T09:20:35.4722488Z * [new branch] gh/jansel/558/base -> origin/gh/jansel/558/base 2025-12-04T09:20:35.4724250Z * [new branch] gh/jansel/558/head -> origin/gh/jansel/558/head 2025-12-04T09:20:35.4726040Z * [new branch] gh/jansel/558/orig -> origin/gh/jansel/558/orig 2025-12-04T09:20:35.4728351Z * [new branch] gh/jansel/559/base -> origin/gh/jansel/559/base 2025-12-04T09:20:35.4730062Z * [new branch] gh/jansel/559/head -> origin/gh/jansel/559/head 2025-12-04T09:20:35.4731868Z * [new branch] gh/jansel/559/orig -> origin/gh/jansel/559/orig 2025-12-04T09:20:35.4734593Z * [new branch] gh/jansel/560/base -> origin/gh/jansel/560/base 2025-12-04T09:20:35.4735970Z * [new branch] gh/jansel/560/head -> origin/gh/jansel/560/head 2025-12-04T09:20:35.4737690Z * [new branch] gh/jansel/560/orig -> origin/gh/jansel/560/orig 2025-12-04T09:20:35.4740138Z * [new branch] gh/jansel/561/base -> origin/gh/jansel/561/base 2025-12-04T09:20:35.4741913Z * [new branch] gh/jansel/561/head -> origin/gh/jansel/561/head 2025-12-04T09:20:35.4743546Z * [new branch] gh/jansel/561/orig -> origin/gh/jansel/561/orig 2025-12-04T09:20:35.4746417Z * [new branch] gh/jansel/562/base -> origin/gh/jansel/562/base 2025-12-04T09:20:35.4748170Z * [new branch] gh/jansel/562/head -> origin/gh/jansel/562/head 2025-12-04T09:20:35.4749886Z * [new branch] gh/jansel/562/orig -> origin/gh/jansel/562/orig 2025-12-04T09:20:35.4752107Z * [new branch] gh/jansel/563/base -> origin/gh/jansel/563/base 2025-12-04T09:20:35.4753892Z * [new branch] gh/jansel/563/head -> origin/gh/jansel/563/head 2025-12-04T09:20:35.4755655Z * [new branch] gh/jansel/563/orig -> origin/gh/jansel/563/orig 2025-12-04T09:20:35.4758926Z * [new branch] gh/jansel/564/base -> origin/gh/jansel/564/base 2025-12-04T09:20:35.4760725Z * [new branch] gh/jansel/564/head -> origin/gh/jansel/564/head 2025-12-04T09:20:35.4762478Z * [new branch] gh/jansel/564/orig -> origin/gh/jansel/564/orig 2025-12-04T09:20:35.4764957Z * [new branch] gh/jansel/565/base -> origin/gh/jansel/565/base 2025-12-04T09:20:35.4766697Z * [new branch] gh/jansel/565/head -> origin/gh/jansel/565/head 2025-12-04T09:20:35.4768436Z * [new branch] gh/jansel/565/orig -> origin/gh/jansel/565/orig 2025-12-04T09:20:35.4770876Z * [new branch] gh/jansel/566/base -> origin/gh/jansel/566/base 2025-12-04T09:20:35.4772931Z * [new branch] gh/jansel/566/head -> origin/gh/jansel/566/head 2025-12-04T09:20:35.4774684Z * [new branch] gh/jansel/566/orig -> origin/gh/jansel/566/orig 2025-12-04T09:20:35.4777107Z * [new branch] gh/jansel/567/base -> origin/gh/jansel/567/base 2025-12-04T09:20:35.4778963Z * [new branch] gh/jansel/567/head -> origin/gh/jansel/567/head 2025-12-04T09:20:35.4780858Z * [new branch] gh/jansel/567/orig -> origin/gh/jansel/567/orig 2025-12-04T09:20:35.4783107Z * [new branch] gh/jansel/568/base -> origin/gh/jansel/568/base 2025-12-04T09:20:35.4785448Z * [new branch] gh/jansel/568/head -> origin/gh/jansel/568/head 2025-12-04T09:20:35.4787242Z * [new branch] gh/jansel/568/orig -> origin/gh/jansel/568/orig 2025-12-04T09:20:35.4789674Z * [new branch] gh/jansel/569/base -> origin/gh/jansel/569/base 2025-12-04T09:20:35.4791439Z * [new branch] gh/jansel/569/head -> origin/gh/jansel/569/head 2025-12-04T09:20:35.4793163Z * [new branch] gh/jansel/569/orig -> origin/gh/jansel/569/orig 2025-12-04T09:20:35.4795540Z * [new branch] gh/jansel/570/base -> origin/gh/jansel/570/base 2025-12-04T09:20:35.4797360Z * [new branch] gh/jansel/570/head -> origin/gh/jansel/570/head 2025-12-04T09:20:35.4799067Z * [new branch] gh/jansel/570/orig -> origin/gh/jansel/570/orig 2025-12-04T09:20:35.4801459Z * [new branch] gh/jansel/571/base -> origin/gh/jansel/571/base 2025-12-04T09:20:35.4803230Z * [new branch] gh/jansel/571/head -> origin/gh/jansel/571/head 2025-12-04T09:20:35.4804968Z * [new branch] gh/jansel/571/orig -> origin/gh/jansel/571/orig 2025-12-04T09:20:35.4807373Z * [new branch] gh/jansel/572/base -> origin/gh/jansel/572/base 2025-12-04T09:20:35.4809133Z * [new branch] gh/jansel/572/head -> origin/gh/jansel/572/head 2025-12-04T09:20:35.4810884Z * [new branch] gh/jansel/572/orig -> origin/gh/jansel/572/orig 2025-12-04T09:20:35.4813246Z * [new branch] gh/jansel/573/base -> origin/gh/jansel/573/base 2025-12-04T09:20:35.4814976Z * [new branch] gh/jansel/573/head -> origin/gh/jansel/573/head 2025-12-04T09:20:35.4816699Z * [new branch] gh/jansel/573/orig -> origin/gh/jansel/573/orig 2025-12-04T09:20:35.4819163Z * [new branch] gh/jansel/574/base -> origin/gh/jansel/574/base 2025-12-04T09:20:35.4821011Z * [new branch] gh/jansel/574/head -> origin/gh/jansel/574/head 2025-12-04T09:20:35.4822822Z * [new branch] gh/jansel/574/orig -> origin/gh/jansel/574/orig 2025-12-04T09:20:35.4825266Z * [new branch] gh/jansel/575/base -> origin/gh/jansel/575/base 2025-12-04T09:20:35.4827047Z * [new branch] gh/jansel/575/head -> origin/gh/jansel/575/head 2025-12-04T09:20:35.4828775Z * [new branch] gh/jansel/575/orig -> origin/gh/jansel/575/orig 2025-12-04T09:20:35.4831287Z * [new branch] gh/jansel/576/base -> origin/gh/jansel/576/base 2025-12-04T09:20:35.4833150Z * [new branch] gh/jansel/576/head -> origin/gh/jansel/576/head 2025-12-04T09:20:35.4834867Z * [new branch] gh/jansel/576/orig -> origin/gh/jansel/576/orig 2025-12-04T09:20:35.4837729Z * [new branch] gh/jbschlosser/247/base -> origin/gh/jbschlosser/247/base 2025-12-04T09:20:35.4839505Z * [new branch] gh/jbschlosser/247/head -> origin/gh/jbschlosser/247/head 2025-12-04T09:20:35.4841648Z * [new branch] gh/jbschlosser/247/orig -> origin/gh/jbschlosser/247/orig 2025-12-04T09:20:35.4844294Z * [new branch] gh/jbschlosser/250/base -> origin/gh/jbschlosser/250/base 2025-12-04T09:20:35.4845945Z * [new branch] gh/jbschlosser/250/head -> origin/gh/jbschlosser/250/head 2025-12-04T09:20:35.4847660Z * [new branch] gh/jbschlosser/250/orig -> origin/gh/jbschlosser/250/orig 2025-12-04T09:20:35.4850706Z * [new branch] gh/jerryzh168/1/base -> origin/gh/jerryzh168/1/base 2025-12-04T09:20:35.4852513Z * [new branch] gh/jerryzh168/1/head -> origin/gh/jerryzh168/1/head 2025-12-04T09:20:35.4854210Z * [new branch] gh/jerryzh168/1/orig -> origin/gh/jerryzh168/1/orig 2025-12-04T09:20:35.4856998Z * [new branch] gh/jiayisunx/59/base -> origin/gh/jiayisunx/59/base 2025-12-04T09:20:35.4859270Z * [new branch] gh/jiayisunx/59/head -> origin/gh/jiayisunx/59/head 2025-12-04T09:20:35.4861137Z * [new branch] gh/jiayisunx/59/orig -> origin/gh/jiayisunx/59/orig 2025-12-04T09:20:35.4863790Z * [new branch] gh/jiayisunx/61/base -> origin/gh/jiayisunx/61/base 2025-12-04T09:20:35.4865531Z * [new branch] gh/jiayisunx/61/head -> origin/gh/jiayisunx/61/head 2025-12-04T09:20:35.4867240Z * [new branch] gh/jiayisunx/61/orig -> origin/gh/jiayisunx/61/orig 2025-12-04T09:20:35.4869612Z * [new branch] gh/jiayisunx/68/base -> origin/gh/jiayisunx/68/base 2025-12-04T09:20:35.4871480Z * [new branch] gh/jiayisunx/68/head -> origin/gh/jiayisunx/68/head 2025-12-04T09:20:35.4873380Z * [new branch] gh/jiayisunx/68/orig -> origin/gh/jiayisunx/68/orig 2025-12-04T09:20:35.4875758Z * [new branch] gh/jiayisunx/77/base -> origin/gh/jiayisunx/77/base 2025-12-04T09:20:35.4877517Z * [new branch] gh/jiayisunx/77/head -> origin/gh/jiayisunx/77/head 2025-12-04T09:20:35.4879343Z * [new branch] gh/jiayisunx/77/orig -> origin/gh/jiayisunx/77/orig 2025-12-04T09:20:35.4881633Z * [new branch] gh/jiayisunx/78/base -> origin/gh/jiayisunx/78/base 2025-12-04T09:20:35.4883341Z * [new branch] gh/jiayisunx/78/head -> origin/gh/jiayisunx/78/head 2025-12-04T09:20:35.4885167Z * [new branch] gh/jiayisunx/78/orig -> origin/gh/jiayisunx/78/orig 2025-12-04T09:20:35.4887412Z * [new branch] gh/jiayisunx/79/base -> origin/gh/jiayisunx/79/base 2025-12-04T09:20:35.4889146Z * [new branch] gh/jiayisunx/79/head -> origin/gh/jiayisunx/79/head 2025-12-04T09:20:35.4890870Z * [new branch] gh/jiayisunx/79/orig -> origin/gh/jiayisunx/79/orig 2025-12-04T09:20:35.4893355Z * [new branch] gh/jiayisunx/82/base -> origin/gh/jiayisunx/82/base 2025-12-04T09:20:35.4895035Z * [new branch] gh/jiayisunx/82/head -> origin/gh/jiayisunx/82/head 2025-12-04T09:20:35.4896761Z * [new branch] gh/jiayisunx/82/orig -> origin/gh/jiayisunx/82/orig 2025-12-04T09:20:35.4899705Z * [new branch] gh/jiayisunx/83/base -> origin/gh/jiayisunx/83/base 2025-12-04T09:20:35.4901522Z * [new branch] gh/jiayisunx/83/head -> origin/gh/jiayisunx/83/head 2025-12-04T09:20:35.4903105Z * [new branch] gh/jiayisunx/83/orig -> origin/gh/jiayisunx/83/orig 2025-12-04T09:20:35.4905364Z * [new branch] gh/jiayisunx/84/base -> origin/gh/jiayisunx/84/base 2025-12-04T09:20:35.4907218Z * [new branch] gh/jiayisunx/84/head -> origin/gh/jiayisunx/84/head 2025-12-04T09:20:35.4909003Z * [new branch] gh/jiayisunx/84/orig -> origin/gh/jiayisunx/84/orig 2025-12-04T09:20:35.4911303Z * [new branch] gh/jiayisunx/85/base -> origin/gh/jiayisunx/85/base 2025-12-04T09:20:35.4913086Z * [new branch] gh/jiayisunx/85/head -> origin/gh/jiayisunx/85/head 2025-12-04T09:20:35.4914788Z * [new branch] gh/jiayisunx/85/orig -> origin/gh/jiayisunx/85/orig 2025-12-04T09:20:35.4917145Z * [new branch] gh/jiayisunx/86/base -> origin/gh/jiayisunx/86/base 2025-12-04T09:20:35.4918815Z * [new branch] gh/jiayisunx/86/head -> origin/gh/jiayisunx/86/head 2025-12-04T09:20:35.4920714Z * [new branch] gh/jiayisunx/86/orig -> origin/gh/jiayisunx/86/orig 2025-12-04T09:20:35.4923051Z * [new branch] gh/jiayisunx/87/base -> origin/gh/jiayisunx/87/base 2025-12-04T09:20:35.4924820Z * [new branch] gh/jiayisunx/87/head -> origin/gh/jiayisunx/87/head 2025-12-04T09:20:35.4926508Z * [new branch] gh/jiayisunx/87/orig -> origin/gh/jiayisunx/87/orig 2025-12-04T09:20:35.4929286Z * [new branch] gh/jiayisunx/88/base -> origin/gh/jiayisunx/88/base 2025-12-04T09:20:35.4931085Z * [new branch] gh/jiayisunx/88/head -> origin/gh/jiayisunx/88/head 2025-12-04T09:20:35.4932872Z * [new branch] gh/jiayisunx/88/orig -> origin/gh/jiayisunx/88/orig 2025-12-04T09:20:35.4935214Z * [new branch] gh/jiayisunx/89/base -> origin/gh/jiayisunx/89/base 2025-12-04T09:20:35.4936879Z * [new branch] gh/jiayisunx/89/head -> origin/gh/jiayisunx/89/head 2025-12-04T09:20:35.4938631Z * [new branch] gh/jiayisunx/89/orig -> origin/gh/jiayisunx/89/orig 2025-12-04T09:20:35.4941100Z * [new branch] gh/jiayisunx/90/base -> origin/gh/jiayisunx/90/base 2025-12-04T09:20:35.4942708Z * [new branch] gh/jiayisunx/90/head -> origin/gh/jiayisunx/90/head 2025-12-04T09:20:35.4944419Z * [new branch] gh/jiayisunx/90/orig -> origin/gh/jiayisunx/90/orig 2025-12-04T09:20:35.4947205Z * [new branch] gh/jjwu@meta.com/1/base -> origin/gh/jjwu@meta.com/1/base 2025-12-04T09:20:35.4948906Z * [new branch] gh/jjwu@meta.com/1/head -> origin/gh/jjwu@meta.com/1/head 2025-12-04T09:20:35.4951764Z * [new branch] gh/jturney/1/base -> origin/gh/jturney/1/base 2025-12-04T09:20:35.4953519Z * [new branch] gh/jturney/1/head -> origin/gh/jturney/1/head 2025-12-04T09:20:35.4955239Z * [new branch] gh/jturney/1/orig -> origin/gh/jturney/1/orig 2025-12-04T09:20:35.4957578Z * [new branch] gh/jturney/2/base -> origin/gh/jturney/2/base 2025-12-04T09:20:35.4959293Z * [new branch] gh/jturney/2/head -> origin/gh/jturney/2/head 2025-12-04T09:20:35.4960982Z * [new branch] gh/jturney/2/orig -> origin/gh/jturney/2/orig 2025-12-04T09:20:35.4963871Z * [new branch] gh/karthickai/10/base -> origin/gh/karthickai/10/base 2025-12-04T09:20:35.4965668Z * [new branch] gh/karthickai/10/head -> origin/gh/karthickai/10/head 2025-12-04T09:20:35.4967451Z * [new branch] gh/karthickai/10/orig -> origin/gh/karthickai/10/orig 2025-12-04T09:20:35.4969796Z * [new branch] gh/karthickai/11/base -> origin/gh/karthickai/11/base 2025-12-04T09:20:35.4971681Z * [new branch] gh/karthickai/11/head -> origin/gh/karthickai/11/head 2025-12-04T09:20:35.4974765Z * [new branch] gh/karthickai/11/orig -> origin/gh/karthickai/11/orig 2025-12-04T09:20:35.4977381Z * [new branch] gh/karthickai/12/base -> origin/gh/karthickai/12/base 2025-12-04T09:20:35.4979222Z * [new branch] gh/karthickai/12/head -> origin/gh/karthickai/12/head 2025-12-04T09:20:35.4981067Z * [new branch] gh/karthickai/12/orig -> origin/gh/karthickai/12/orig 2025-12-04T09:20:35.4983394Z * [new branch] gh/karthickai/13/base -> origin/gh/karthickai/13/base 2025-12-04T09:20:35.4985185Z * [new branch] gh/karthickai/13/head -> origin/gh/karthickai/13/head 2025-12-04T09:20:35.4986934Z * [new branch] gh/karthickai/13/orig -> origin/gh/karthickai/13/orig 2025-12-04T09:20:35.4989350Z * [new branch] gh/karthickai/14/base -> origin/gh/karthickai/14/base 2025-12-04T09:20:35.4991140Z * [new branch] gh/karthickai/14/head -> origin/gh/karthickai/14/head 2025-12-04T09:20:35.4993127Z * [new branch] gh/karthickai/14/orig -> origin/gh/karthickai/14/orig 2025-12-04T09:20:35.4995567Z * [new branch] gh/karthickai/15/base -> origin/gh/karthickai/15/base 2025-12-04T09:20:35.4997374Z * [new branch] gh/karthickai/15/head -> origin/gh/karthickai/15/head 2025-12-04T09:20:35.4999047Z * [new branch] gh/karthickai/15/orig -> origin/gh/karthickai/15/orig 2025-12-04T09:20:35.5001331Z * [new branch] gh/karthickai/16/base -> origin/gh/karthickai/16/base 2025-12-04T09:20:35.5003174Z * [new branch] gh/karthickai/16/head -> origin/gh/karthickai/16/head 2025-12-04T09:20:35.5005005Z * [new branch] gh/karthickai/16/orig -> origin/gh/karthickai/16/orig 2025-12-04T09:20:35.5007208Z * [new branch] gh/karthickai/17/base -> origin/gh/karthickai/17/base 2025-12-04T09:20:35.5008892Z * [new branch] gh/karthickai/17/head -> origin/gh/karthickai/17/head 2025-12-04T09:20:35.5010609Z * [new branch] gh/karthickai/17/orig -> origin/gh/karthickai/17/orig 2025-12-04T09:20:35.5013049Z * [new branch] gh/karthickai/18/base -> origin/gh/karthickai/18/base 2025-12-04T09:20:35.5014914Z * [new branch] gh/karthickai/18/head -> origin/gh/karthickai/18/head 2025-12-04T09:20:35.5016965Z * [new branch] gh/karthickai/18/orig -> origin/gh/karthickai/18/orig 2025-12-04T09:20:35.5019370Z * [new branch] gh/karthickai/19/base -> origin/gh/karthickai/19/base 2025-12-04T09:20:35.5021292Z * [new branch] gh/karthickai/19/head -> origin/gh/karthickai/19/head 2025-12-04T09:20:35.5022877Z * [new branch] gh/karthickai/19/orig -> origin/gh/karthickai/19/orig 2025-12-04T09:20:35.5025793Z * [new branch] gh/karthickai/20/base -> origin/gh/karthickai/20/base 2025-12-04T09:20:35.5028258Z * [new branch] gh/karthickai/20/head -> origin/gh/karthickai/20/head 2025-12-04T09:20:35.5030023Z * [new branch] gh/karthickai/20/orig -> origin/gh/karthickai/20/orig 2025-12-04T09:20:35.5032657Z * [new branch] gh/karthickai/21/base -> origin/gh/karthickai/21/base 2025-12-04T09:20:35.5034513Z * [new branch] gh/karthickai/21/head -> origin/gh/karthickai/21/head 2025-12-04T09:20:35.5036374Z * [new branch] gh/karthickai/21/orig -> origin/gh/karthickai/21/orig 2025-12-04T09:20:35.5038969Z * [new branch] gh/karthickai/22/base -> origin/gh/karthickai/22/base 2025-12-04T09:20:35.5040498Z * [new branch] gh/karthickai/22/head -> origin/gh/karthickai/22/head 2025-12-04T09:20:35.5042226Z * [new branch] gh/karthickai/22/orig -> origin/gh/karthickai/22/orig 2025-12-04T09:20:35.5044642Z * [new branch] gh/karthickai/23/base -> origin/gh/karthickai/23/base 2025-12-04T09:20:35.5046520Z * [new branch] gh/karthickai/23/head -> origin/gh/karthickai/23/head 2025-12-04T09:20:35.5048282Z * [new branch] gh/karthickai/23/orig -> origin/gh/karthickai/23/orig 2025-12-04T09:20:35.5050748Z * [new branch] gh/karthickai/24/base -> origin/gh/karthickai/24/base 2025-12-04T09:20:35.5052475Z * [new branch] gh/karthickai/24/head -> origin/gh/karthickai/24/head 2025-12-04T09:20:35.5054149Z * [new branch] gh/karthickai/24/orig -> origin/gh/karthickai/24/orig 2025-12-04T09:20:35.5056877Z * [new branch] gh/karthickai/25/base -> origin/gh/karthickai/25/base 2025-12-04T09:20:35.5089544Z * [new branch] gh/karthickai/25/head -> origin/gh/karthickai/25/head 2025-12-04T09:20:35.5089913Z * [new branch] gh/karthickai/25/orig -> origin/gh/karthickai/25/orig 2025-12-04T09:20:35.5090143Z * [new branch] gh/karthickai/26/base -> origin/gh/karthickai/26/base 2025-12-04T09:20:35.5090543Z * [new branch] gh/karthickai/26/head -> origin/gh/karthickai/26/head 2025-12-04T09:20:35.5090739Z * [new branch] gh/karthickai/26/orig -> origin/gh/karthickai/26/orig 2025-12-04T09:20:35.5090937Z * [new branch] gh/karthickai/6/base -> origin/gh/karthickai/6/base 2025-12-04T09:20:35.5091139Z * [new branch] gh/karthickai/6/head -> origin/gh/karthickai/6/head 2025-12-04T09:20:35.5091343Z * [new branch] gh/karthickai/6/orig -> origin/gh/karthickai/6/orig 2025-12-04T09:20:35.5091522Z * [new branch] gh/krocki/1/base -> origin/gh/krocki/1/base 2025-12-04T09:20:35.5091812Z * [new branch] gh/krocki/1/head -> origin/gh/krocki/1/head 2025-12-04T09:20:35.5091980Z * [new branch] gh/krocki/1/orig -> origin/gh/krocki/1/orig 2025-12-04T09:20:35.5092140Z * [new branch] gh/krocki/2/base -> origin/gh/krocki/2/base 2025-12-04T09:20:35.5092305Z * [new branch] gh/krocki/2/head -> origin/gh/krocki/2/head 2025-12-04T09:20:35.5092465Z * [new branch] gh/krocki/2/orig -> origin/gh/krocki/2/orig 2025-12-04T09:20:35.5092681Z * [new branch] gh/kurtamohler/60/base -> origin/gh/kurtamohler/60/base 2025-12-04T09:20:35.5092892Z * [new branch] gh/kurtamohler/60/head -> origin/gh/kurtamohler/60/head 2025-12-04T09:20:35.5094001Z * [new branch] gh/kurtamohler/60/orig -> origin/gh/kurtamohler/60/orig 2025-12-04T09:20:35.5096481Z * [new branch] gh/kurtamohler/61/base -> origin/gh/kurtamohler/61/base 2025-12-04T09:20:35.5098106Z * [new branch] gh/kurtamohler/61/head -> origin/gh/kurtamohler/61/head 2025-12-04T09:20:35.5099927Z * [new branch] gh/kurtamohler/61/orig -> origin/gh/kurtamohler/61/orig 2025-12-04T09:20:35.5102171Z * [new branch] gh/kurtamohler/62/base -> origin/gh/kurtamohler/62/base 2025-12-04T09:20:35.5103918Z * [new branch] gh/kurtamohler/62/head -> origin/gh/kurtamohler/62/head 2025-12-04T09:20:35.5105602Z * [new branch] gh/kurtamohler/62/orig -> origin/gh/kurtamohler/62/orig 2025-12-04T09:20:35.5107966Z * [new branch] gh/kurtamohler/63/base -> origin/gh/kurtamohler/63/base 2025-12-04T09:20:35.5109717Z * [new branch] gh/kurtamohler/63/head -> origin/gh/kurtamohler/63/head 2025-12-04T09:20:35.5111552Z * [new branch] gh/kurtamohler/63/orig -> origin/gh/kurtamohler/63/orig 2025-12-04T09:20:35.5113910Z * [new branch] gh/kurtamohler/64/base -> origin/gh/kurtamohler/64/base 2025-12-04T09:20:35.5115617Z * [new branch] gh/kurtamohler/64/head -> origin/gh/kurtamohler/64/head 2025-12-04T09:20:35.5117334Z * [new branch] gh/kurtamohler/64/orig -> origin/gh/kurtamohler/64/orig 2025-12-04T09:20:35.5119771Z * [new branch] gh/kurtamohler/65/base -> origin/gh/kurtamohler/65/base 2025-12-04T09:20:35.5121492Z * [new branch] gh/kurtamohler/65/head -> origin/gh/kurtamohler/65/head 2025-12-04T09:20:35.5123667Z * [new branch] gh/kurtamohler/65/orig -> origin/gh/kurtamohler/65/orig 2025-12-04T09:20:35.5126008Z * [new branch] gh/kurtamohler/66/base -> origin/gh/kurtamohler/66/base 2025-12-04T09:20:35.5127854Z * [new branch] gh/kurtamohler/66/head -> origin/gh/kurtamohler/66/head 2025-12-04T09:20:35.5129612Z * [new branch] gh/kurtamohler/66/orig -> origin/gh/kurtamohler/66/orig 2025-12-04T09:20:35.5132016Z * [new branch] gh/kurtamohler/67/base -> origin/gh/kurtamohler/67/base 2025-12-04T09:20:35.5133752Z * [new branch] gh/kurtamohler/67/head -> origin/gh/kurtamohler/67/head 2025-12-04T09:20:35.5135634Z * [new branch] gh/kurtamohler/67/orig -> origin/gh/kurtamohler/67/orig 2025-12-04T09:20:35.5138483Z * [new branch] gh/kwen2501/130/base -> origin/gh/kwen2501/130/base 2025-12-04T09:20:35.5140408Z * [new branch] gh/kwen2501/130/head -> origin/gh/kwen2501/130/head 2025-12-04T09:20:35.5142117Z * [new branch] gh/kwen2501/130/orig -> origin/gh/kwen2501/130/orig 2025-12-04T09:20:35.5144505Z * [new branch] gh/kwen2501/170/base -> origin/gh/kwen2501/170/base 2025-12-04T09:20:35.5146262Z * [new branch] gh/kwen2501/170/head -> origin/gh/kwen2501/170/head 2025-12-04T09:20:35.5148697Z * [new branch] gh/kwen2501/187/base -> origin/gh/kwen2501/187/base 2025-12-04T09:20:35.5150495Z * [new branch] gh/kwen2501/187/head -> origin/gh/kwen2501/187/head 2025-12-04T09:20:35.5152228Z * [new branch] gh/kwen2501/187/orig -> origin/gh/kwen2501/187/orig 2025-12-04T09:20:35.5154545Z * [new branch] gh/kwen2501/188/base -> origin/gh/kwen2501/188/base 2025-12-04T09:20:35.5156242Z * [new branch] gh/kwen2501/188/head -> origin/gh/kwen2501/188/head 2025-12-04T09:20:35.5157988Z * [new branch] gh/kwen2501/188/orig -> origin/gh/kwen2501/188/orig 2025-12-04T09:20:35.5160419Z * [new branch] gh/kwen2501/211/base -> origin/gh/kwen2501/211/base 2025-12-04T09:20:35.5162149Z * [new branch] gh/kwen2501/211/head -> origin/gh/kwen2501/211/head 2025-12-04T09:20:35.5164736Z * [new branch] gh/kwen2501/224/base -> origin/gh/kwen2501/224/base 2025-12-04T09:20:35.5166824Z * [new branch] gh/kwen2501/224/head -> origin/gh/kwen2501/224/head 2025-12-04T09:20:35.5168570Z * [new branch] gh/kwen2501/224/orig -> origin/gh/kwen2501/224/orig 2025-12-04T09:20:35.5170834Z * [new branch] gh/kwen2501/228/base -> origin/gh/kwen2501/228/base 2025-12-04T09:20:35.5172902Z * [new branch] gh/kwen2501/228/head -> origin/gh/kwen2501/228/head 2025-12-04T09:20:35.5174605Z * [new branch] gh/kwen2501/228/orig -> origin/gh/kwen2501/228/orig 2025-12-04T09:20:35.5177083Z * [new branch] gh/kwen2501/234/base -> origin/gh/kwen2501/234/base 2025-12-04T09:20:35.5178740Z * [new branch] gh/kwen2501/234/head -> origin/gh/kwen2501/234/head 2025-12-04T09:20:35.5180701Z * [new branch] gh/kwen2501/234/orig -> origin/gh/kwen2501/234/orig 2025-12-04T09:20:35.5183500Z * [new branch] gh/kwen2501/235/base -> origin/gh/kwen2501/235/base 2025-12-04T09:20:35.5185350Z * [new branch] gh/kwen2501/235/head -> origin/gh/kwen2501/235/head 2025-12-04T09:20:35.5187113Z * [new branch] gh/kwen2501/235/orig -> origin/gh/kwen2501/235/orig 2025-12-04T09:20:35.5189440Z * [new branch] gh/kwen2501/236/base -> origin/gh/kwen2501/236/base 2025-12-04T09:20:35.5191198Z * [new branch] gh/kwen2501/236/head -> origin/gh/kwen2501/236/head 2025-12-04T09:20:35.5192879Z * [new branch] gh/kwen2501/236/orig -> origin/gh/kwen2501/236/orig 2025-12-04T09:20:35.5195157Z * [new branch] gh/kwen2501/237/base -> origin/gh/kwen2501/237/base 2025-12-04T09:20:35.5196884Z * [new branch] gh/kwen2501/237/head -> origin/gh/kwen2501/237/head 2025-12-04T09:20:35.5198582Z * [new branch] gh/kwen2501/237/orig -> origin/gh/kwen2501/237/orig 2025-12-04T09:20:35.5200889Z * [new branch] gh/kwen2501/238/base -> origin/gh/kwen2501/238/base 2025-12-04T09:20:35.5202592Z * [new branch] gh/kwen2501/238/head -> origin/gh/kwen2501/238/head 2025-12-04T09:20:35.5204666Z * [new branch] gh/kwen2501/238/orig -> origin/gh/kwen2501/238/orig 2025-12-04T09:20:35.5206887Z * [new branch] gh/kwen2501/240/base -> origin/gh/kwen2501/240/base 2025-12-04T09:20:35.5208607Z * [new branch] gh/kwen2501/240/head -> origin/gh/kwen2501/240/head 2025-12-04T09:20:35.5210330Z * [new branch] gh/kwen2501/240/orig -> origin/gh/kwen2501/240/orig 2025-12-04T09:20:35.5213186Z * [new branch] gh/kwen2501/241/base -> origin/gh/kwen2501/241/base 2025-12-04T09:20:35.5215021Z * [new branch] gh/kwen2501/241/head -> origin/gh/kwen2501/241/head 2025-12-04T09:20:35.5216750Z * [new branch] gh/kwen2501/241/orig -> origin/gh/kwen2501/241/orig 2025-12-04T09:20:35.5219178Z * [new branch] gh/kwen2501/247/base -> origin/gh/kwen2501/247/base 2025-12-04T09:20:35.5221489Z * [new branch] gh/kwen2501/247/head -> origin/gh/kwen2501/247/head 2025-12-04T09:20:35.5223056Z * [new branch] gh/kwen2501/247/orig -> origin/gh/kwen2501/247/orig 2025-12-04T09:20:35.5225364Z * [new branch] gh/kwen2501/252/base -> origin/gh/kwen2501/252/base 2025-12-04T09:20:35.5227074Z * [new branch] gh/kwen2501/252/head -> origin/gh/kwen2501/252/head 2025-12-04T09:20:35.5228878Z * [new branch] gh/kwen2501/252/orig -> origin/gh/kwen2501/252/orig 2025-12-04T09:20:35.5231668Z * [new branch] gh/kwen2501/259/base -> origin/gh/kwen2501/259/base 2025-12-04T09:20:35.5233557Z * [new branch] gh/kwen2501/259/head -> origin/gh/kwen2501/259/head 2025-12-04T09:20:35.5235351Z * [new branch] gh/kwen2501/259/orig -> origin/gh/kwen2501/259/orig 2025-12-04T09:20:35.5237877Z * [new branch] gh/kwen2501/260/base -> origin/gh/kwen2501/260/base 2025-12-04T09:20:35.5239728Z * [new branch] gh/kwen2501/260/head -> origin/gh/kwen2501/260/head 2025-12-04T09:20:35.5241514Z * [new branch] gh/kwen2501/260/orig -> origin/gh/kwen2501/260/orig 2025-12-04T09:20:35.5243962Z * [new branch] gh/kwen2501/268/base -> origin/gh/kwen2501/268/base 2025-12-04T09:20:35.5245744Z * [new branch] gh/kwen2501/268/head -> origin/gh/kwen2501/268/head 2025-12-04T09:20:35.5247453Z * [new branch] gh/kwen2501/268/orig -> origin/gh/kwen2501/268/orig 2025-12-04T09:20:35.5249839Z * [new branch] gh/kwen2501/269/base -> origin/gh/kwen2501/269/base 2025-12-04T09:20:35.5251625Z * [new branch] gh/kwen2501/269/head -> origin/gh/kwen2501/269/head 2025-12-04T09:20:35.5253362Z * [new branch] gh/kwen2501/269/orig -> origin/gh/kwen2501/269/orig 2025-12-04T09:20:35.5255821Z * [new branch] gh/kwen2501/270/base -> origin/gh/kwen2501/270/base 2025-12-04T09:20:35.5257699Z * [new branch] gh/kwen2501/270/head -> origin/gh/kwen2501/270/head 2025-12-04T09:20:35.5259488Z * [new branch] gh/kwen2501/270/orig -> origin/gh/kwen2501/270/orig 2025-12-04T09:20:35.5262093Z * [new branch] gh/kwen2501/271/base -> origin/gh/kwen2501/271/base 2025-12-04T09:20:35.5263812Z * [new branch] gh/kwen2501/271/head -> origin/gh/kwen2501/271/head 2025-12-04T09:20:35.5265572Z * [new branch] gh/kwen2501/271/orig -> origin/gh/kwen2501/271/orig 2025-12-04T09:20:35.5268058Z * [new branch] gh/kwen2501/274/base -> origin/gh/kwen2501/274/base 2025-12-04T09:20:35.5269883Z * [new branch] gh/kwen2501/274/head -> origin/gh/kwen2501/274/head 2025-12-04T09:20:35.5271773Z * [new branch] gh/kwen2501/274/orig -> origin/gh/kwen2501/274/orig 2025-12-04T09:20:35.5274347Z * [new branch] gh/kwen2501/275/base -> origin/gh/kwen2501/275/base 2025-12-04T09:20:35.5276125Z * [new branch] gh/kwen2501/275/head -> origin/gh/kwen2501/275/head 2025-12-04T09:20:35.5278072Z * [new branch] gh/kwen2501/275/orig -> origin/gh/kwen2501/275/orig 2025-12-04T09:20:35.5280421Z * [new branch] gh/kwen2501/276/base -> origin/gh/kwen2501/276/base 2025-12-04T09:20:35.5282187Z * [new branch] gh/kwen2501/276/head -> origin/gh/kwen2501/276/head 2025-12-04T09:20:35.5283884Z * [new branch] gh/kwen2501/276/orig -> origin/gh/kwen2501/276/orig 2025-12-04T09:20:35.5286320Z * [new branch] gh/kwen2501/277/base -> origin/gh/kwen2501/277/base 2025-12-04T09:20:35.5288050Z * [new branch] gh/kwen2501/277/head -> origin/gh/kwen2501/277/head 2025-12-04T09:20:35.5289795Z * [new branch] gh/kwen2501/277/orig -> origin/gh/kwen2501/277/orig 2025-12-04T09:20:35.5292210Z * [new branch] gh/kwen2501/278/base -> origin/gh/kwen2501/278/base 2025-12-04T09:20:35.5293924Z * [new branch] gh/kwen2501/278/head -> origin/gh/kwen2501/278/head 2025-12-04T09:20:35.5295658Z * [new branch] gh/kwen2501/278/orig -> origin/gh/kwen2501/278/orig 2025-12-04T09:20:35.5298407Z * [new branch] gh/kwen2501/279/base -> origin/gh/kwen2501/279/base 2025-12-04T09:20:35.5300386Z * [new branch] gh/kwen2501/279/head -> origin/gh/kwen2501/279/head 2025-12-04T09:20:35.5302172Z * [new branch] gh/kwen2501/279/orig -> origin/gh/kwen2501/279/orig 2025-12-04T09:20:35.5304640Z * [new branch] gh/kwen2501/280/base -> origin/gh/kwen2501/280/base 2025-12-04T09:20:35.5306453Z * [new branch] gh/kwen2501/280/head -> origin/gh/kwen2501/280/head 2025-12-04T09:20:35.5308268Z * [new branch] gh/kwen2501/280/orig -> origin/gh/kwen2501/280/orig 2025-12-04T09:20:35.5310653Z * [new branch] gh/kwen2501/281/base -> origin/gh/kwen2501/281/base 2025-12-04T09:20:35.5312352Z * [new branch] gh/kwen2501/281/head -> origin/gh/kwen2501/281/head 2025-12-04T09:20:35.5314107Z * [new branch] gh/kwen2501/281/orig -> origin/gh/kwen2501/281/orig 2025-12-04T09:20:35.5316999Z * [new branch] gh/kwen2501/282/base -> origin/gh/kwen2501/282/base 2025-12-04T09:20:35.5318822Z * [new branch] gh/kwen2501/282/head -> origin/gh/kwen2501/282/head 2025-12-04T09:20:35.5320524Z * [new branch] gh/kwen2501/282/orig -> origin/gh/kwen2501/282/orig 2025-12-04T09:20:35.5322919Z * [new branch] gh/kwen2501/283/base -> origin/gh/kwen2501/283/base 2025-12-04T09:20:35.5324728Z * [new branch] gh/kwen2501/283/head -> origin/gh/kwen2501/283/head 2025-12-04T09:20:35.5326884Z * [new branch] gh/kwen2501/283/orig -> origin/gh/kwen2501/283/orig 2025-12-04T09:20:35.5329365Z * [new branch] gh/kwen2501/284/base -> origin/gh/kwen2501/284/base 2025-12-04T09:20:35.5331205Z * [new branch] gh/kwen2501/284/head -> origin/gh/kwen2501/284/head 2025-12-04T09:20:35.5332874Z * [new branch] gh/kwen2501/284/orig -> origin/gh/kwen2501/284/orig 2025-12-04T09:20:35.5335281Z * [new branch] gh/kwen2501/285/base -> origin/gh/kwen2501/285/base 2025-12-04T09:20:35.5336988Z * [new branch] gh/kwen2501/285/head -> origin/gh/kwen2501/285/head 2025-12-04T09:20:35.5338674Z * [new branch] gh/kwen2501/285/orig -> origin/gh/kwen2501/285/orig 2025-12-04T09:20:35.5341251Z * [new branch] gh/kwen2501/286/base -> origin/gh/kwen2501/286/base 2025-12-04T09:20:35.5343061Z * [new branch] gh/kwen2501/286/head -> origin/gh/kwen2501/286/head 2025-12-04T09:20:35.5344846Z * [new branch] gh/kwen2501/286/orig -> origin/gh/kwen2501/286/orig 2025-12-04T09:20:35.5347216Z * [new branch] gh/kwen2501/287/base -> origin/gh/kwen2501/287/base 2025-12-04T09:20:35.5349161Z * [new branch] gh/kwen2501/287/head -> origin/gh/kwen2501/287/head 2025-12-04T09:20:35.5350763Z * [new branch] gh/kwen2501/287/orig -> origin/gh/kwen2501/287/orig 2025-12-04T09:20:35.5353181Z * [new branch] gh/kwen2501/288/base -> origin/gh/kwen2501/288/base 2025-12-04T09:20:35.5355135Z * [new branch] gh/kwen2501/288/head -> origin/gh/kwen2501/288/head 2025-12-04T09:20:35.5356991Z * [new branch] gh/kwen2501/288/orig -> origin/gh/kwen2501/288/orig 2025-12-04T09:20:35.5359772Z * [new branch] gh/laithsakka/251/base -> origin/gh/laithsakka/251/base 2025-12-04T09:20:35.5361492Z * [new branch] gh/laithsakka/251/head -> origin/gh/laithsakka/251/head 2025-12-04T09:20:35.5363239Z * [new branch] gh/laithsakka/251/orig -> origin/gh/laithsakka/251/orig 2025-12-04T09:20:35.5365592Z * [new branch] gh/laithsakka/276/base -> origin/gh/laithsakka/276/base 2025-12-04T09:20:35.5367245Z * [new branch] gh/laithsakka/276/head -> origin/gh/laithsakka/276/head 2025-12-04T09:20:35.5368973Z * [new branch] gh/laithsakka/276/orig -> origin/gh/laithsakka/276/orig 2025-12-04T09:20:35.5372206Z * [new branch] gh/laithsakka/28/base -> origin/gh/laithsakka/28/base 2025-12-04T09:20:35.5375588Z * [new branch] gh/laithsakka/29/base -> origin/gh/laithsakka/29/base 2025-12-04T09:20:35.5377782Z * [new branch] gh/laithsakka/30/base -> origin/gh/laithsakka/30/base 2025-12-04T09:20:35.5379619Z * [new branch] gh/laithsakka/30/head -> origin/gh/laithsakka/30/head 2025-12-04T09:20:35.5381789Z * [new branch] gh/laithsakka/31/base -> origin/gh/laithsakka/31/base 2025-12-04T09:20:35.5383463Z * [new branch] gh/laithsakka/31/head -> origin/gh/laithsakka/31/head 2025-12-04T09:20:35.5385886Z * [new branch] gh/laithsakka/313/base -> origin/gh/laithsakka/313/base 2025-12-04T09:20:35.5387618Z * [new branch] gh/laithsakka/313/head -> origin/gh/laithsakka/313/head 2025-12-04T09:20:35.5389362Z * [new branch] gh/laithsakka/313/orig -> origin/gh/laithsakka/313/orig 2025-12-04T09:20:35.5391896Z * [new branch] gh/laithsakka/316/base -> origin/gh/laithsakka/316/base 2025-12-04T09:20:35.5393537Z * [new branch] gh/laithsakka/316/head -> origin/gh/laithsakka/316/head 2025-12-04T09:20:35.5395293Z * [new branch] gh/laithsakka/316/orig -> origin/gh/laithsakka/316/orig 2025-12-04T09:20:35.5398155Z * [new branch] gh/laithsakka/317/base -> origin/gh/laithsakka/317/base 2025-12-04T09:20:35.5399857Z * [new branch] gh/laithsakka/317/head -> origin/gh/laithsakka/317/head 2025-12-04T09:20:35.5401688Z * [new branch] gh/laithsakka/317/orig -> origin/gh/laithsakka/317/orig 2025-12-04T09:20:35.5404031Z * [new branch] gh/laithsakka/319/base -> origin/gh/laithsakka/319/base 2025-12-04T09:20:35.5405778Z * [new branch] gh/laithsakka/319/head -> origin/gh/laithsakka/319/head 2025-12-04T09:20:35.5407512Z * [new branch] gh/laithsakka/319/orig -> origin/gh/laithsakka/319/orig 2025-12-04T09:20:35.5409712Z * [new branch] gh/laithsakka/32/base -> origin/gh/laithsakka/32/base 2025-12-04T09:20:35.5411346Z * [new branch] gh/laithsakka/32/head -> origin/gh/laithsakka/32/head 2025-12-04T09:20:35.5413810Z * [new branch] gh/laithsakka/320/base -> origin/gh/laithsakka/320/base 2025-12-04T09:20:35.5415914Z * [new branch] gh/laithsakka/320/head -> origin/gh/laithsakka/320/head 2025-12-04T09:20:35.5417727Z * [new branch] gh/laithsakka/320/orig -> origin/gh/laithsakka/320/orig 2025-12-04T09:20:35.5420181Z * [new branch] gh/laithsakka/321/base -> origin/gh/laithsakka/321/base 2025-12-04T09:20:35.5422145Z * [new branch] gh/laithsakka/321/head -> origin/gh/laithsakka/321/head 2025-12-04T09:20:35.5423690Z * [new branch] gh/laithsakka/321/orig -> origin/gh/laithsakka/321/orig 2025-12-04T09:20:35.5426267Z * [new branch] gh/laithsakka/322/base -> origin/gh/laithsakka/322/base 2025-12-04T09:20:35.5428077Z * [new branch] gh/laithsakka/322/head -> origin/gh/laithsakka/322/head 2025-12-04T09:20:35.5429751Z * [new branch] gh/laithsakka/322/orig -> origin/gh/laithsakka/322/orig 2025-12-04T09:20:35.5432102Z * [new branch] gh/laithsakka/323/base -> origin/gh/laithsakka/323/base 2025-12-04T09:20:35.5433926Z * [new branch] gh/laithsakka/323/head -> origin/gh/laithsakka/323/head 2025-12-04T09:20:35.5435637Z * [new branch] gh/laithsakka/323/orig -> origin/gh/laithsakka/323/orig 2025-12-04T09:20:35.5438117Z * [new branch] gh/laithsakka/324/base -> origin/gh/laithsakka/324/base 2025-12-04T09:20:35.5439760Z * [new branch] gh/laithsakka/324/head -> origin/gh/laithsakka/324/head 2025-12-04T09:20:35.5441549Z * [new branch] gh/laithsakka/324/orig -> origin/gh/laithsakka/324/orig 2025-12-04T09:20:35.5444073Z * [new branch] gh/laithsakka/325/base -> origin/gh/laithsakka/325/base 2025-12-04T09:20:35.5445878Z * [new branch] gh/laithsakka/325/head -> origin/gh/laithsakka/325/head 2025-12-04T09:20:35.5447600Z * [new branch] gh/laithsakka/325/orig -> origin/gh/laithsakka/325/orig 2025-12-04T09:20:35.5450517Z * [new branch] gh/laithsakka/326/base -> origin/gh/laithsakka/326/base 2025-12-04T09:20:35.5452265Z * [new branch] gh/laithsakka/326/head -> origin/gh/laithsakka/326/head 2025-12-04T09:20:35.5454059Z * [new branch] gh/laithsakka/326/orig -> origin/gh/laithsakka/326/orig 2025-12-04T09:20:35.5456554Z * [new branch] gh/laithsakka/327/base -> origin/gh/laithsakka/327/base 2025-12-04T09:20:35.5458298Z * [new branch] gh/laithsakka/327/head -> origin/gh/laithsakka/327/head 2025-12-04T09:20:35.5460146Z * [new branch] gh/laithsakka/327/orig -> origin/gh/laithsakka/327/orig 2025-12-04T09:20:35.5462514Z * [new branch] gh/laithsakka/328/base -> origin/gh/laithsakka/328/base 2025-12-04T09:20:35.5464229Z * [new branch] gh/laithsakka/328/head -> origin/gh/laithsakka/328/head 2025-12-04T09:20:35.5465945Z * [new branch] gh/laithsakka/328/orig -> origin/gh/laithsakka/328/orig 2025-12-04T09:20:35.5468748Z * [new branch] gh/liangel/4/base -> origin/gh/liangel/4/base 2025-12-04T09:20:35.5470513Z * [new branch] gh/liangel/4/head -> origin/gh/liangel/4/head 2025-12-04T09:20:35.5472559Z * [new branch] gh/liangel/4/orig -> origin/gh/liangel/4/orig 2025-12-04T09:20:35.5476864Z * [new branch] gh/lucaskabela/1/base -> origin/gh/lucaskabela/1/base 2025-12-04T09:20:35.5478623Z * [new branch] gh/lucaskabela/1/head -> origin/gh/lucaskabela/1/head 2025-12-04T09:20:35.5481383Z * [new branch] gh/lw/4/base -> origin/gh/lw/4/base 2025-12-04T09:20:35.5483188Z * [new branch] gh/lw/4/head -> origin/gh/lw/4/head 2025-12-04T09:20:35.5484909Z * [new branch] gh/lw/4/orig -> origin/gh/lw/4/orig 2025-12-04T09:20:35.5487232Z * [new branch] gh/lw/5/base -> origin/gh/lw/5/base 2025-12-04T09:20:35.5488976Z * [new branch] gh/lw/5/head -> origin/gh/lw/5/head 2025-12-04T09:20:35.5490732Z * [new branch] gh/lw/5/orig -> origin/gh/lw/5/orig 2025-12-04T09:20:35.5492980Z * [new branch] gh/lw/6/base -> origin/gh/lw/6/base 2025-12-04T09:20:35.5494953Z * [new branch] gh/lw/6/head -> origin/gh/lw/6/head 2025-12-04T09:20:35.5496572Z * [new branch] gh/lw/6/orig -> origin/gh/lw/6/orig 2025-12-04T09:20:35.5499841Z * [new branch] gh/malfet/14/base -> origin/gh/malfet/14/base 2025-12-04T09:20:35.5502128Z * [new branch] gh/malfet/417/base -> origin/gh/malfet/417/base 2025-12-04T09:20:35.5504001Z * [new branch] gh/malfet/417/head -> origin/gh/malfet/417/head 2025-12-04T09:20:35.5505707Z * [new branch] gh/malfet/417/orig -> origin/gh/malfet/417/orig 2025-12-04T09:20:35.5508025Z * [new branch] gh/malfet/506/base -> origin/gh/malfet/506/base 2025-12-04T09:20:35.5510006Z * [new branch] gh/malfet/506/head -> origin/gh/malfet/506/head 2025-12-04T09:20:35.5511698Z * [new branch] gh/malfet/506/orig -> origin/gh/malfet/506/orig 2025-12-04T09:20:35.5514048Z * [new branch] gh/malfet/517/base -> origin/gh/malfet/517/base 2025-12-04T09:20:35.5515741Z * [new branch] gh/malfet/517/head -> origin/gh/malfet/517/head 2025-12-04T09:20:35.5518140Z * [new branch] gh/malfet/528/base -> origin/gh/malfet/528/base 2025-12-04T09:20:35.5520309Z * [new branch] gh/malfet/528/head -> origin/gh/malfet/528/head 2025-12-04T09:20:35.5522095Z * [new branch] gh/malfet/528/orig -> origin/gh/malfet/528/orig 2025-12-04T09:20:35.5524460Z * [new branch] gh/malfet/537/base -> origin/gh/malfet/537/base 2025-12-04T09:20:35.5526218Z * [new branch] gh/malfet/537/head -> origin/gh/malfet/537/head 2025-12-04T09:20:35.5528063Z * [new branch] gh/malfet/537/orig -> origin/gh/malfet/537/orig 2025-12-04T09:20:35.5530403Z * [new branch] gh/malfet/546/base -> origin/gh/malfet/546/base 2025-12-04T09:20:35.5532123Z * [new branch] gh/malfet/546/head -> origin/gh/malfet/546/head 2025-12-04T09:20:35.5533802Z * [new branch] gh/malfet/546/orig -> origin/gh/malfet/546/orig 2025-12-04T09:20:35.5536115Z * [new branch] gh/malfet/565/base -> origin/gh/malfet/565/base 2025-12-04T09:20:35.5537866Z * [new branch] gh/malfet/565/head -> origin/gh/malfet/565/head 2025-12-04T09:20:35.5539681Z * [new branch] gh/malfet/565/orig -> origin/gh/malfet/565/orig 2025-12-04T09:20:35.5542132Z * [new branch] gh/malfet/575/base -> origin/gh/malfet/575/base 2025-12-04T09:20:35.5543894Z * [new branch] gh/malfet/575/head -> origin/gh/malfet/575/head 2025-12-04T09:20:35.5545628Z * [new branch] gh/malfet/575/orig -> origin/gh/malfet/575/orig 2025-12-04T09:20:35.5547961Z * [new branch] gh/malfet/580/base -> origin/gh/malfet/580/base 2025-12-04T09:20:35.5549704Z * [new branch] gh/malfet/580/head -> origin/gh/malfet/580/head 2025-12-04T09:20:35.5551463Z * [new branch] gh/malfet/580/orig -> origin/gh/malfet/580/orig 2025-12-04T09:20:35.5553886Z * [new branch] gh/malfet/581/base -> origin/gh/malfet/581/base 2025-12-04T09:20:35.5555575Z * [new branch] gh/malfet/581/head -> origin/gh/malfet/581/head 2025-12-04T09:20:35.5557286Z * [new branch] gh/malfet/581/orig -> origin/gh/malfet/581/orig 2025-12-04T09:20:35.5560048Z * [new branch] gh/malfet/583/base -> origin/gh/malfet/583/base 2025-12-04T09:20:35.5561674Z * [new branch] gh/malfet/583/head -> origin/gh/malfet/583/head 2025-12-04T09:20:35.5563394Z * [new branch] gh/malfet/583/orig -> origin/gh/malfet/583/orig 2025-12-04T09:20:35.5565805Z * [new branch] gh/malfet/586/base -> origin/gh/malfet/586/base 2025-12-04T09:20:35.5567721Z * [new branch] gh/malfet/586/head -> origin/gh/malfet/586/head 2025-12-04T09:20:35.5569323Z * [new branch] gh/malfet/586/orig -> origin/gh/malfet/586/orig 2025-12-04T09:20:35.5571675Z * [new branch] gh/malfet/587/base -> origin/gh/malfet/587/base 2025-12-04T09:20:35.5573570Z * [new branch] gh/malfet/587/head -> origin/gh/malfet/587/head 2025-12-04T09:20:35.5575345Z * [new branch] gh/malfet/587/orig -> origin/gh/malfet/587/orig 2025-12-04T09:20:35.5577603Z * [new branch] gh/malfet/588/base -> origin/gh/malfet/588/base 2025-12-04T09:20:35.5579435Z * [new branch] gh/malfet/588/head -> origin/gh/malfet/588/head 2025-12-04T09:20:35.5581274Z * [new branch] gh/malfet/588/orig -> origin/gh/malfet/588/orig 2025-12-04T09:20:35.5584182Z * [new branch] gh/malfet/589/base -> origin/gh/malfet/589/base 2025-12-04T09:20:35.5585948Z * [new branch] gh/malfet/589/head -> origin/gh/malfet/589/head 2025-12-04T09:20:35.5587652Z * [new branch] gh/malfet/589/orig -> origin/gh/malfet/589/orig 2025-12-04T09:20:35.5590116Z * [new branch] gh/malfet/590/base -> origin/gh/malfet/590/base 2025-12-04T09:20:35.5591960Z * [new branch] gh/malfet/590/head -> origin/gh/malfet/590/head 2025-12-04T09:20:35.5593716Z * [new branch] gh/malfet/590/orig -> origin/gh/malfet/590/orig 2025-12-04T09:20:35.5596462Z * [new branch] gh/malfet/591/base -> origin/gh/malfet/591/base 2025-12-04T09:20:35.5598212Z * [new branch] gh/malfet/591/head -> origin/gh/malfet/591/head 2025-12-04T09:20:35.5600034Z * [new branch] gh/malfet/591/orig -> origin/gh/malfet/591/orig 2025-12-04T09:20:35.5602313Z * [new branch] gh/malfet/592/base -> origin/gh/malfet/592/base 2025-12-04T09:20:35.5604116Z * [new branch] gh/malfet/592/head -> origin/gh/malfet/592/head 2025-12-04T09:20:35.5605857Z * [new branch] gh/malfet/592/orig -> origin/gh/malfet/592/orig 2025-12-04T09:20:35.5608322Z * [new branch] gh/malfet/593/base -> origin/gh/malfet/593/base 2025-12-04T09:20:35.5609972Z * [new branch] gh/malfet/593/head -> origin/gh/malfet/593/head 2025-12-04T09:20:35.5611647Z * [new branch] gh/malfet/593/orig -> origin/gh/malfet/593/orig 2025-12-04T09:20:35.5614169Z * [new branch] gh/malfet/594/base -> origin/gh/malfet/594/base 2025-12-04T09:20:35.5615911Z * [new branch] gh/malfet/594/head -> origin/gh/malfet/594/head 2025-12-04T09:20:35.5617647Z * [new branch] gh/malfet/594/orig -> origin/gh/malfet/594/orig 2025-12-04T09:20:35.5620097Z * [new branch] gh/malfet/595/base -> origin/gh/malfet/595/base 2025-12-04T09:20:35.5621833Z * [new branch] gh/malfet/595/head -> origin/gh/malfet/595/head 2025-12-04T09:20:35.5623620Z * [new branch] gh/malfet/595/orig -> origin/gh/malfet/595/orig 2025-12-04T09:20:35.5626439Z * [new branch] gh/malfet/596/base -> origin/gh/malfet/596/base 2025-12-04T09:20:35.5628153Z * [new branch] gh/malfet/596/head -> origin/gh/malfet/596/head 2025-12-04T09:20:35.5629915Z * [new branch] gh/malfet/596/orig -> origin/gh/malfet/596/orig 2025-12-04T09:20:35.5632328Z * [new branch] gh/malfet/597/base -> origin/gh/malfet/597/base 2025-12-04T09:20:35.5634044Z * [new branch] gh/malfet/597/head -> origin/gh/malfet/597/head 2025-12-04T09:20:35.5635769Z * [new branch] gh/malfet/597/orig -> origin/gh/malfet/597/orig 2025-12-04T09:20:35.5638263Z * [new branch] gh/malfet/598/base -> origin/gh/malfet/598/base 2025-12-04T09:20:35.5640204Z * [new branch] gh/malfet/598/head -> origin/gh/malfet/598/head 2025-12-04T09:20:35.5641785Z * [new branch] gh/malfet/598/orig -> origin/gh/malfet/598/orig 2025-12-04T09:20:35.5644238Z * [new branch] gh/malfet/599/base -> origin/gh/malfet/599/base 2025-12-04T09:20:35.5645979Z * [new branch] gh/malfet/599/head -> origin/gh/malfet/599/head 2025-12-04T09:20:35.5647751Z * [new branch] gh/malfet/599/orig -> origin/gh/malfet/599/orig 2025-12-04T09:20:35.5650166Z * [new branch] gh/malfet/600/base -> origin/gh/malfet/600/base 2025-12-04T09:20:35.5651997Z * [new branch] gh/malfet/600/head -> origin/gh/malfet/600/head 2025-12-04T09:20:35.5653717Z * [new branch] gh/malfet/600/orig -> origin/gh/malfet/600/orig 2025-12-04T09:20:35.5656270Z * [new branch] gh/malfet/601/base -> origin/gh/malfet/601/base 2025-12-04T09:20:35.5658013Z * [new branch] gh/malfet/601/head -> origin/gh/malfet/601/head 2025-12-04T09:20:35.5659837Z * [new branch] gh/malfet/601/orig -> origin/gh/malfet/601/orig 2025-12-04T09:20:35.5662437Z * [new branch] gh/malfet/602/base -> origin/gh/malfet/602/base 2025-12-04T09:20:35.5664155Z * [new branch] gh/malfet/602/head -> origin/gh/malfet/602/head 2025-12-04T09:20:35.5665848Z * [new branch] gh/malfet/602/orig -> origin/gh/malfet/602/orig 2025-12-04T09:20:35.5668628Z * [new branch] gh/malfet/603/base -> origin/gh/malfet/603/base 2025-12-04T09:20:35.5670287Z * [new branch] gh/malfet/603/head -> origin/gh/malfet/603/head 2025-12-04T09:20:35.5672256Z * [new branch] gh/malfet/603/orig -> origin/gh/malfet/603/orig 2025-12-04T09:20:35.5674602Z * [new branch] gh/malfet/604/base -> origin/gh/malfet/604/base 2025-12-04T09:20:35.5676396Z * [new branch] gh/malfet/604/head -> origin/gh/malfet/604/head 2025-12-04T09:20:35.5678085Z * [new branch] gh/malfet/604/orig -> origin/gh/malfet/604/orig 2025-12-04T09:20:35.5680506Z * [new branch] gh/malfet/605/base -> origin/gh/malfet/605/base 2025-12-04T09:20:35.5682333Z * [new branch] gh/malfet/605/head -> origin/gh/malfet/605/head 2025-12-04T09:20:35.5684161Z * [new branch] gh/malfet/605/orig -> origin/gh/malfet/605/orig 2025-12-04T09:20:35.5686656Z * [new branch] gh/malfet/606/base -> origin/gh/malfet/606/base 2025-12-04T09:20:35.5688417Z * [new branch] gh/malfet/606/head -> origin/gh/malfet/606/head 2025-12-04T09:20:35.5690178Z * [new branch] gh/malfet/606/orig -> origin/gh/malfet/606/orig 2025-12-04T09:20:35.5692550Z * [new branch] gh/malfet/607/base -> origin/gh/malfet/607/base 2025-12-04T09:20:35.5694353Z * [new branch] gh/malfet/607/head -> origin/gh/malfet/607/head 2025-12-04T09:20:35.5696135Z * [new branch] gh/malfet/607/orig -> origin/gh/malfet/607/orig 2025-12-04T09:20:35.5698503Z * [new branch] gh/malfet/608/base -> origin/gh/malfet/608/base 2025-12-04T09:20:35.5700394Z * [new branch] gh/malfet/608/head -> origin/gh/malfet/608/head 2025-12-04T09:20:35.5702017Z * [new branch] gh/malfet/608/orig -> origin/gh/malfet/608/orig 2025-12-04T09:20:35.5704457Z * [new branch] gh/malfet/609/base -> origin/gh/malfet/609/base 2025-12-04T09:20:35.5706351Z * [new branch] gh/malfet/609/head -> origin/gh/malfet/609/head 2025-12-04T09:20:35.5708126Z * [new branch] gh/malfet/609/orig -> origin/gh/malfet/609/orig 2025-12-04T09:20:35.5710788Z * [new branch] gh/malfet/610/base -> origin/gh/malfet/610/base 2025-12-04T09:20:35.5712458Z * [new branch] gh/malfet/610/head -> origin/gh/malfet/610/head 2025-12-04T09:20:35.5714183Z * [new branch] gh/malfet/610/orig -> origin/gh/malfet/610/orig 2025-12-04T09:20:35.5716767Z * [new branch] gh/malfet/611/base -> origin/gh/malfet/611/base 2025-12-04T09:20:35.5718510Z * [new branch] gh/malfet/611/head -> origin/gh/malfet/611/head 2025-12-04T09:20:35.5720231Z * [new branch] gh/malfet/611/orig -> origin/gh/malfet/611/orig 2025-12-04T09:20:35.5722446Z * [new branch] gh/malfet/612/base -> origin/gh/malfet/612/base 2025-12-04T09:20:35.5724218Z * [new branch] gh/malfet/612/head -> origin/gh/malfet/612/head 2025-12-04T09:20:35.5725931Z * [new branch] gh/malfet/612/orig -> origin/gh/malfet/612/orig 2025-12-04T09:20:35.5728346Z * [new branch] gh/malfet/64/base -> origin/gh/malfet/64/base 2025-12-04T09:20:35.5730061Z * [new branch] gh/malfet/64/head -> origin/gh/malfet/64/head 2025-12-04T09:20:35.5732950Z * [new branch] gh/manuelcandales/11/base -> origin/gh/manuelcandales/11/base 2025-12-04T09:20:35.5734694Z * [new branch] gh/manuelcandales/11/head -> origin/gh/manuelcandales/11/head 2025-12-04T09:20:35.5736457Z * [new branch] gh/manuelcandales/11/orig -> origin/gh/manuelcandales/11/orig 2025-12-04T09:20:35.5739460Z * [new branch] gh/markkm/1/base -> origin/gh/markkm/1/base 2025-12-04T09:20:35.5742301Z * [new branch] gh/masnesral/1/base -> origin/gh/masnesral/1/base 2025-12-04T09:20:35.5744563Z * [new branch] gh/masnesral/1/head -> origin/gh/masnesral/1/head 2025-12-04T09:20:35.5746327Z * [new branch] gh/masnesral/1/orig -> origin/gh/masnesral/1/orig 2025-12-04T09:20:35.5749720Z * [new branch] gh/mhorowitz/0/base -> origin/gh/mhorowitz/0/base 2025-12-04T09:20:35.5751593Z * [new branch] gh/mhorowitz/0/head -> origin/gh/mhorowitz/0/head 2025-12-04T09:20:35.5754265Z * [new branch] gh/mhorowitz/1/base -> origin/gh/mhorowitz/1/base 2025-12-04T09:20:35.5756019Z * [new branch] gh/mhorowitz/1/head -> origin/gh/mhorowitz/1/head 2025-12-04T09:20:35.5758235Z * [new branch] gh/mhorowitz/2/base -> origin/gh/mhorowitz/2/base 2025-12-04T09:20:35.5759991Z * [new branch] gh/mhorowitz/2/head -> origin/gh/mhorowitz/2/head 2025-12-04T09:20:35.5762147Z * [new branch] gh/mhorowitz/3/base -> origin/gh/mhorowitz/3/base 2025-12-04T09:20:35.5764450Z * [new branch] gh/mhorowitz/3/head -> origin/gh/mhorowitz/3/head 2025-12-04T09:20:35.5766738Z * [new branch] gh/mhorowitz/4/base -> origin/gh/mhorowitz/4/base 2025-12-04T09:20:35.5768414Z * [new branch] gh/mhorowitz/4/head -> origin/gh/mhorowitz/4/head 2025-12-04T09:20:35.5770614Z * [new branch] gh/mhorowitz/5/base -> origin/gh/mhorowitz/5/base 2025-12-04T09:20:35.5773935Z * [new branch] gh/mhorowitz/5/head -> origin/gh/mhorowitz/5/head 2025-12-04T09:20:35.5776239Z * [new branch] gh/mhorowitz/6/base -> origin/gh/mhorowitz/6/base 2025-12-04T09:20:35.5777965Z * [new branch] gh/mhorowitz/6/head -> origin/gh/mhorowitz/6/head 2025-12-04T09:20:35.5780959Z * [new branch] gh/mikaylagawarecki/234/base -> origin/gh/mikaylagawarecki/234/base 2025-12-04T09:20:35.5782780Z * [new branch] gh/mikaylagawarecki/234/head -> origin/gh/mikaylagawarecki/234/head 2025-12-04T09:20:35.5785123Z * [new branch] gh/mikaylagawarecki/235/base -> origin/gh/mikaylagawarecki/235/base 2025-12-04T09:20:35.5787049Z * [new branch] gh/mikaylagawarecki/235/head -> origin/gh/mikaylagawarecki/235/head 2025-12-04T09:20:35.5789172Z * [new branch] gh/mikaylagawarecki/236/base -> origin/gh/mikaylagawarecki/236/base 2025-12-04T09:20:35.5790808Z * [new branch] gh/mikaylagawarecki/236/head -> origin/gh/mikaylagawarecki/236/head 2025-12-04T09:20:35.5792967Z * [new branch] gh/mikaylagawarecki/237/base -> origin/gh/mikaylagawarecki/237/base 2025-12-04T09:20:35.5794757Z * [new branch] gh/mikaylagawarecki/237/head -> origin/gh/mikaylagawarecki/237/head 2025-12-04T09:20:35.5797080Z * [new branch] gh/mikaylagawarecki/238/base -> origin/gh/mikaylagawarecki/238/base 2025-12-04T09:20:35.5798881Z * [new branch] gh/mikaylagawarecki/238/head -> origin/gh/mikaylagawarecki/238/head 2025-12-04T09:20:35.5801202Z * [new branch] gh/mikaylagawarecki/336/base -> origin/gh/mikaylagawarecki/336/base 2025-12-04T09:20:35.5802973Z * [new branch] gh/mikaylagawarecki/336/head -> origin/gh/mikaylagawarecki/336/head 2025-12-04T09:20:35.5804662Z * [new branch] gh/mikaylagawarecki/336/orig -> origin/gh/mikaylagawarecki/336/orig 2025-12-04T09:20:35.5807060Z * [new branch] gh/mikaylagawarecki/341/base -> origin/gh/mikaylagawarecki/341/base 2025-12-04T09:20:35.5808741Z * [new branch] gh/mikaylagawarecki/341/head -> origin/gh/mikaylagawarecki/341/head 2025-12-04T09:20:35.5810451Z * [new branch] gh/mikaylagawarecki/341/orig -> origin/gh/mikaylagawarecki/341/orig 2025-12-04T09:20:35.5812930Z * [new branch] gh/mikaylagawarecki/342/base -> origin/gh/mikaylagawarecki/342/base 2025-12-04T09:20:35.5814685Z * [new branch] gh/mikaylagawarecki/342/head -> origin/gh/mikaylagawarecki/342/head 2025-12-04T09:20:35.5816539Z * [new branch] gh/mikaylagawarecki/342/orig -> origin/gh/mikaylagawarecki/342/orig 2025-12-04T09:20:35.5818985Z * [new branch] gh/mikaylagawarecki/345/base -> origin/gh/mikaylagawarecki/345/base 2025-12-04T09:20:35.5820950Z * [new branch] gh/mikaylagawarecki/345/head -> origin/gh/mikaylagawarecki/345/head 2025-12-04T09:20:35.5822636Z * [new branch] gh/mikaylagawarecki/345/orig -> origin/gh/mikaylagawarecki/345/orig 2025-12-04T09:20:35.5825153Z * [new branch] gh/mikaylagawarecki/346/base -> origin/gh/mikaylagawarecki/346/base 2025-12-04T09:20:35.5826914Z * [new branch] gh/mikaylagawarecki/346/head -> origin/gh/mikaylagawarecki/346/head 2025-12-04T09:20:35.5828722Z * [new branch] gh/mikaylagawarecki/346/orig -> origin/gh/mikaylagawarecki/346/orig 2025-12-04T09:20:35.5831108Z * [new branch] gh/mikaylagawarecki/347/base -> origin/gh/mikaylagawarecki/347/base 2025-12-04T09:20:35.5832797Z * [new branch] gh/mikaylagawarecki/347/head -> origin/gh/mikaylagawarecki/347/head 2025-12-04T09:20:35.5834495Z * [new branch] gh/mikaylagawarecki/347/orig -> origin/gh/mikaylagawarecki/347/orig 2025-12-04T09:20:35.5836927Z * [new branch] gh/mikaylagawarecki/350/base -> origin/gh/mikaylagawarecki/350/base 2025-12-04T09:20:35.5838695Z * [new branch] gh/mikaylagawarecki/350/head -> origin/gh/mikaylagawarecki/350/head 2025-12-04T09:20:35.5840474Z * [new branch] gh/mikaylagawarecki/350/orig -> origin/gh/mikaylagawarecki/350/orig 2025-12-04T09:20:35.5843087Z * [new branch] gh/mikaylagawarecki/351/base -> origin/gh/mikaylagawarecki/351/base 2025-12-04T09:20:35.5844966Z * [new branch] gh/mikaylagawarecki/351/head -> origin/gh/mikaylagawarecki/351/head 2025-12-04T09:20:35.5846742Z * [new branch] gh/mikaylagawarecki/351/orig -> origin/gh/mikaylagawarecki/351/orig 2025-12-04T09:20:35.5849289Z * [new branch] gh/mikaylagawarecki/352/base -> origin/gh/mikaylagawarecki/352/base 2025-12-04T09:20:35.5851069Z * [new branch] gh/mikaylagawarecki/352/head -> origin/gh/mikaylagawarecki/352/head 2025-12-04T09:20:35.5852869Z * [new branch] gh/mikaylagawarecki/352/orig -> origin/gh/mikaylagawarecki/352/orig 2025-12-04T09:20:35.5855334Z * [new branch] gh/mikaylagawarecki/353/base -> origin/gh/mikaylagawarecki/353/base 2025-12-04T09:20:35.5857358Z * [new branch] gh/mikaylagawarecki/353/head -> origin/gh/mikaylagawarecki/353/head 2025-12-04T09:20:35.5859090Z * [new branch] gh/mikaylagawarecki/353/orig -> origin/gh/mikaylagawarecki/353/orig 2025-12-04T09:20:35.5861433Z * [new branch] gh/mikaylagawarecki/354/base -> origin/gh/mikaylagawarecki/354/base 2025-12-04T09:20:35.5863130Z * [new branch] gh/mikaylagawarecki/354/head -> origin/gh/mikaylagawarecki/354/head 2025-12-04T09:20:35.5864947Z * [new branch] gh/mikaylagawarecki/354/orig -> origin/gh/mikaylagawarecki/354/orig 2025-12-04T09:20:35.5867736Z * [new branch] gh/mikaylagawarecki/356/base -> origin/gh/mikaylagawarecki/356/base 2025-12-04T09:20:35.5869597Z * [new branch] gh/mikaylagawarecki/356/head -> origin/gh/mikaylagawarecki/356/head 2025-12-04T09:20:35.5871567Z * [new branch] gh/mikaylagawarecki/356/orig -> origin/gh/mikaylagawarecki/356/orig 2025-12-04T09:20:35.5873912Z * [new branch] gh/mikaylagawarecki/357/base -> origin/gh/mikaylagawarecki/357/base 2025-12-04T09:20:35.5875593Z * [new branch] gh/mikaylagawarecki/357/head -> origin/gh/mikaylagawarecki/357/head 2025-12-04T09:20:35.5877382Z * [new branch] gh/mikaylagawarecki/357/orig -> origin/gh/mikaylagawarecki/357/orig 2025-12-04T09:20:35.5879926Z * [new branch] gh/mikaylagawarecki/359/base -> origin/gh/mikaylagawarecki/359/base 2025-12-04T09:20:35.5881805Z * [new branch] gh/mikaylagawarecki/359/head -> origin/gh/mikaylagawarecki/359/head 2025-12-04T09:20:35.5883509Z * [new branch] gh/mikaylagawarecki/359/orig -> origin/gh/mikaylagawarecki/359/orig 2025-12-04T09:20:35.5885958Z * [new branch] gh/mikaylagawarecki/360/base -> origin/gh/mikaylagawarecki/360/base 2025-12-04T09:20:35.5887774Z * [new branch] gh/mikaylagawarecki/360/head -> origin/gh/mikaylagawarecki/360/head 2025-12-04T09:20:35.5889528Z * [new branch] gh/mikaylagawarecki/360/orig -> origin/gh/mikaylagawarecki/360/orig 2025-12-04T09:20:35.5891919Z * [new branch] gh/mikaylagawarecki/361/base -> origin/gh/mikaylagawarecki/361/base 2025-12-04T09:20:35.5893741Z * [new branch] gh/mikaylagawarecki/361/head -> origin/gh/mikaylagawarecki/361/head 2025-12-04T09:20:35.5895454Z * [new branch] gh/mikaylagawarecki/361/orig -> origin/gh/mikaylagawarecki/361/orig 2025-12-04T09:20:35.5897906Z * [new branch] gh/mikaylagawarecki/362/base -> origin/gh/mikaylagawarecki/362/base 2025-12-04T09:20:35.5899899Z * [new branch] gh/mikaylagawarecki/362/head -> origin/gh/mikaylagawarecki/362/head 2025-12-04T09:20:35.5901681Z * [new branch] gh/mikaylagawarecki/362/orig -> origin/gh/mikaylagawarecki/362/orig 2025-12-04T09:20:35.5904245Z * [new branch] gh/mikaylagawarecki/363/base -> origin/gh/mikaylagawarecki/363/base 2025-12-04T09:20:35.5906303Z * [new branch] gh/mikaylagawarecki/363/head -> origin/gh/mikaylagawarecki/363/head 2025-12-04T09:20:35.5907998Z * [new branch] gh/mikaylagawarecki/363/orig -> origin/gh/mikaylagawarecki/363/orig 2025-12-04T09:20:35.5910740Z * [new branch] gh/mikaylagawarecki/364/base -> origin/gh/mikaylagawarecki/364/base 2025-12-04T09:20:35.5912651Z * [new branch] gh/mikaylagawarecki/364/head -> origin/gh/mikaylagawarecki/364/head 2025-12-04T09:20:35.5914376Z * [new branch] gh/mikaylagawarecki/364/orig -> origin/gh/mikaylagawarecki/364/orig 2025-12-04T09:20:35.5916946Z * [new branch] gh/mikaylagawarecki/365/base -> origin/gh/mikaylagawarecki/365/base 2025-12-04T09:20:35.5918886Z * [new branch] gh/mikaylagawarecki/365/head -> origin/gh/mikaylagawarecki/365/head 2025-12-04T09:20:35.5920713Z * [new branch] gh/mikaylagawarecki/365/orig -> origin/gh/mikaylagawarecki/365/orig 2025-12-04T09:20:35.5922980Z * [new branch] gh/mikaylagawarecki/366/base -> origin/gh/mikaylagawarecki/366/base 2025-12-04T09:20:35.5924683Z * [new branch] gh/mikaylagawarecki/366/head -> origin/gh/mikaylagawarecki/366/head 2025-12-04T09:20:35.5926432Z * [new branch] gh/mikaylagawarecki/366/orig -> origin/gh/mikaylagawarecki/366/orig 2025-12-04T09:20:35.5928993Z * [new branch] gh/mikaylagawarecki/367/base -> origin/gh/mikaylagawarecki/367/base 2025-12-04T09:20:35.5930880Z * [new branch] gh/mikaylagawarecki/367/head -> origin/gh/mikaylagawarecki/367/head 2025-12-04T09:20:35.5932594Z * [new branch] gh/mikaylagawarecki/367/orig -> origin/gh/mikaylagawarecki/367/orig 2025-12-04T09:20:35.5935105Z * [new branch] gh/mikaylagawarecki/368/base -> origin/gh/mikaylagawarecki/368/base 2025-12-04T09:20:35.5936886Z * [new branch] gh/mikaylagawarecki/368/head -> origin/gh/mikaylagawarecki/368/head 2025-12-04T09:20:35.5938667Z * [new branch] gh/mikaylagawarecki/368/orig -> origin/gh/mikaylagawarecki/368/orig 2025-12-04T09:20:35.5941221Z * [new branch] gh/mikaylagawarecki/369/base -> origin/gh/mikaylagawarecki/369/base 2025-12-04T09:20:35.5943182Z * [new branch] gh/mikaylagawarecki/369/head -> origin/gh/mikaylagawarecki/369/head 2025-12-04T09:20:35.5944944Z * [new branch] gh/mikaylagawarecki/369/orig -> origin/gh/mikaylagawarecki/369/orig 2025-12-04T09:20:35.5947401Z * [new branch] gh/mikaylagawarecki/370/base -> origin/gh/mikaylagawarecki/370/base 2025-12-04T09:20:35.5949251Z * [new branch] gh/mikaylagawarecki/370/head -> origin/gh/mikaylagawarecki/370/head 2025-12-04T09:20:35.5950986Z * [new branch] gh/mikaylagawarecki/370/orig -> origin/gh/mikaylagawarecki/370/orig 2025-12-04T09:20:35.5953424Z * [new branch] gh/mikaylagawarecki/371/base -> origin/gh/mikaylagawarecki/371/base 2025-12-04T09:20:35.5955103Z * [new branch] gh/mikaylagawarecki/371/head -> origin/gh/mikaylagawarecki/371/head 2025-12-04T09:20:35.5956851Z * [new branch] gh/mikaylagawarecki/371/orig -> origin/gh/mikaylagawarecki/371/orig 2025-12-04T09:20:35.5959321Z * [new branch] gh/mikaylagawarecki/372/base -> origin/gh/mikaylagawarecki/372/base 2025-12-04T09:20:35.5961097Z * [new branch] gh/mikaylagawarecki/372/head -> origin/gh/mikaylagawarecki/372/head 2025-12-04T09:20:35.5962823Z * [new branch] gh/mikaylagawarecki/372/orig -> origin/gh/mikaylagawarecki/372/orig 2025-12-04T09:20:35.5965164Z * [new branch] gh/mikaylagawarecki/373/base -> origin/gh/mikaylagawarecki/373/base 2025-12-04T09:20:35.5966996Z * [new branch] gh/mikaylagawarecki/373/head -> origin/gh/mikaylagawarecki/373/head 2025-12-04T09:20:35.5968711Z * [new branch] gh/mikaylagawarecki/373/orig -> origin/gh/mikaylagawarecki/373/orig 2025-12-04T09:20:35.5971318Z * [new branch] gh/mikaylagawarecki/374/base -> origin/gh/mikaylagawarecki/374/base 2025-12-04T09:20:35.5973149Z * [new branch] gh/mikaylagawarecki/374/head -> origin/gh/mikaylagawarecki/374/head 2025-12-04T09:20:35.5974846Z * [new branch] gh/mikaylagawarecki/374/orig -> origin/gh/mikaylagawarecki/374/orig 2025-12-04T09:20:35.5977203Z * [new branch] gh/mikaylagawarecki/375/base -> origin/gh/mikaylagawarecki/375/base 2025-12-04T09:20:35.5979019Z * [new branch] gh/mikaylagawarecki/375/head -> origin/gh/mikaylagawarecki/375/head 2025-12-04T09:20:35.5980930Z * [new branch] gh/mikaylagawarecki/375/orig -> origin/gh/mikaylagawarecki/375/orig 2025-12-04T09:20:35.5983365Z * [new branch] gh/mikaylagawarecki/376/base -> origin/gh/mikaylagawarecki/376/base 2025-12-04T09:20:35.5985262Z * [new branch] gh/mikaylagawarecki/376/head -> origin/gh/mikaylagawarecki/376/head 2025-12-04T09:20:35.5986990Z * [new branch] gh/mikaylagawarecki/376/orig -> origin/gh/mikaylagawarecki/376/orig 2025-12-04T09:20:35.5989374Z * [new branch] gh/mikaylagawarecki/377/base -> origin/gh/mikaylagawarecki/377/base 2025-12-04T09:20:35.5991321Z * [new branch] gh/mikaylagawarecki/377/head -> origin/gh/mikaylagawarecki/377/head 2025-12-04T09:20:35.5992984Z * [new branch] gh/mikaylagawarecki/377/orig -> origin/gh/mikaylagawarecki/377/orig 2025-12-04T09:20:35.5995415Z * [new branch] gh/mikaylagawarecki/378/base -> origin/gh/mikaylagawarecki/378/base 2025-12-04T09:20:35.5997246Z * [new branch] gh/mikaylagawarecki/378/head -> origin/gh/mikaylagawarecki/378/head 2025-12-04T09:20:35.5999030Z * [new branch] gh/mikaylagawarecki/378/orig -> origin/gh/mikaylagawarecki/378/orig 2025-12-04T09:20:35.6001414Z * [new branch] gh/mikaylagawarecki/379/base -> origin/gh/mikaylagawarecki/379/base 2025-12-04T09:20:35.6003143Z * [new branch] gh/mikaylagawarecki/379/head -> origin/gh/mikaylagawarecki/379/head 2025-12-04T09:20:35.6004836Z * [new branch] gh/mikaylagawarecki/379/orig -> origin/gh/mikaylagawarecki/379/orig 2025-12-04T09:20:35.6007158Z * [new branch] gh/mikaylagawarecki/380/base -> origin/gh/mikaylagawarecki/380/base 2025-12-04T09:20:35.6008843Z * [new branch] gh/mikaylagawarecki/380/head -> origin/gh/mikaylagawarecki/380/head 2025-12-04T09:20:35.6010547Z * [new branch] gh/mikaylagawarecki/380/orig -> origin/gh/mikaylagawarecki/380/orig 2025-12-04T09:20:35.6013315Z * [new branch] gh/mikaylagawarecki/381/base -> origin/gh/mikaylagawarecki/381/base 2025-12-04T09:20:35.6015141Z * [new branch] gh/mikaylagawarecki/381/head -> origin/gh/mikaylagawarecki/381/head 2025-12-04T09:20:35.6016895Z * [new branch] gh/mikaylagawarecki/381/orig -> origin/gh/mikaylagawarecki/381/orig 2025-12-04T09:20:35.6019207Z * [new branch] gh/mikaylagawarecki/382/base -> origin/gh/mikaylagawarecki/382/base 2025-12-04T09:20:35.6020978Z * [new branch] gh/mikaylagawarecki/382/head -> origin/gh/mikaylagawarecki/382/head 2025-12-04T09:20:35.6022701Z * [new branch] gh/mikaylagawarecki/382/orig -> origin/gh/mikaylagawarecki/382/orig 2025-12-04T09:20:35.6025188Z * [new branch] gh/mikaylagawarecki/383/base -> origin/gh/mikaylagawarecki/383/base 2025-12-04T09:20:35.6027018Z * [new branch] gh/mikaylagawarecki/383/head -> origin/gh/mikaylagawarecki/383/head 2025-12-04T09:20:35.6028843Z * [new branch] gh/mikaylagawarecki/383/orig -> origin/gh/mikaylagawarecki/383/orig 2025-12-04T09:20:35.6031207Z * [new branch] gh/mikaylagawarecki/384/base -> origin/gh/mikaylagawarecki/384/base 2025-12-04T09:20:35.6032971Z * [new branch] gh/mikaylagawarecki/384/head -> origin/gh/mikaylagawarecki/384/head 2025-12-04T09:20:35.6034707Z * [new branch] gh/mikaylagawarecki/384/orig -> origin/gh/mikaylagawarecki/384/orig 2025-12-04T09:20:35.6037176Z * [new branch] gh/mikaylagawarecki/385/base -> origin/gh/mikaylagawarecki/385/base 2025-12-04T09:20:35.6039007Z * [new branch] gh/mikaylagawarecki/385/head -> origin/gh/mikaylagawarecki/385/head 2025-12-04T09:20:35.6040769Z * [new branch] gh/mikaylagawarecki/385/orig -> origin/gh/mikaylagawarecki/385/orig 2025-12-04T09:20:35.6043385Z * [new branch] gh/mikaylagawarecki/386/base -> origin/gh/mikaylagawarecki/386/base 2025-12-04T09:20:35.6045127Z * [new branch] gh/mikaylagawarecki/386/head -> origin/gh/mikaylagawarecki/386/head 2025-12-04T09:20:35.6046868Z * [new branch] gh/mikaylagawarecki/386/orig -> origin/gh/mikaylagawarecki/386/orig 2025-12-04T09:20:35.6049421Z * [new branch] gh/mikaylagawarecki/387/base -> origin/gh/mikaylagawarecki/387/base 2025-12-04T09:20:35.6051104Z * [new branch] gh/mikaylagawarecki/387/head -> origin/gh/mikaylagawarecki/387/head 2025-12-04T09:20:35.6052878Z * [new branch] gh/mikaylagawarecki/387/orig -> origin/gh/mikaylagawarecki/387/orig 2025-12-04T09:20:35.6055131Z * [new branch] gh/mikaylagawarecki/388/base -> origin/gh/mikaylagawarecki/388/base 2025-12-04T09:20:35.6056922Z * [new branch] gh/mikaylagawarecki/388/head -> origin/gh/mikaylagawarecki/388/head 2025-12-04T09:20:35.6058670Z * [new branch] gh/mikaylagawarecki/388/orig -> origin/gh/mikaylagawarecki/388/orig 2025-12-04T09:20:35.6061270Z * [new branch] gh/mikaylagawarecki/389/base -> origin/gh/mikaylagawarecki/389/base 2025-12-04T09:20:35.6062919Z * [new branch] gh/mikaylagawarecki/389/head -> origin/gh/mikaylagawarecki/389/head 2025-12-04T09:20:35.6064640Z * [new branch] gh/mikaylagawarecki/389/orig -> origin/gh/mikaylagawarecki/389/orig 2025-12-04T09:20:35.6067159Z * [new branch] gh/mikaylagawarecki/390/base -> origin/gh/mikaylagawarecki/390/base 2025-12-04T09:20:35.6068816Z * [new branch] gh/mikaylagawarecki/390/head -> origin/gh/mikaylagawarecki/390/head 2025-12-04T09:20:35.6070489Z * [new branch] gh/mikaylagawarecki/390/orig -> origin/gh/mikaylagawarecki/390/orig 2025-12-04T09:20:35.6073287Z * [new branch] gh/mikaylagawarecki/391/base -> origin/gh/mikaylagawarecki/391/base 2025-12-04T09:20:35.6074983Z * [new branch] gh/mikaylagawarecki/391/head -> origin/gh/mikaylagawarecki/391/head 2025-12-04T09:20:35.6076731Z * [new branch] gh/mikaylagawarecki/391/orig -> origin/gh/mikaylagawarecki/391/orig 2025-12-04T09:20:35.6079206Z * [new branch] gh/mikaylagawarecki/392/base -> origin/gh/mikaylagawarecki/392/base 2025-12-04T09:20:35.6080951Z * [new branch] gh/mikaylagawarecki/392/head -> origin/gh/mikaylagawarecki/392/head 2025-12-04T09:20:35.6082729Z * [new branch] gh/mikaylagawarecki/392/orig -> origin/gh/mikaylagawarecki/392/orig 2025-12-04T09:20:35.6085625Z * [new branch] gh/mlazos/41/base -> origin/gh/mlazos/41/base 2025-12-04T09:20:35.6087340Z * [new branch] gh/mlazos/41/head -> origin/gh/mlazos/41/head 2025-12-04T09:20:35.6089055Z * [new branch] gh/mlazos/41/orig -> origin/gh/mlazos/41/orig 2025-12-04T09:20:35.6091458Z * [new branch] gh/mlazos/42/base -> origin/gh/mlazos/42/base 2025-12-04T09:20:35.6093147Z * [new branch] gh/mlazos/42/head -> origin/gh/mlazos/42/head 2025-12-04T09:20:35.6094834Z * [new branch] gh/mlazos/42/orig -> origin/gh/mlazos/42/orig 2025-12-04T09:20:35.6097041Z * [new branch] gh/mlazos/43/base -> origin/gh/mlazos/43/base 2025-12-04T09:20:35.6098918Z * [new branch] gh/mlazos/43/head -> origin/gh/mlazos/43/head 2025-12-04T09:20:35.6100789Z * [new branch] gh/mlazos/43/orig -> origin/gh/mlazos/43/orig 2025-12-04T09:20:35.6103853Z * [new branch] gh/mlazos/44/base -> origin/gh/mlazos/44/base 2025-12-04T09:20:35.6105626Z * [new branch] gh/mlazos/44/head -> origin/gh/mlazos/44/head 2025-12-04T09:20:35.6107376Z * [new branch] gh/mlazos/44/orig -> origin/gh/mlazos/44/orig 2025-12-04T09:20:35.6109811Z * [new branch] gh/mlazos/47/base -> origin/gh/mlazos/47/base 2025-12-04T09:20:35.6111522Z * [new branch] gh/mlazos/47/head -> origin/gh/mlazos/47/head 2025-12-04T09:20:35.6113427Z * [new branch] gh/mlazos/47/orig -> origin/gh/mlazos/47/orig 2025-12-04T09:20:35.6115685Z * [new branch] gh/mlazos/48/base -> origin/gh/mlazos/48/base 2025-12-04T09:20:35.6117725Z * [new branch] gh/mlazos/48/head -> origin/gh/mlazos/48/head 2025-12-04T09:20:35.6119046Z * [new branch] gh/mlazos/48/orig -> origin/gh/mlazos/48/orig 2025-12-04T09:20:35.6121450Z * [new branch] gh/mlazos/49/base -> origin/gh/mlazos/49/base 2025-12-04T09:20:35.6123128Z * [new branch] gh/mlazos/49/head -> origin/gh/mlazos/49/head 2025-12-04T09:20:35.6124778Z * [new branch] gh/mlazos/49/orig -> origin/gh/mlazos/49/orig 2025-12-04T09:20:35.6127078Z * [new branch] gh/mlazos/50/base -> origin/gh/mlazos/50/base 2025-12-04T09:20:35.6128779Z * [new branch] gh/mlazos/50/head -> origin/gh/mlazos/50/head 2025-12-04T09:20:35.6130623Z * [new branch] gh/mlazos/50/orig -> origin/gh/mlazos/50/orig 2025-12-04T09:20:35.6132895Z * [new branch] gh/mlazos/51/base -> origin/gh/mlazos/51/base 2025-12-04T09:20:35.6134735Z * [new branch] gh/mlazos/51/head -> origin/gh/mlazos/51/head 2025-12-04T09:20:35.6136441Z * [new branch] gh/mlazos/51/orig -> origin/gh/mlazos/51/orig 2025-12-04T09:20:35.6138868Z * [new branch] gh/mlazos/52/base -> origin/gh/mlazos/52/base 2025-12-04T09:20:35.6140671Z * [new branch] gh/mlazos/52/head -> origin/gh/mlazos/52/head 2025-12-04T09:20:35.6142376Z * [new branch] gh/mlazos/52/orig -> origin/gh/mlazos/52/orig 2025-12-04T09:20:35.6144737Z * [new branch] gh/mlazos/53/base -> origin/gh/mlazos/53/base 2025-12-04T09:20:35.6146563Z * [new branch] gh/mlazos/53/head -> origin/gh/mlazos/53/head 2025-12-04T09:20:35.6148221Z * [new branch] gh/mlazos/53/orig -> origin/gh/mlazos/53/orig 2025-12-04T09:20:35.6150501Z * [new branch] gh/mlazos/54/base -> origin/gh/mlazos/54/base 2025-12-04T09:20:35.6152192Z * [new branch] gh/mlazos/54/head -> origin/gh/mlazos/54/head 2025-12-04T09:20:35.6154008Z * [new branch] gh/mlazos/54/orig -> origin/gh/mlazos/54/orig 2025-12-04T09:20:35.6156360Z * [new branch] gh/mlazos/55/base -> origin/gh/mlazos/55/base 2025-12-04T09:20:35.6158669Z * [new branch] gh/mlazos/55/head -> origin/gh/mlazos/55/head 2025-12-04T09:20:35.6160416Z * [new branch] gh/mlazos/55/orig -> origin/gh/mlazos/55/orig 2025-12-04T09:20:35.6162858Z * [new branch] gh/mlazos/56/base -> origin/gh/mlazos/56/base 2025-12-04T09:20:35.6164571Z * [new branch] gh/mlazos/56/head -> origin/gh/mlazos/56/head 2025-12-04T09:20:35.6166324Z * [new branch] gh/mlazos/56/orig -> origin/gh/mlazos/56/orig 2025-12-04T09:20:35.6168644Z * [new branch] gh/mlazos/57/base -> origin/gh/mlazos/57/base 2025-12-04T09:20:35.6170389Z * [new branch] gh/mlazos/57/head -> origin/gh/mlazos/57/head 2025-12-04T09:20:35.6173789Z * [new branch] gh/mlazos/57/orig -> origin/gh/mlazos/57/orig 2025-12-04T09:20:35.6176403Z * [new branch] gh/mlazos/58/base -> origin/gh/mlazos/58/base 2025-12-04T09:20:35.6178077Z * [new branch] gh/mlazos/58/head -> origin/gh/mlazos/58/head 2025-12-04T09:20:35.6179961Z * [new branch] gh/mlazos/58/orig -> origin/gh/mlazos/58/orig 2025-12-04T09:20:35.6182369Z * [new branch] gh/mlazos/59/base -> origin/gh/mlazos/59/base 2025-12-04T09:20:35.6184075Z * [new branch] gh/mlazos/59/head -> origin/gh/mlazos/59/head 2025-12-04T09:20:35.6185774Z * [new branch] gh/mlazos/59/orig -> origin/gh/mlazos/59/orig 2025-12-04T09:20:35.6188152Z * [new branch] gh/mlazos/60/base -> origin/gh/mlazos/60/base 2025-12-04T09:20:35.6190058Z * [new branch] gh/mlazos/60/head -> origin/gh/mlazos/60/head 2025-12-04T09:20:35.6191655Z * [new branch] gh/mlazos/60/orig -> origin/gh/mlazos/60/orig 2025-12-04T09:20:35.6194441Z * [new branch] gh/mlazos/61/base -> origin/gh/mlazos/61/base 2025-12-04T09:20:35.6197068Z * [new branch] gh/mlazos/61/head -> origin/gh/mlazos/61/head 2025-12-04T09:20:35.6198070Z * [new branch] gh/mlazos/61/orig -> origin/gh/mlazos/61/orig 2025-12-04T09:20:35.6200220Z * [new branch] gh/mlazos/62/base -> origin/gh/mlazos/62/base 2025-12-04T09:20:35.6201866Z * [new branch] gh/mlazos/62/head -> origin/gh/mlazos/62/head 2025-12-04T09:20:35.6203679Z * [new branch] gh/mlazos/62/orig -> origin/gh/mlazos/62/orig 2025-12-04T09:20:35.6206316Z * [new branch] gh/mlazos/63/base -> origin/gh/mlazos/63/base 2025-12-04T09:20:35.6208080Z * [new branch] gh/mlazos/63/head -> origin/gh/mlazos/63/head 2025-12-04T09:20:35.6209805Z * [new branch] gh/mlazos/63/orig -> origin/gh/mlazos/63/orig 2025-12-04T09:20:35.6212111Z * [new branch] gh/mlazos/64/base -> origin/gh/mlazos/64/base 2025-12-04T09:20:35.6213833Z * [new branch] gh/mlazos/64/head -> origin/gh/mlazos/64/head 2025-12-04T09:20:35.6215581Z * [new branch] gh/mlazos/64/orig -> origin/gh/mlazos/64/orig 2025-12-04T09:20:35.6218108Z * [new branch] gh/mlazos/65/base -> origin/gh/mlazos/65/base 2025-12-04T09:20:35.6219987Z * [new branch] gh/mlazos/65/head -> origin/gh/mlazos/65/head 2025-12-04T09:20:35.6221740Z * [new branch] gh/mlazos/65/orig -> origin/gh/mlazos/65/orig 2025-12-04T09:20:35.6224113Z * [new branch] gh/mlazos/66/base -> origin/gh/mlazos/66/base 2025-12-04T09:20:35.6225813Z * [new branch] gh/mlazos/66/head -> origin/gh/mlazos/66/head 2025-12-04T09:20:35.6227518Z * [new branch] gh/mlazos/66/orig -> origin/gh/mlazos/66/orig 2025-12-04T09:20:35.6229964Z * [new branch] gh/mlazos/67/base -> origin/gh/mlazos/67/base 2025-12-04T09:20:35.6231740Z * [new branch] gh/mlazos/67/head -> origin/gh/mlazos/67/head 2025-12-04T09:20:35.6233498Z * [new branch] gh/mlazos/67/orig -> origin/gh/mlazos/67/orig 2025-12-04T09:20:35.6235885Z * [new branch] gh/mlazos/68/base -> origin/gh/mlazos/68/base 2025-12-04T09:20:35.6237576Z * [new branch] gh/mlazos/68/head -> origin/gh/mlazos/68/head 2025-12-04T09:20:35.6239330Z * [new branch] gh/mlazos/68/orig -> origin/gh/mlazos/68/orig 2025-12-04T09:20:35.6241659Z * [new branch] gh/mlazos/69/base -> origin/gh/mlazos/69/base 2025-12-04T09:20:35.6243433Z * [new branch] gh/mlazos/69/head -> origin/gh/mlazos/69/head 2025-12-04T09:20:35.6245191Z * [new branch] gh/mlazos/69/orig -> origin/gh/mlazos/69/orig 2025-12-04T09:20:35.6247646Z * [new branch] gh/mlazos/70/base -> origin/gh/mlazos/70/base 2025-12-04T09:20:35.6249314Z * [new branch] gh/mlazos/70/head -> origin/gh/mlazos/70/head 2025-12-04T09:20:35.6251074Z * [new branch] gh/mlazos/70/orig -> origin/gh/mlazos/70/orig 2025-12-04T09:20:35.6253488Z * [new branch] gh/mlazos/71/base -> origin/gh/mlazos/71/base 2025-12-04T09:20:35.6255260Z * [new branch] gh/mlazos/71/head -> origin/gh/mlazos/71/head 2025-12-04T09:20:35.6256975Z * [new branch] gh/mlazos/71/orig -> origin/gh/mlazos/71/orig 2025-12-04T09:20:35.6259348Z * [new branch] gh/mlazos/72/base -> origin/gh/mlazos/72/base 2025-12-04T09:20:35.6261244Z * [new branch] gh/mlazos/72/head -> origin/gh/mlazos/72/head 2025-12-04T09:20:35.6262981Z * [new branch] gh/mlazos/72/orig -> origin/gh/mlazos/72/orig 2025-12-04T09:20:35.6265390Z * [new branch] gh/mlazos/73/base -> origin/gh/mlazos/73/base 2025-12-04T09:20:35.6267164Z * [new branch] gh/mlazos/73/head -> origin/gh/mlazos/73/head 2025-12-04T09:20:35.6268909Z * [new branch] gh/mlazos/73/orig -> origin/gh/mlazos/73/orig 2025-12-04T09:20:35.6272086Z * [new branch] gh/mrmiywj/1/base -> origin/gh/mrmiywj/1/base 2025-12-04T09:20:35.6274010Z * [new branch] gh/mrmiywj/1/head -> origin/gh/mrmiywj/1/head 2025-12-04T09:20:35.6276950Z * [new branch] gh/muchulee8/73/base -> origin/gh/muchulee8/73/base 2025-12-04T09:20:35.6278814Z * [new branch] gh/muchulee8/73/head -> origin/gh/muchulee8/73/head 2025-12-04T09:20:35.6280648Z * [new branch] gh/muchulee8/73/orig -> origin/gh/muchulee8/73/orig 2025-12-04T09:20:35.6283485Z * [new branch] gh/naveenthangudu/1/base -> origin/gh/naveenthangudu/1/base 2025-12-04T09:20:35.6285255Z * [new branch] gh/naveenthangudu/1/head -> origin/gh/naveenthangudu/1/head 2025-12-04T09:20:35.6287110Z * [new branch] gh/naveenthangudu/1/orig -> origin/gh/naveenthangudu/1/orig 2025-12-04T09:20:35.6289434Z * [new branch] gh/naveenthangudu/2/base -> origin/gh/naveenthangudu/2/base 2025-12-04T09:20:35.6291126Z * [new branch] gh/naveenthangudu/2/head -> origin/gh/naveenthangudu/2/head 2025-12-04T09:20:35.6292859Z * [new branch] gh/naveenthangudu/2/orig -> origin/gh/naveenthangudu/2/orig 2025-12-04T09:20:35.6295129Z * [new branch] gh/naveenthangudu/3/base -> origin/gh/naveenthangudu/3/base 2025-12-04T09:20:35.6296847Z * [new branch] gh/naveenthangudu/3/head -> origin/gh/naveenthangudu/3/head 2025-12-04T09:20:35.6298662Z * [new branch] gh/naveenthangudu/3/orig -> origin/gh/naveenthangudu/3/orig 2025-12-04T09:20:35.6301136Z * [new branch] gh/naveenthangudu/4/base -> origin/gh/naveenthangudu/4/base 2025-12-04T09:20:35.6302952Z * [new branch] gh/naveenthangudu/4/head -> origin/gh/naveenthangudu/4/head 2025-12-04T09:20:35.6304905Z * [new branch] gh/naveenthangudu/4/orig -> origin/gh/naveenthangudu/4/orig 2025-12-04T09:20:35.6307205Z * [new branch] gh/naveenthangudu/5/base -> origin/gh/naveenthangudu/5/base 2025-12-04T09:20:35.6308941Z * [new branch] gh/naveenthangudu/5/head -> origin/gh/naveenthangudu/5/head 2025-12-04T09:20:35.6310821Z * [new branch] gh/naveenthangudu/5/orig -> origin/gh/naveenthangudu/5/orig 2025-12-04T09:20:35.6313225Z * [new branch] gh/naveenthangudu/6/base -> origin/gh/naveenthangudu/6/base 2025-12-04T09:20:35.6314961Z * [new branch] gh/naveenthangudu/6/head -> origin/gh/naveenthangudu/6/head 2025-12-04T09:20:35.6316640Z * [new branch] gh/naveenthangudu/6/orig -> origin/gh/naveenthangudu/6/orig 2025-12-04T09:20:35.6318955Z * [new branch] gh/naveenthangudu/7/base -> origin/gh/naveenthangudu/7/base 2025-12-04T09:20:35.6320701Z * [new branch] gh/naveenthangudu/7/head -> origin/gh/naveenthangudu/7/head 2025-12-04T09:20:35.6322446Z * [new branch] gh/naveenthangudu/7/orig -> origin/gh/naveenthangudu/7/orig 2025-12-04T09:20:35.6324677Z * [new branch] gh/naveenthangudu/8/base -> origin/gh/naveenthangudu/8/base 2025-12-04T09:20:35.6326583Z * [new branch] gh/naveenthangudu/8/head -> origin/gh/naveenthangudu/8/head 2025-12-04T09:20:35.6328329Z * [new branch] gh/naveenthangudu/8/orig -> origin/gh/naveenthangudu/8/orig 2025-12-04T09:20:35.6330846Z * [new branch] gh/naveenthangudu/9/base -> origin/gh/naveenthangudu/9/base 2025-12-04T09:20:35.6332482Z * [new branch] gh/naveenthangudu/9/head -> origin/gh/naveenthangudu/9/head 2025-12-04T09:20:35.6334383Z * [new branch] gh/naveenthangudu/9/orig -> origin/gh/naveenthangudu/9/orig 2025-12-04T09:20:35.6337133Z * [new branch] gh/nikitaved/1/base -> origin/gh/nikitaved/1/base 2025-12-04T09:20:35.6338886Z * [new branch] gh/nikitaved/1/head -> origin/gh/nikitaved/1/head 2025-12-04T09:20:35.6340753Z * [new branch] gh/nikitaved/1/orig -> origin/gh/nikitaved/1/orig 2025-12-04T09:20:35.6343248Z * [new branch] gh/nikitaved/10/base -> origin/gh/nikitaved/10/base 2025-12-04T09:20:35.6345043Z * [new branch] gh/nikitaved/10/head -> origin/gh/nikitaved/10/head 2025-12-04T09:20:35.6346718Z * [new branch] gh/nikitaved/10/orig -> origin/gh/nikitaved/10/orig 2025-12-04T09:20:35.6348975Z * [new branch] gh/nikitaved/11/base -> origin/gh/nikitaved/11/base 2025-12-04T09:20:35.6350810Z * [new branch] gh/nikitaved/11/head -> origin/gh/nikitaved/11/head 2025-12-04T09:20:35.6352491Z * [new branch] gh/nikitaved/11/orig -> origin/gh/nikitaved/11/orig 2025-12-04T09:20:35.6354769Z * [new branch] gh/nikitaved/12/base -> origin/gh/nikitaved/12/base 2025-12-04T09:20:35.6356535Z * [new branch] gh/nikitaved/12/head -> origin/gh/nikitaved/12/head 2025-12-04T09:20:35.6358315Z * [new branch] gh/nikitaved/12/orig -> origin/gh/nikitaved/12/orig 2025-12-04T09:20:35.6360641Z * [new branch] gh/nikitaved/13/base -> origin/gh/nikitaved/13/base 2025-12-04T09:20:35.6362489Z * [new branch] gh/nikitaved/13/head -> origin/gh/nikitaved/13/head 2025-12-04T09:20:35.6364258Z * [new branch] gh/nikitaved/13/orig -> origin/gh/nikitaved/13/orig 2025-12-04T09:20:35.6366673Z * [new branch] gh/nikitaved/14/base -> origin/gh/nikitaved/14/base 2025-12-04T09:20:35.6368507Z * [new branch] gh/nikitaved/14/head -> origin/gh/nikitaved/14/head 2025-12-04T09:20:35.6370207Z * [new branch] gh/nikitaved/14/orig -> origin/gh/nikitaved/14/orig 2025-12-04T09:20:35.6373905Z * [new branch] gh/nikitaved/15/base -> origin/gh/nikitaved/15/base 2025-12-04T09:20:35.6375825Z * [new branch] gh/nikitaved/15/head -> origin/gh/nikitaved/15/head 2025-12-04T09:20:35.6377506Z * [new branch] gh/nikitaved/15/orig -> origin/gh/nikitaved/15/orig 2025-12-04T09:20:35.6379944Z * [new branch] gh/nikitaved/16/base -> origin/gh/nikitaved/16/base 2025-12-04T09:20:35.6381694Z * [new branch] gh/nikitaved/16/head -> origin/gh/nikitaved/16/head 2025-12-04T09:20:35.6383338Z * [new branch] gh/nikitaved/16/orig -> origin/gh/nikitaved/16/orig 2025-12-04T09:20:35.6385682Z * [new branch] gh/nikitaved/2/base -> origin/gh/nikitaved/2/base 2025-12-04T09:20:35.6387419Z * [new branch] gh/nikitaved/2/head -> origin/gh/nikitaved/2/head 2025-12-04T09:20:35.6389211Z * [new branch] gh/nikitaved/2/orig -> origin/gh/nikitaved/2/orig 2025-12-04T09:20:35.6391432Z * [new branch] gh/nikitaved/4/base -> origin/gh/nikitaved/4/base 2025-12-04T09:20:35.6393313Z * [new branch] gh/nikitaved/4/head -> origin/gh/nikitaved/4/head 2025-12-04T09:20:35.6395071Z * [new branch] gh/nikitaved/4/orig -> origin/gh/nikitaved/4/orig 2025-12-04T09:20:35.6397428Z * [new branch] gh/nikitaved/5/base -> origin/gh/nikitaved/5/base 2025-12-04T09:20:35.6399196Z * [new branch] gh/nikitaved/5/head -> origin/gh/nikitaved/5/head 2025-12-04T09:20:35.6401147Z * [new branch] gh/nikitaved/5/orig -> origin/gh/nikitaved/5/orig 2025-12-04T09:20:35.6403286Z * [new branch] gh/nikitaved/6/base -> origin/gh/nikitaved/6/base 2025-12-04T09:20:35.6404992Z * [new branch] gh/nikitaved/6/head -> origin/gh/nikitaved/6/head 2025-12-04T09:20:35.6406688Z * [new branch] gh/nikitaved/6/orig -> origin/gh/nikitaved/6/orig 2025-12-04T09:20:35.6409063Z * [new branch] gh/nikitaved/8/base -> origin/gh/nikitaved/8/base 2025-12-04T09:20:35.6410768Z * [new branch] gh/nikitaved/8/head -> origin/gh/nikitaved/8/head 2025-12-04T09:20:35.6412921Z * [new branch] gh/nikitaved/8/orig -> origin/gh/nikitaved/8/orig 2025-12-04T09:20:35.6415305Z * [new branch] gh/nikitaved/9/base -> origin/gh/nikitaved/9/base 2025-12-04T09:20:35.6417043Z * [new branch] gh/nikitaved/9/head -> origin/gh/nikitaved/9/head 2025-12-04T09:20:35.6418769Z * [new branch] gh/nikitaved/9/orig -> origin/gh/nikitaved/9/orig 2025-12-04T09:20:35.6421753Z * [new branch] gh/oulgen/10/base -> origin/gh/oulgen/10/base 2025-12-04T09:20:35.6423472Z * [new branch] gh/oulgen/10/head -> origin/gh/oulgen/10/head 2025-12-04T09:20:35.6425139Z * [new branch] gh/oulgen/10/orig -> origin/gh/oulgen/10/orig 2025-12-04T09:20:35.6427448Z * [new branch] gh/oulgen/11/base -> origin/gh/oulgen/11/base 2025-12-04T09:20:35.6429174Z * [new branch] gh/oulgen/11/head -> origin/gh/oulgen/11/head 2025-12-04T09:20:35.6430895Z * [new branch] gh/oulgen/11/orig -> origin/gh/oulgen/11/orig 2025-12-04T09:20:35.6433183Z * [new branch] gh/oulgen/12/base -> origin/gh/oulgen/12/base 2025-12-04T09:20:35.6434887Z * [new branch] gh/oulgen/12/head -> origin/gh/oulgen/12/head 2025-12-04T09:20:35.6436647Z * [new branch] gh/oulgen/12/orig -> origin/gh/oulgen/12/orig 2025-12-04T09:20:35.6439073Z * [new branch] gh/oulgen/13/base -> origin/gh/oulgen/13/base 2025-12-04T09:20:35.6440819Z * [new branch] gh/oulgen/13/head -> origin/gh/oulgen/13/head 2025-12-04T09:20:35.6442543Z * [new branch] gh/oulgen/13/orig -> origin/gh/oulgen/13/orig 2025-12-04T09:20:35.6444875Z * [new branch] gh/oulgen/14/base -> origin/gh/oulgen/14/base 2025-12-04T09:20:35.6446955Z * [new branch] gh/oulgen/14/head -> origin/gh/oulgen/14/head 2025-12-04T09:20:35.6448445Z * [new branch] gh/oulgen/14/orig -> origin/gh/oulgen/14/orig 2025-12-04T09:20:35.6450795Z * [new branch] gh/oulgen/15/base -> origin/gh/oulgen/15/base 2025-12-04T09:20:35.6452502Z * [new branch] gh/oulgen/15/head -> origin/gh/oulgen/15/head 2025-12-04T09:20:35.6454374Z * [new branch] gh/oulgen/15/orig -> origin/gh/oulgen/15/orig 2025-12-04T09:20:35.6456572Z * [new branch] gh/oulgen/16/base -> origin/gh/oulgen/16/base 2025-12-04T09:20:35.6458280Z * [new branch] gh/oulgen/16/head -> origin/gh/oulgen/16/head 2025-12-04T09:20:35.6460161Z * [new branch] gh/oulgen/16/orig -> origin/gh/oulgen/16/orig 2025-12-04T09:20:35.6462725Z * [new branch] gh/oulgen/17/base -> origin/gh/oulgen/17/base 2025-12-04T09:20:35.6464267Z * [new branch] gh/oulgen/17/head -> origin/gh/oulgen/17/head 2025-12-04T09:20:35.6465919Z * [new branch] gh/oulgen/17/orig -> origin/gh/oulgen/17/orig 2025-12-04T09:20:35.6468280Z * [new branch] gh/oulgen/18/base -> origin/gh/oulgen/18/base 2025-12-04T09:20:35.6469984Z * [new branch] gh/oulgen/18/head -> origin/gh/oulgen/18/head 2025-12-04T09:20:35.6472117Z * [new branch] gh/oulgen/18/orig -> origin/gh/oulgen/18/orig 2025-12-04T09:20:35.6474253Z * [new branch] gh/oulgen/19/base -> origin/gh/oulgen/19/base 2025-12-04T09:20:35.6476044Z * [new branch] gh/oulgen/19/head -> origin/gh/oulgen/19/head 2025-12-04T09:20:35.6477810Z * [new branch] gh/oulgen/19/orig -> origin/gh/oulgen/19/orig 2025-12-04T09:20:35.6480139Z * [new branch] gh/oulgen/20/base -> origin/gh/oulgen/20/base 2025-12-04T09:20:35.6482256Z * [new branch] gh/oulgen/20/head -> origin/gh/oulgen/20/head 2025-12-04T09:20:35.6483996Z * [new branch] gh/oulgen/20/orig -> origin/gh/oulgen/20/orig 2025-12-04T09:20:35.6486362Z * [new branch] gh/oulgen/21/base -> origin/gh/oulgen/21/base 2025-12-04T09:20:35.6488093Z * [new branch] gh/oulgen/21/head -> origin/gh/oulgen/21/head 2025-12-04T09:20:35.6489805Z * [new branch] gh/oulgen/21/orig -> origin/gh/oulgen/21/orig 2025-12-04T09:20:35.6492118Z * [new branch] gh/oulgen/22/base -> origin/gh/oulgen/22/base 2025-12-04T09:20:35.6493875Z * [new branch] gh/oulgen/22/head -> origin/gh/oulgen/22/head 2025-12-04T09:20:35.6495605Z * [new branch] gh/oulgen/22/orig -> origin/gh/oulgen/22/orig 2025-12-04T09:20:35.6498370Z * [new branch] gh/oulgen/23/base -> origin/gh/oulgen/23/base 2025-12-04T09:20:35.6500178Z * [new branch] gh/oulgen/23/head -> origin/gh/oulgen/23/head 2025-12-04T09:20:35.6501903Z * [new branch] gh/oulgen/23/orig -> origin/gh/oulgen/23/orig 2025-12-04T09:20:35.6504692Z * [new branch] gh/oulgen/24/base -> origin/gh/oulgen/24/base 2025-12-04T09:20:35.6506516Z * [new branch] gh/oulgen/24/head -> origin/gh/oulgen/24/head 2025-12-04T09:20:35.6508223Z * [new branch] gh/oulgen/24/orig -> origin/gh/oulgen/24/orig 2025-12-04T09:20:35.6510563Z * [new branch] gh/oulgen/25/base -> origin/gh/oulgen/25/base 2025-12-04T09:20:35.6512245Z * [new branch] gh/oulgen/25/head -> origin/gh/oulgen/25/head 2025-12-04T09:20:35.6513992Z * [new branch] gh/oulgen/25/orig -> origin/gh/oulgen/25/orig 2025-12-04T09:20:35.6516315Z * [new branch] gh/oulgen/26/base -> origin/gh/oulgen/26/base 2025-12-04T09:20:35.6518057Z * [new branch] gh/oulgen/26/head -> origin/gh/oulgen/26/head 2025-12-04T09:20:35.6519777Z * [new branch] gh/oulgen/26/orig -> origin/gh/oulgen/26/orig 2025-12-04T09:20:35.6522020Z * [new branch] gh/oulgen/4/base -> origin/gh/oulgen/4/base 2025-12-04T09:20:35.6523731Z * [new branch] gh/oulgen/4/head -> origin/gh/oulgen/4/head 2025-12-04T09:20:35.6525439Z * [new branch] gh/oulgen/4/orig -> origin/gh/oulgen/4/orig 2025-12-04T09:20:35.6528211Z * [new branch] gh/oulgen/7/base -> origin/gh/oulgen/7/base 2025-12-04T09:20:35.6529954Z * [new branch] gh/oulgen/7/head -> origin/gh/oulgen/7/head 2025-12-04T09:20:35.6531673Z * [new branch] gh/oulgen/7/orig -> origin/gh/oulgen/7/orig 2025-12-04T09:20:35.6534189Z * [new branch] gh/oulgen/8/base -> origin/gh/oulgen/8/base 2025-12-04T09:20:35.6535946Z * [new branch] gh/oulgen/8/head -> origin/gh/oulgen/8/head 2025-12-04T09:20:35.6537653Z * [new branch] gh/oulgen/8/orig -> origin/gh/oulgen/8/orig 2025-12-04T09:20:35.6540086Z * [new branch] gh/oulgen/9/base -> origin/gh/oulgen/9/base 2025-12-04T09:20:35.6541849Z * [new branch] gh/oulgen/9/head -> origin/gh/oulgen/9/head 2025-12-04T09:20:35.6543783Z * [new branch] gh/oulgen/9/orig -> origin/gh/oulgen/9/orig 2025-12-04T09:20:35.6546168Z * [new branch] gh/patvig/mtia-serialization -> origin/gh/patvig/mtia-serialization 2025-12-04T09:20:35.6549136Z * [new branch] gh/pearu/108/base -> origin/gh/pearu/108/base 2025-12-04T09:20:35.6551367Z * [new branch] gh/pearu/108/head -> origin/gh/pearu/108/head 2025-12-04T09:20:35.6553168Z * [new branch] gh/pearu/108/orig -> origin/gh/pearu/108/orig 2025-12-04T09:20:35.6555452Z * [new branch] gh/pearu/109/base -> origin/gh/pearu/109/base 2025-12-04T09:20:35.6557226Z * [new branch] gh/pearu/109/head -> origin/gh/pearu/109/head 2025-12-04T09:20:35.6558988Z * [new branch] gh/pearu/109/orig -> origin/gh/pearu/109/orig 2025-12-04T09:20:35.6561354Z * [new branch] gh/pearu/110/base -> origin/gh/pearu/110/base 2025-12-04T09:20:35.6563168Z * [new branch] gh/pearu/110/head -> origin/gh/pearu/110/head 2025-12-04T09:20:35.6564895Z * [new branch] gh/pearu/110/orig -> origin/gh/pearu/110/orig 2025-12-04T09:20:35.6567311Z * [new branch] gh/pearu/111/base -> origin/gh/pearu/111/base 2025-12-04T09:20:35.6569528Z * [new branch] gh/pearu/111/head -> origin/gh/pearu/111/head 2025-12-04T09:20:35.6571464Z * [new branch] gh/pearu/111/orig -> origin/gh/pearu/111/orig 2025-12-04T09:20:35.6573894Z * [new branch] gh/pearu/112/base -> origin/gh/pearu/112/base 2025-12-04T09:20:35.6575693Z * [new branch] gh/pearu/112/head -> origin/gh/pearu/112/head 2025-12-04T09:20:35.6577486Z * [new branch] gh/pearu/112/orig -> origin/gh/pearu/112/orig 2025-12-04T09:20:35.6579817Z * [new branch] gh/pearu/115/base -> origin/gh/pearu/115/base 2025-12-04T09:20:35.6581630Z * [new branch] gh/pearu/115/head -> origin/gh/pearu/115/head 2025-12-04T09:20:35.6583335Z * [new branch] gh/pearu/115/orig -> origin/gh/pearu/115/orig 2025-12-04T09:20:35.6585649Z * [new branch] gh/pearu/116/base -> origin/gh/pearu/116/base 2025-12-04T09:20:35.6587284Z * [new branch] gh/pearu/116/head -> origin/gh/pearu/116/head 2025-12-04T09:20:35.6589022Z * [new branch] gh/pearu/116/orig -> origin/gh/pearu/116/orig 2025-12-04T09:20:35.6591771Z * [new branch] gh/pearu/117/base -> origin/gh/pearu/117/base 2025-12-04T09:20:35.6593567Z * [new branch] gh/pearu/117/head -> origin/gh/pearu/117/head 2025-12-04T09:20:35.6595303Z * [new branch] gh/pearu/117/orig -> origin/gh/pearu/117/orig 2025-12-04T09:20:35.6597643Z * [new branch] gh/pearu/118/base -> origin/gh/pearu/118/base 2025-12-04T09:20:35.6599386Z * [new branch] gh/pearu/118/head -> origin/gh/pearu/118/head 2025-12-04T09:20:35.6601093Z * [new branch] gh/pearu/118/orig -> origin/gh/pearu/118/orig 2025-12-04T09:20:35.6603419Z * [new branch] gh/pearu/119/base -> origin/gh/pearu/119/base 2025-12-04T09:20:35.6605222Z * [new branch] gh/pearu/119/head -> origin/gh/pearu/119/head 2025-12-04T09:20:35.6606900Z * [new branch] gh/pearu/119/orig -> origin/gh/pearu/119/orig 2025-12-04T09:20:35.6609278Z * [new branch] gh/pearu/139/base -> origin/gh/pearu/139/base 2025-12-04T09:20:35.6610996Z * [new branch] gh/pearu/139/head -> origin/gh/pearu/139/head 2025-12-04T09:20:35.6612757Z * [new branch] gh/pearu/139/orig -> origin/gh/pearu/139/orig 2025-12-04T09:20:35.6615120Z * [new branch] gh/pearu/140/base -> origin/gh/pearu/140/base 2025-12-04T09:20:35.6617459Z * [new branch] gh/pearu/140/head -> origin/gh/pearu/140/head 2025-12-04T09:20:35.6619227Z * [new branch] gh/pearu/140/orig -> origin/gh/pearu/140/orig 2025-12-04T09:20:35.6621686Z * [new branch] gh/pearu/142/base -> origin/gh/pearu/142/base 2025-12-04T09:20:35.6623372Z * [new branch] gh/pearu/142/head -> origin/gh/pearu/142/head 2025-12-04T09:20:35.6625066Z * [new branch] gh/pearu/142/orig -> origin/gh/pearu/142/orig 2025-12-04T09:20:35.6627416Z * [new branch] gh/pearu/143/base -> origin/gh/pearu/143/base 2025-12-04T09:20:35.6629217Z * [new branch] gh/pearu/143/head -> origin/gh/pearu/143/head 2025-12-04T09:20:35.6631139Z * [new branch] gh/pearu/143/orig -> origin/gh/pearu/143/orig 2025-12-04T09:20:35.6633344Z * [new branch] gh/pearu/147/base -> origin/gh/pearu/147/base 2025-12-04T09:20:35.6635147Z * [new branch] gh/pearu/147/head -> origin/gh/pearu/147/head 2025-12-04T09:20:35.6636874Z * [new branch] gh/pearu/147/orig -> origin/gh/pearu/147/orig 2025-12-04T09:20:35.6639219Z * [new branch] gh/pearu/149/base -> origin/gh/pearu/149/base 2025-12-04T09:20:35.6640915Z * [new branch] gh/pearu/149/head -> origin/gh/pearu/149/head 2025-12-04T09:20:35.6642623Z * [new branch] gh/pearu/149/orig -> origin/gh/pearu/149/orig 2025-12-04T09:20:35.6645862Z * [new branch] gh/pearu/150/base -> origin/gh/pearu/150/base 2025-12-04T09:20:35.6647678Z * [new branch] gh/pearu/150/head -> origin/gh/pearu/150/head 2025-12-04T09:20:35.6649434Z * [new branch] gh/pearu/150/orig -> origin/gh/pearu/150/orig 2025-12-04T09:20:35.6651940Z * [new branch] gh/pearu/151/base -> origin/gh/pearu/151/base 2025-12-04T09:20:35.6653713Z * [new branch] gh/pearu/151/head -> origin/gh/pearu/151/head 2025-12-04T09:20:35.6655434Z * [new branch] gh/pearu/151/orig -> origin/gh/pearu/151/orig 2025-12-04T09:20:35.6657807Z * [new branch] gh/pearu/152/base -> origin/gh/pearu/152/base 2025-12-04T09:20:35.6659805Z * [new branch] gh/pearu/152/head -> origin/gh/pearu/152/head 2025-12-04T09:20:35.6661364Z * [new branch] gh/pearu/152/orig -> origin/gh/pearu/152/orig 2025-12-04T09:20:35.6663871Z * [new branch] gh/pearu/153/base -> origin/gh/pearu/153/base 2025-12-04T09:20:35.6665542Z * [new branch] gh/pearu/153/head -> origin/gh/pearu/153/head 2025-12-04T09:20:35.6667253Z * [new branch] gh/pearu/153/orig -> origin/gh/pearu/153/orig 2025-12-04T09:20:35.6669657Z * [new branch] gh/pearu/154/base -> origin/gh/pearu/154/base 2025-12-04T09:20:35.6671441Z * [new branch] gh/pearu/154/head -> origin/gh/pearu/154/head 2025-12-04T09:20:35.6673325Z * [new branch] gh/pearu/154/orig -> origin/gh/pearu/154/orig 2025-12-04T09:20:35.6675698Z * [new branch] gh/pearu/155/base -> origin/gh/pearu/155/base 2025-12-04T09:20:35.6677470Z * [new branch] gh/pearu/155/head -> origin/gh/pearu/155/head 2025-12-04T09:20:35.6679260Z * [new branch] gh/pearu/155/orig -> origin/gh/pearu/155/orig 2025-12-04T09:20:35.6681605Z * [new branch] gh/pearu/156/base -> origin/gh/pearu/156/base 2025-12-04T09:20:35.6683444Z * [new branch] gh/pearu/156/head -> origin/gh/pearu/156/head 2025-12-04T09:20:35.6685195Z * [new branch] gh/pearu/156/orig -> origin/gh/pearu/156/orig 2025-12-04T09:20:35.6687971Z * [new branch] gh/pearu/56/base -> origin/gh/pearu/56/base 2025-12-04T09:20:35.6690038Z * [new branch] gh/pearu/56/head -> origin/gh/pearu/56/head 2025-12-04T09:20:35.6691765Z * [new branch] gh/pearu/56/orig -> origin/gh/pearu/56/orig 2025-12-04T09:20:35.6694244Z * [new branch] gh/pearu/97/base -> origin/gh/pearu/97/base 2025-12-04T09:20:35.6696016Z * [new branch] gh/pearu/97/head -> origin/gh/pearu/97/head 2025-12-04T09:20:35.6698317Z * [new branch] gh/pearu/97/orig -> origin/gh/pearu/97/orig 2025-12-04T09:20:35.6701332Z * [new branch] gh/pianpwk/21/base -> origin/gh/pianpwk/21/base 2025-12-04T09:20:35.6703176Z * [new branch] gh/pianpwk/21/head -> origin/gh/pianpwk/21/head 2025-12-04T09:20:35.6705599Z * [new branch] gh/pianpwk/28/base -> origin/gh/pianpwk/28/base 2025-12-04T09:20:35.6707256Z * [new branch] gh/pianpwk/28/head -> origin/gh/pianpwk/28/head 2025-12-04T09:20:35.6709039Z * [new branch] gh/pianpwk/28/orig -> origin/gh/pianpwk/28/orig 2025-12-04T09:20:35.6711474Z * [new branch] gh/pianpwk/29/base -> origin/gh/pianpwk/29/base 2025-12-04T09:20:35.6713209Z * [new branch] gh/pianpwk/29/head -> origin/gh/pianpwk/29/head 2025-12-04T09:20:35.6714929Z * [new branch] gh/pianpwk/29/orig -> origin/gh/pianpwk/29/orig 2025-12-04T09:20:35.6717380Z * [new branch] gh/pianpwk/30/base -> origin/gh/pianpwk/30/base 2025-12-04T09:20:35.6719107Z * [new branch] gh/pianpwk/30/head -> origin/gh/pianpwk/30/head 2025-12-04T09:20:35.6720913Z * [new branch] gh/pianpwk/30/orig -> origin/gh/pianpwk/30/orig 2025-12-04T09:20:35.6723269Z * [new branch] gh/pianpwk/31/base -> origin/gh/pianpwk/31/base 2025-12-04T09:20:35.6725144Z * [new branch] gh/pianpwk/31/head -> origin/gh/pianpwk/31/head 2025-12-04T09:20:35.6726881Z * [new branch] gh/pianpwk/31/orig -> origin/gh/pianpwk/31/orig 2025-12-04T09:20:35.6729071Z * [new branch] gh/pianpwk/32/base -> origin/gh/pianpwk/32/base 2025-12-04T09:20:35.6730812Z * [new branch] gh/pianpwk/32/head -> origin/gh/pianpwk/32/head 2025-12-04T09:20:35.6732539Z * [new branch] gh/pianpwk/32/orig -> origin/gh/pianpwk/32/orig 2025-12-04T09:20:35.6734795Z * [new branch] gh/pianpwk/33/base -> origin/gh/pianpwk/33/base 2025-12-04T09:20:35.6737003Z * [new branch] gh/pianpwk/33/head -> origin/gh/pianpwk/33/head 2025-12-04T09:20:35.6738741Z * [new branch] gh/pianpwk/33/orig -> origin/gh/pianpwk/33/orig 2025-12-04T09:20:35.6741407Z * [new branch] gh/pianpwk/34/base -> origin/gh/pianpwk/34/base 2025-12-04T09:20:35.6743344Z * [new branch] gh/pianpwk/34/head -> origin/gh/pianpwk/34/head 2025-12-04T09:20:35.6745303Z * [new branch] gh/pianpwk/34/orig -> origin/gh/pianpwk/34/orig 2025-12-04T09:20:35.6747610Z * [new branch] gh/pianpwk/35/base -> origin/gh/pianpwk/35/base 2025-12-04T09:20:35.6749499Z * [new branch] gh/pianpwk/35/head -> origin/gh/pianpwk/35/head 2025-12-04T09:20:35.6751349Z * [new branch] gh/pianpwk/35/orig -> origin/gh/pianpwk/35/orig 2025-12-04T09:20:35.6754151Z * [new branch] gh/rec/141/base -> origin/gh/rec/141/base 2025-12-04T09:20:35.6756036Z * [new branch] gh/rec/141/head -> origin/gh/rec/141/head 2025-12-04T09:20:35.6758329Z * [new branch] gh/rec/153/base -> origin/gh/rec/153/base 2025-12-04T09:20:35.6760062Z * [new branch] gh/rec/153/head -> origin/gh/rec/153/head 2025-12-04T09:20:35.6761781Z * [new branch] gh/rec/153/orig -> origin/gh/rec/153/orig 2025-12-04T09:20:35.6764190Z * [new branch] gh/rec/154/base -> origin/gh/rec/154/base 2025-12-04T09:20:35.6765827Z * [new branch] gh/rec/154/head -> origin/gh/rec/154/head 2025-12-04T09:20:35.6767545Z * [new branch] gh/rec/154/orig -> origin/gh/rec/154/orig 2025-12-04T09:20:35.6769808Z * [new branch] gh/rec/164/base -> origin/gh/rec/164/base 2025-12-04T09:20:35.6771705Z * [new branch] gh/rec/164/head -> origin/gh/rec/164/head 2025-12-04T09:20:35.6774785Z * [new branch] gh/rec/164/orig -> origin/gh/rec/164/orig 2025-12-04T09:20:35.6777338Z * [new branch] gh/rec/166/base -> origin/gh/rec/166/base 2025-12-04T09:20:35.6778941Z * [new branch] gh/rec/166/head -> origin/gh/rec/166/head 2025-12-04T09:20:35.6780762Z * [new branch] gh/rec/166/orig -> origin/gh/rec/166/orig 2025-12-04T09:20:35.6783099Z * [new branch] gh/rec/167/base -> origin/gh/rec/167/base 2025-12-04T09:20:35.6784829Z * [new branch] gh/rec/167/head -> origin/gh/rec/167/head 2025-12-04T09:20:35.6786556Z * [new branch] gh/rec/167/orig -> origin/gh/rec/167/orig 2025-12-04T09:20:35.6788911Z * [new branch] gh/rec/168/base -> origin/gh/rec/168/base 2025-12-04T09:20:35.6790648Z * [new branch] gh/rec/168/head -> origin/gh/rec/168/head 2025-12-04T09:20:35.6792526Z * [new branch] gh/rec/168/orig -> origin/gh/rec/168/orig 2025-12-04T09:20:35.6794860Z * [new branch] gh/rec/169/base -> origin/gh/rec/169/base 2025-12-04T09:20:35.6796693Z * [new branch] gh/rec/169/head -> origin/gh/rec/169/head 2025-12-04T09:20:35.6798432Z * [new branch] gh/rec/169/orig -> origin/gh/rec/169/orig 2025-12-04T09:20:35.6800874Z * [new branch] gh/rec/170/base -> origin/gh/rec/170/base 2025-12-04T09:20:35.6802545Z * [new branch] gh/rec/170/head -> origin/gh/rec/170/head 2025-12-04T09:20:35.6804332Z * [new branch] gh/rec/170/orig -> origin/gh/rec/170/orig 2025-12-04T09:20:35.6807167Z * [new branch] gh/rec/171/base -> origin/gh/rec/171/base 2025-12-04T09:20:35.6808895Z * [new branch] gh/rec/171/head -> origin/gh/rec/171/head 2025-12-04T09:20:35.6810672Z * [new branch] gh/rec/171/orig -> origin/gh/rec/171/orig 2025-12-04T09:20:35.6812861Z * [new branch] gh/rec/172/base -> origin/gh/rec/172/base 2025-12-04T09:20:35.6814605Z * [new branch] gh/rec/172/head -> origin/gh/rec/172/head 2025-12-04T09:20:35.6816277Z * [new branch] gh/rec/172/orig -> origin/gh/rec/172/orig 2025-12-04T09:20:35.6818660Z * [new branch] gh/rec/173/base -> origin/gh/rec/173/base 2025-12-04T09:20:35.6820598Z * [new branch] gh/rec/173/head -> origin/gh/rec/173/head 2025-12-04T09:20:35.6822301Z * [new branch] gh/rec/173/orig -> origin/gh/rec/173/orig 2025-12-04T09:20:35.6824585Z * [new branch] gh/rec/174/base -> origin/gh/rec/174/base 2025-12-04T09:20:35.6826265Z * [new branch] gh/rec/174/head -> origin/gh/rec/174/head 2025-12-04T09:20:35.6827991Z * [new branch] gh/rec/174/orig -> origin/gh/rec/174/orig 2025-12-04T09:20:35.6830280Z * [new branch] gh/rec/175/base -> origin/gh/rec/175/base 2025-12-04T09:20:35.6832022Z * [new branch] gh/rec/175/head -> origin/gh/rec/175/head 2025-12-04T09:20:35.6833778Z * [new branch] gh/rec/175/orig -> origin/gh/rec/175/orig 2025-12-04T09:20:35.6836208Z * [new branch] gh/rec/176/base -> origin/gh/rec/176/base 2025-12-04T09:20:35.6837862Z * [new branch] gh/rec/176/head -> origin/gh/rec/176/head 2025-12-04T09:20:35.6839571Z * [new branch] gh/rec/176/orig -> origin/gh/rec/176/orig 2025-12-04T09:20:35.6841852Z * [new branch] gh/rec/177/base -> origin/gh/rec/177/base 2025-12-04T09:20:35.6843665Z * [new branch] gh/rec/177/head -> origin/gh/rec/177/head 2025-12-04T09:20:35.6845396Z * [new branch] gh/rec/177/orig -> origin/gh/rec/177/orig 2025-12-04T09:20:35.6848301Z * [new branch] gh/robert-hardwick/3/base -> origin/gh/robert-hardwick/3/base 2025-12-04T09:20:35.6850546Z * [new branch] gh/robert-hardwick/3/head -> origin/gh/robert-hardwick/3/head 2025-12-04T09:20:35.6852311Z * [new branch] gh/robert-hardwick/3/orig -> origin/gh/robert-hardwick/3/orig 2025-12-04T09:20:35.6854729Z * [new branch] gh/robert-hardwick/4/base -> origin/gh/robert-hardwick/4/base 2025-12-04T09:20:35.6856440Z * [new branch] gh/robert-hardwick/4/head -> origin/gh/robert-hardwick/4/head 2025-12-04T09:20:35.6858132Z * [new branch] gh/robert-hardwick/4/orig -> origin/gh/robert-hardwick/4/orig 2025-12-04T09:20:35.6860666Z * [new branch] gh/robert-hardwick/5/base -> origin/gh/robert-hardwick/5/base 2025-12-04T09:20:35.6862410Z * [new branch] gh/robert-hardwick/5/head -> origin/gh/robert-hardwick/5/head 2025-12-04T09:20:35.6864204Z * [new branch] gh/robert-hardwick/5/orig -> origin/gh/robert-hardwick/5/orig 2025-12-04T09:20:35.6866643Z * [new branch] gh/robert-hardwick/6/base -> origin/gh/robert-hardwick/6/base 2025-12-04T09:20:35.6868348Z * [new branch] gh/robert-hardwick/6/head -> origin/gh/robert-hardwick/6/head 2025-12-04T09:20:35.6870059Z * [new branch] gh/robert-hardwick/6/orig -> origin/gh/robert-hardwick/6/orig 2025-12-04T09:20:35.6872459Z * [new branch] gh/robert-hardwick/7/base -> origin/gh/robert-hardwick/7/base 2025-12-04T09:20:35.6874311Z * [new branch] gh/robert-hardwick/7/head -> origin/gh/robert-hardwick/7/head 2025-12-04T09:20:35.6876066Z * [new branch] gh/robert-hardwick/7/orig -> origin/gh/robert-hardwick/7/orig 2025-12-04T09:20:35.6878379Z * [new branch] gh/robert-hardwick/8/base -> origin/gh/robert-hardwick/8/base 2025-12-04T09:20:35.6880583Z * [new branch] gh/robert-hardwick/8/head -> origin/gh/robert-hardwick/8/head 2025-12-04T09:20:35.6882327Z * [new branch] gh/robert-hardwick/8/orig -> origin/gh/robert-hardwick/8/orig 2025-12-04T09:20:35.6884727Z * [new branch] gh/robert-hardwick/9/base -> origin/gh/robert-hardwick/9/base 2025-12-04T09:20:35.6886498Z * [new branch] gh/robert-hardwick/9/head -> origin/gh/robert-hardwick/9/head 2025-12-04T09:20:35.6888211Z * [new branch] gh/robert-hardwick/9/orig -> origin/gh/robert-hardwick/9/orig 2025-12-04T09:20:35.6891495Z * [new branch] gh/rtimpe/1/base -> origin/gh/rtimpe/1/base 2025-12-04T09:20:35.6893268Z * [new branch] gh/rtimpe/1/head -> origin/gh/rtimpe/1/head 2025-12-04T09:20:35.6895649Z * [new branch] gh/rtimpe/2/base -> origin/gh/rtimpe/2/base 2025-12-04T09:20:35.6897270Z * [new branch] gh/rtimpe/2/head -> origin/gh/rtimpe/2/head 2025-12-04T09:20:35.6899695Z * [new branch] gh/rtimpe/22/base -> origin/gh/rtimpe/22/base 2025-12-04T09:20:35.6901379Z * [new branch] gh/rtimpe/22/head -> origin/gh/rtimpe/22/head 2025-12-04T09:20:35.6903173Z * [new branch] gh/rtimpe/22/orig -> origin/gh/rtimpe/22/orig 2025-12-04T09:20:35.6905356Z * [new branch] gh/rtimpe/23/base -> origin/gh/rtimpe/23/base 2025-12-04T09:20:35.6907298Z * [new branch] gh/rtimpe/23/head -> origin/gh/rtimpe/23/head 2025-12-04T09:20:35.6908936Z * [new branch] gh/rtimpe/23/orig -> origin/gh/rtimpe/23/orig 2025-12-04T09:20:35.6911221Z * [new branch] gh/rtimpe/24/base -> origin/gh/rtimpe/24/base 2025-12-04T09:20:35.6912962Z * [new branch] gh/rtimpe/24/head -> origin/gh/rtimpe/24/head 2025-12-04T09:20:35.6914685Z * [new branch] gh/rtimpe/24/orig -> origin/gh/rtimpe/24/orig 2025-12-04T09:20:35.6917012Z * [new branch] gh/rtimpe/25/base -> origin/gh/rtimpe/25/base 2025-12-04T09:20:35.6918751Z * [new branch] gh/rtimpe/25/head -> origin/gh/rtimpe/25/head 2025-12-04T09:20:35.6920541Z * [new branch] gh/rtimpe/25/orig -> origin/gh/rtimpe/25/orig 2025-12-04T09:20:35.6922852Z * [new branch] gh/rtimpe/26/base -> origin/gh/rtimpe/26/base 2025-12-04T09:20:35.6924602Z * [new branch] gh/rtimpe/26/head -> origin/gh/rtimpe/26/head 2025-12-04T09:20:35.6926326Z * [new branch] gh/rtimpe/26/orig -> origin/gh/rtimpe/26/orig 2025-12-04T09:20:35.6928603Z * [new branch] gh/rtimpe/27/base -> origin/gh/rtimpe/27/base 2025-12-04T09:20:35.6930814Z * [new branch] gh/rtimpe/27/head -> origin/gh/rtimpe/27/head 2025-12-04T09:20:35.6932653Z * [new branch] gh/rtimpe/27/orig -> origin/gh/rtimpe/27/orig 2025-12-04T09:20:35.6934960Z * [new branch] gh/rtimpe/28/base -> origin/gh/rtimpe/28/base 2025-12-04T09:20:35.6936741Z * [new branch] gh/rtimpe/28/head -> origin/gh/rtimpe/28/head 2025-12-04T09:20:35.6938958Z * [new branch] gh/rtimpe/28/orig -> origin/gh/rtimpe/28/orig 2025-12-04T09:20:35.6941430Z * [new branch] gh/rtimpe/29/base -> origin/gh/rtimpe/29/base 2025-12-04T09:20:35.6943031Z * [new branch] gh/rtimpe/29/head -> origin/gh/rtimpe/29/head 2025-12-04T09:20:35.6944742Z * [new branch] gh/rtimpe/29/orig -> origin/gh/rtimpe/29/orig 2025-12-04T09:20:35.6947052Z * [new branch] gh/rtimpe/3/base -> origin/gh/rtimpe/3/base 2025-12-04T09:20:35.6948702Z * [new branch] gh/rtimpe/3/head -> origin/gh/rtimpe/3/head 2025-12-04T09:20:35.6951031Z * [new branch] gh/rtimpe/30/base -> origin/gh/rtimpe/30/base 2025-12-04T09:20:35.6952743Z * [new branch] gh/rtimpe/30/head -> origin/gh/rtimpe/30/head 2025-12-04T09:20:35.6954460Z * [new branch] gh/rtimpe/30/orig -> origin/gh/rtimpe/30/orig 2025-12-04T09:20:35.6956815Z * [new branch] gh/rtimpe/31/base -> origin/gh/rtimpe/31/base 2025-12-04T09:20:35.6958550Z * [new branch] gh/rtimpe/31/head -> origin/gh/rtimpe/31/head 2025-12-04T09:20:35.6960439Z * [new branch] gh/rtimpe/31/orig -> origin/gh/rtimpe/31/orig 2025-12-04T09:20:35.6963334Z * [new branch] gh/rtimpe/32/base -> origin/gh/rtimpe/32/base 2025-12-04T09:20:35.6965056Z * [new branch] gh/rtimpe/32/head -> origin/gh/rtimpe/32/head 2025-12-04T09:20:35.6966786Z * [new branch] gh/rtimpe/32/orig -> origin/gh/rtimpe/32/orig 2025-12-04T09:20:35.6969200Z * [new branch] gh/rtimpe/33/base -> origin/gh/rtimpe/33/base 2025-12-04T09:20:35.6971169Z * [new branch] gh/rtimpe/33/head -> origin/gh/rtimpe/33/head 2025-12-04T09:20:35.6972913Z * [new branch] gh/rtimpe/33/orig -> origin/gh/rtimpe/33/orig 2025-12-04T09:20:35.6975169Z * [new branch] gh/rtimpe/34/base -> origin/gh/rtimpe/34/base 2025-12-04T09:20:35.6976894Z * [new branch] gh/rtimpe/34/head -> origin/gh/rtimpe/34/head 2025-12-04T09:20:35.6978764Z * [new branch] gh/rtimpe/34/orig -> origin/gh/rtimpe/34/orig 2025-12-04T09:20:35.6981173Z * [new branch] gh/rtimpe/35/base -> origin/gh/rtimpe/35/base 2025-12-04T09:20:35.6982890Z * [new branch] gh/rtimpe/35/head -> origin/gh/rtimpe/35/head 2025-12-04T09:20:35.6984670Z * [new branch] gh/rtimpe/35/orig -> origin/gh/rtimpe/35/orig 2025-12-04T09:20:35.6987610Z * [new branch] gh/rtimpe/4/base -> origin/gh/rtimpe/4/base 2025-12-04T09:20:35.6989355Z * [new branch] gh/rtimpe/4/head -> origin/gh/rtimpe/4/head 2025-12-04T09:20:35.6992286Z * [new branch] gh/ruisizhang123/1/base -> origin/gh/ruisizhang123/1/base 2025-12-04T09:20:35.6994011Z * [new branch] gh/ruisizhang123/1/head -> origin/gh/ruisizhang123/1/head 2025-12-04T09:20:35.6995725Z * [new branch] gh/ruisizhang123/1/orig -> origin/gh/ruisizhang123/1/orig 2025-12-04T09:20:35.6998260Z * [new branch] gh/ruisizhang123/4/base -> origin/gh/ruisizhang123/4/base 2025-12-04T09:20:35.6999985Z * [new branch] gh/ruisizhang123/4/head -> origin/gh/ruisizhang123/4/head 2025-12-04T09:20:35.7001751Z * [new branch] gh/ruisizhang123/4/orig -> origin/gh/ruisizhang123/4/orig 2025-12-04T09:20:35.7003996Z * [new branch] gh/ruisizhang123/5/base -> origin/gh/ruisizhang123/5/base 2025-12-04T09:20:35.7005743Z * [new branch] gh/ruisizhang123/5/head -> origin/gh/ruisizhang123/5/head 2025-12-04T09:20:35.7007533Z * [new branch] gh/ruisizhang123/5/orig -> origin/gh/ruisizhang123/5/orig 2025-12-04T09:20:35.7009852Z * [new branch] gh/ruisizhang123/6/base -> origin/gh/ruisizhang123/6/base 2025-12-04T09:20:35.7011607Z * [new branch] gh/ruisizhang123/6/head -> origin/gh/ruisizhang123/6/head 2025-12-04T09:20:35.7013385Z * [new branch] gh/ruisizhang123/6/orig -> origin/gh/ruisizhang123/6/orig 2025-12-04T09:20:35.7015732Z * [new branch] gh/ruisizhang123/7/base -> origin/gh/ruisizhang123/7/base 2025-12-04T09:20:35.7017441Z * [new branch] gh/ruisizhang123/7/head -> origin/gh/ruisizhang123/7/head 2025-12-04T09:20:35.7019220Z * [new branch] gh/ruisizhang123/7/orig -> origin/gh/ruisizhang123/7/orig 2025-12-04T09:20:35.7021619Z * [new branch] gh/ruisizhang123/8/base -> origin/gh/ruisizhang123/8/base 2025-12-04T09:20:35.7023440Z * [new branch] gh/ruisizhang123/8/head -> origin/gh/ruisizhang123/8/head 2025-12-04T09:20:35.7025144Z * [new branch] gh/ruisizhang123/8/orig -> origin/gh/ruisizhang123/8/orig 2025-12-04T09:20:35.7027475Z * [new branch] gh/ruisizhang123/9/base -> origin/gh/ruisizhang123/9/base 2025-12-04T09:20:35.7029182Z * [new branch] gh/ruisizhang123/9/head -> origin/gh/ruisizhang123/9/head 2025-12-04T09:20:35.7030968Z * [new branch] gh/ruisizhang123/9/orig -> origin/gh/ruisizhang123/9/orig 2025-12-04T09:20:35.7033871Z * [new branch] gh/seemethere/52/base -> origin/gh/seemethere/52/base 2025-12-04T09:20:35.7035641Z * [new branch] gh/seemethere/52/head -> origin/gh/seemethere/52/head 2025-12-04T09:20:35.7037384Z * [new branch] gh/seemethere/52/orig -> origin/gh/seemethere/52/orig 2025-12-04T09:20:35.7039705Z * [new branch] gh/seemethere/53/base -> origin/gh/seemethere/53/base 2025-12-04T09:20:35.7041472Z * [new branch] gh/seemethere/53/head -> origin/gh/seemethere/53/head 2025-12-04T09:20:35.7043126Z * [new branch] gh/seemethere/53/orig -> origin/gh/seemethere/53/orig 2025-12-04T09:20:35.7045563Z * [new branch] gh/seemethere/54/base -> origin/gh/seemethere/54/base 2025-12-04T09:20:35.7047298Z * [new branch] gh/seemethere/54/head -> origin/gh/seemethere/54/head 2025-12-04T09:20:35.7049168Z * [new branch] gh/seemethere/54/orig -> origin/gh/seemethere/54/orig 2025-12-04T09:20:35.7051303Z * [new branch] gh/seemethere/55/base -> origin/gh/seemethere/55/base 2025-12-04T09:20:35.7053130Z * [new branch] gh/seemethere/55/head -> origin/gh/seemethere/55/head 2025-12-04T09:20:35.7054924Z * [new branch] gh/seemethere/55/orig -> origin/gh/seemethere/55/orig 2025-12-04T09:20:35.7057159Z * [new branch] gh/seemethere/59/base -> origin/gh/seemethere/59/base 2025-12-04T09:20:35.7058907Z * [new branch] gh/seemethere/59/head -> origin/gh/seemethere/59/head 2025-12-04T09:20:35.7060774Z * [new branch] gh/seemethere/59/orig -> origin/gh/seemethere/59/orig 2025-12-04T09:20:35.7063042Z * [new branch] gh/seemethere/62/base -> origin/gh/seemethere/62/base 2025-12-04T09:20:35.7064757Z * [new branch] gh/seemethere/62/head -> origin/gh/seemethere/62/head 2025-12-04T09:20:35.7066471Z * [new branch] gh/seemethere/62/orig -> origin/gh/seemethere/62/orig 2025-12-04T09:20:35.7068796Z * [new branch] gh/seemethere/63/base -> origin/gh/seemethere/63/base 2025-12-04T09:20:35.7070557Z * [new branch] gh/seemethere/63/head -> origin/gh/seemethere/63/head 2025-12-04T09:20:35.7072555Z * [new branch] gh/seemethere/63/orig -> origin/gh/seemethere/63/orig 2025-12-04T09:20:35.7074836Z * [new branch] gh/seemethere/71/base -> origin/gh/seemethere/71/base 2025-12-04T09:20:35.7076637Z * [new branch] gh/seemethere/71/head -> origin/gh/seemethere/71/head 2025-12-04T09:20:35.7078351Z * [new branch] gh/seemethere/71/orig -> origin/gh/seemethere/71/orig 2025-12-04T09:20:35.7080724Z * [new branch] gh/seemethere/72/base -> origin/gh/seemethere/72/base 2025-12-04T09:20:35.7082666Z * [new branch] gh/seemethere/72/head -> origin/gh/seemethere/72/head 2025-12-04T09:20:35.7084385Z * [new branch] gh/seemethere/72/orig -> origin/gh/seemethere/72/orig 2025-12-04T09:20:35.7086683Z * [new branch] gh/seemethere/73/base -> origin/gh/seemethere/73/base 2025-12-04T09:20:35.7088446Z * [new branch] gh/seemethere/73/head -> origin/gh/seemethere/73/head 2025-12-04T09:20:35.7090199Z * [new branch] gh/seemethere/73/orig -> origin/gh/seemethere/73/orig 2025-12-04T09:20:35.7092631Z * [new branch] gh/seemethere/74/base -> origin/gh/seemethere/74/base 2025-12-04T09:20:35.7094363Z * [new branch] gh/seemethere/74/head -> origin/gh/seemethere/74/head 2025-12-04T09:20:35.7096124Z * [new branch] gh/seemethere/74/orig -> origin/gh/seemethere/74/orig 2025-12-04T09:20:35.7098512Z * [new branch] gh/seemethere/75/base -> origin/gh/seemethere/75/base 2025-12-04T09:20:35.7100189Z * [new branch] gh/seemethere/75/head -> origin/gh/seemethere/75/head 2025-12-04T09:20:35.7102504Z * [new branch] gh/seemethere/75/orig -> origin/gh/seemethere/75/orig 2025-12-04T09:20:35.7104796Z * [new branch] gh/seemethere/76/base -> origin/gh/seemethere/76/base 2025-12-04T09:20:35.7106628Z * [new branch] gh/seemethere/76/head -> origin/gh/seemethere/76/head 2025-12-04T09:20:35.7108337Z * [new branch] gh/seemethere/76/orig -> origin/gh/seemethere/76/orig 2025-12-04T09:20:35.7111298Z * [new branch] gh/shunting314/145/base -> origin/gh/shunting314/145/base 2025-12-04T09:20:35.7113145Z * [new branch] gh/shunting314/145/head -> origin/gh/shunting314/145/head 2025-12-04T09:20:35.7115069Z * [new branch] gh/shunting314/145/orig -> origin/gh/shunting314/145/orig 2025-12-04T09:20:35.7117722Z * [new branch] gh/shunting314/176/base -> origin/gh/shunting314/176/base 2025-12-04T09:20:35.7119506Z * [new branch] gh/shunting314/176/head -> origin/gh/shunting314/176/head 2025-12-04T09:20:35.7121250Z * [new branch] gh/shunting314/176/orig -> origin/gh/shunting314/176/orig 2025-12-04T09:20:35.7123570Z * [new branch] gh/shunting314/249/base -> origin/gh/shunting314/249/base 2025-12-04T09:20:35.7125395Z * [new branch] gh/shunting314/249/head -> origin/gh/shunting314/249/head 2025-12-04T09:20:35.7127253Z * [new branch] gh/shunting314/249/orig -> origin/gh/shunting314/249/orig 2025-12-04T09:20:35.7129715Z * [new branch] gh/shunting314/253/base -> origin/gh/shunting314/253/base 2025-12-04T09:20:35.7131299Z * [new branch] gh/shunting314/253/head -> origin/gh/shunting314/253/head 2025-12-04T09:20:35.7132986Z * [new branch] gh/shunting314/253/orig -> origin/gh/shunting314/253/orig 2025-12-04T09:20:35.7135477Z * [new branch] gh/shunting314/256/base -> origin/gh/shunting314/256/base 2025-12-04T09:20:35.7137312Z * [new branch] gh/shunting314/256/head -> origin/gh/shunting314/256/head 2025-12-04T09:20:35.7139037Z * [new branch] gh/shunting314/256/orig -> origin/gh/shunting314/256/orig 2025-12-04T09:20:35.7141684Z * [new branch] gh/shunting314/257/base -> origin/gh/shunting314/257/base 2025-12-04T09:20:35.7143487Z * [new branch] gh/shunting314/257/head -> origin/gh/shunting314/257/head 2025-12-04T09:20:35.7145267Z * [new branch] gh/shunting314/257/orig -> origin/gh/shunting314/257/orig 2025-12-04T09:20:35.7147820Z * [new branch] gh/shunting314/258/base -> origin/gh/shunting314/258/base 2025-12-04T09:20:35.7149504Z * [new branch] gh/shunting314/258/head -> origin/gh/shunting314/258/head 2025-12-04T09:20:35.7151326Z * [new branch] gh/shunting314/258/orig -> origin/gh/shunting314/258/orig 2025-12-04T09:20:35.7153462Z * [new branch] gh/shunting314/259/base -> origin/gh/shunting314/259/base 2025-12-04T09:20:35.7155163Z * [new branch] gh/shunting314/259/head -> origin/gh/shunting314/259/head 2025-12-04T09:20:35.7156879Z * [new branch] gh/shunting314/259/orig -> origin/gh/shunting314/259/orig 2025-12-04T09:20:35.7159336Z * [new branch] gh/shunting314/260/base -> origin/gh/shunting314/260/base 2025-12-04T09:20:35.7161205Z * [new branch] gh/shunting314/260/head -> origin/gh/shunting314/260/head 2025-12-04T09:20:35.7163032Z * [new branch] gh/shunting314/260/orig -> origin/gh/shunting314/260/orig 2025-12-04T09:20:35.7165414Z * [new branch] gh/shunting314/261/base -> origin/gh/shunting314/261/base 2025-12-04T09:20:35.7167381Z * [new branch] gh/shunting314/261/head -> origin/gh/shunting314/261/head 2025-12-04T09:20:35.7169153Z * [new branch] gh/shunting314/261/orig -> origin/gh/shunting314/261/orig 2025-12-04T09:20:35.7171801Z * [new branch] gh/shunting314/262/base -> origin/gh/shunting314/262/base 2025-12-04T09:20:35.7174288Z * [new branch] gh/shunting314/262/head -> origin/gh/shunting314/262/head 2025-12-04T09:20:35.7176080Z * [new branch] gh/shunting314/262/orig -> origin/gh/shunting314/262/orig 2025-12-04T09:20:35.7179418Z * [new branch] gh/shunting314/263/base -> origin/gh/shunting314/263/base 2025-12-04T09:20:35.7181348Z * [new branch] gh/shunting314/263/head -> origin/gh/shunting314/263/head 2025-12-04T09:20:35.7183109Z * [new branch] gh/shunting314/263/orig -> origin/gh/shunting314/263/orig 2025-12-04T09:20:35.7185463Z * [new branch] gh/shunting314/264/base -> origin/gh/shunting314/264/base 2025-12-04T09:20:35.7187389Z * [new branch] gh/shunting314/264/head -> origin/gh/shunting314/264/head 2025-12-04T09:20:35.7189027Z * [new branch] gh/shunting314/264/orig -> origin/gh/shunting314/264/orig 2025-12-04T09:20:35.7191396Z * [new branch] gh/shunting314/265/base -> origin/gh/shunting314/265/base 2025-12-04T09:20:35.7193139Z * [new branch] gh/shunting314/265/head -> origin/gh/shunting314/265/head 2025-12-04T09:20:35.7194893Z * [new branch] gh/shunting314/265/orig -> origin/gh/shunting314/265/orig 2025-12-04T09:20:35.7197260Z * [new branch] gh/shunting314/266/base -> origin/gh/shunting314/266/base 2025-12-04T09:20:35.7199163Z * [new branch] gh/shunting314/266/head -> origin/gh/shunting314/266/head 2025-12-04T09:20:35.7200963Z * [new branch] gh/shunting314/266/orig -> origin/gh/shunting314/266/orig 2025-12-04T09:20:35.7203409Z * [new branch] gh/shunting314/267/base -> origin/gh/shunting314/267/base 2025-12-04T09:20:35.7205326Z * [new branch] gh/shunting314/267/head -> origin/gh/shunting314/267/head 2025-12-04T09:20:35.7207116Z * [new branch] gh/shunting314/267/orig -> origin/gh/shunting314/267/orig 2025-12-04T09:20:35.7210052Z * [new branch] gh/shunting314/268/base -> origin/gh/shunting314/268/base 2025-12-04T09:20:35.7211792Z * [new branch] gh/shunting314/268/head -> origin/gh/shunting314/268/head 2025-12-04T09:20:35.7213610Z * [new branch] gh/shunting314/268/orig -> origin/gh/shunting314/268/orig 2025-12-04T09:20:35.7216052Z * [new branch] gh/shunting314/269/base -> origin/gh/shunting314/269/base 2025-12-04T09:20:35.7217813Z * [new branch] gh/shunting314/269/head -> origin/gh/shunting314/269/head 2025-12-04T09:20:35.7219579Z * [new branch] gh/shunting314/269/orig -> origin/gh/shunting314/269/orig 2025-12-04T09:20:35.7222520Z * [new branch] gh/silverguo/1/base -> origin/gh/silverguo/1/base 2025-12-04T09:20:35.7224300Z * [new branch] gh/silverguo/1/head -> origin/gh/silverguo/1/head 2025-12-04T09:20:35.7226505Z * [new branch] gh/silverguo/2/base -> origin/gh/silverguo/2/base 2025-12-04T09:20:35.7228161Z * [new branch] gh/silverguo/2/head -> origin/gh/silverguo/2/head 2025-12-04T09:20:35.7230797Z * [new branch] gh/silverguo/3/base -> origin/gh/silverguo/3/base 2025-12-04T09:20:35.7232688Z * [new branch] gh/silverguo/3/head -> origin/gh/silverguo/3/head 2025-12-04T09:20:35.7234941Z * [new branch] gh/silverguo/4/base -> origin/gh/silverguo/4/base 2025-12-04T09:20:35.7236688Z * [new branch] gh/silverguo/4/head -> origin/gh/silverguo/4/head 2025-12-04T09:20:35.7239454Z * [new branch] gh/slayton58/39/base -> origin/gh/slayton58/39/base 2025-12-04T09:20:35.7241211Z * [new branch] gh/slayton58/39/head -> origin/gh/slayton58/39/head 2025-12-04T09:20:35.7243011Z * [new branch] gh/slayton58/39/orig -> origin/gh/slayton58/39/orig 2025-12-04T09:20:35.7245716Z * [new branch] gh/slayton58/42/base -> origin/gh/slayton58/42/base 2025-12-04T09:20:35.7247455Z * [new branch] gh/slayton58/42/head -> origin/gh/slayton58/42/head 2025-12-04T09:20:35.7249319Z * [new branch] gh/slayton58/42/orig -> origin/gh/slayton58/42/orig 2025-12-04T09:20:35.7251793Z * [new branch] gh/slayton58/43/base -> origin/gh/slayton58/43/base 2025-12-04T09:20:35.7253450Z * [new branch] gh/slayton58/43/head -> origin/gh/slayton58/43/head 2025-12-04T09:20:35.7255186Z * [new branch] gh/slayton58/43/orig -> origin/gh/slayton58/43/orig 2025-12-04T09:20:35.7257626Z * [new branch] gh/slayton58/44/base -> origin/gh/slayton58/44/base 2025-12-04T09:20:35.7260144Z * [new branch] gh/slayton58/44/head -> origin/gh/slayton58/44/head 2025-12-04T09:20:35.7261964Z * [new branch] gh/slayton58/44/orig -> origin/gh/slayton58/44/orig 2025-12-04T09:20:35.7264738Z * [new branch] gh/slayton58/45/base -> origin/gh/slayton58/45/base 2025-12-04T09:20:35.7266467Z * [new branch] gh/slayton58/45/head -> origin/gh/slayton58/45/head 2025-12-04T09:20:35.7268285Z * [new branch] gh/slayton58/45/orig -> origin/gh/slayton58/45/orig 2025-12-04T09:20:35.7270662Z * [new branch] gh/slayton58/46/base -> origin/gh/slayton58/46/base 2025-12-04T09:20:35.7272753Z * [new branch] gh/slayton58/46/head -> origin/gh/slayton58/46/head 2025-12-04T09:20:35.7274512Z * [new branch] gh/slayton58/46/orig -> origin/gh/slayton58/46/orig 2025-12-04T09:20:35.7276918Z * [new branch] gh/slayton58/6/base -> origin/gh/slayton58/6/base 2025-12-04T09:20:35.7278672Z * [new branch] gh/slayton58/6/head -> origin/gh/slayton58/6/head 2025-12-04T09:20:35.7281321Z * [new branch] gh/slayton58/7/base -> origin/gh/slayton58/7/base 2025-12-04T09:20:35.7282940Z * [new branch] gh/slayton58/7/head -> origin/gh/slayton58/7/head 2025-12-04T09:20:35.7286330Z * [new branch] gh/soulitzer/269/base -> origin/gh/soulitzer/269/base 2025-12-04T09:20:35.7288055Z * [new branch] gh/soulitzer/269/head -> origin/gh/soulitzer/269/head 2025-12-04T09:20:35.7289827Z * [new branch] gh/soulitzer/269/orig -> origin/gh/soulitzer/269/orig 2025-12-04T09:20:35.7292319Z * [new branch] gh/soulitzer/276/base -> origin/gh/soulitzer/276/base 2025-12-04T09:20:35.7294119Z * [new branch] gh/soulitzer/276/head -> origin/gh/soulitzer/276/head 2025-12-04T09:20:35.7295825Z * [new branch] gh/soulitzer/276/orig -> origin/gh/soulitzer/276/orig 2025-12-04T09:20:35.7298530Z * [new branch] gh/soulitzer/287/base -> origin/gh/soulitzer/287/base 2025-12-04T09:20:35.7300343Z * [new branch] gh/soulitzer/287/head -> origin/gh/soulitzer/287/head 2025-12-04T09:20:35.7302056Z * [new branch] gh/soulitzer/287/orig -> origin/gh/soulitzer/287/orig 2025-12-04T09:20:35.7304581Z * [new branch] gh/soulitzer/296/base -> origin/gh/soulitzer/296/base 2025-12-04T09:20:35.7306344Z * [new branch] gh/soulitzer/296/head -> origin/gh/soulitzer/296/head 2025-12-04T09:20:35.7308297Z * [new branch] gh/soulitzer/296/orig -> origin/gh/soulitzer/296/orig 2025-12-04T09:20:35.7310644Z * [new branch] gh/soulitzer/299/base -> origin/gh/soulitzer/299/base 2025-12-04T09:20:35.7312466Z * [new branch] gh/soulitzer/299/head -> origin/gh/soulitzer/299/head 2025-12-04T09:20:35.7314201Z * [new branch] gh/soulitzer/299/orig -> origin/gh/soulitzer/299/orig 2025-12-04T09:20:35.7316673Z * [new branch] gh/soulitzer/300/base -> origin/gh/soulitzer/300/base 2025-12-04T09:20:35.7318477Z * [new branch] gh/soulitzer/300/head -> origin/gh/soulitzer/300/head 2025-12-04T09:20:35.7320294Z * [new branch] gh/soulitzer/300/orig -> origin/gh/soulitzer/300/orig 2025-12-04T09:20:35.7322790Z * [new branch] gh/soulitzer/301/base -> origin/gh/soulitzer/301/base 2025-12-04T09:20:35.7325081Z * [new branch] gh/soulitzer/301/head -> origin/gh/soulitzer/301/head 2025-12-04T09:20:35.7326810Z * [new branch] gh/soulitzer/301/orig -> origin/gh/soulitzer/301/orig 2025-12-04T09:20:35.7329107Z * [new branch] gh/soulitzer/313/base -> origin/gh/soulitzer/313/base 2025-12-04T09:20:35.7330879Z * [new branch] gh/soulitzer/313/head -> origin/gh/soulitzer/313/head 2025-12-04T09:20:35.7332696Z * [new branch] gh/soulitzer/313/orig -> origin/gh/soulitzer/313/orig 2025-12-04T09:20:35.7334950Z * [new branch] gh/soulitzer/319/base -> origin/gh/soulitzer/319/base 2025-12-04T09:20:35.7336665Z * [new branch] gh/soulitzer/319/head -> origin/gh/soulitzer/319/head 2025-12-04T09:20:35.7338402Z * [new branch] gh/soulitzer/319/orig -> origin/gh/soulitzer/319/orig 2025-12-04T09:20:35.7341161Z * [new branch] gh/soulitzer/320/base -> origin/gh/soulitzer/320/base 2025-12-04T09:20:35.7342830Z * [new branch] gh/soulitzer/320/head -> origin/gh/soulitzer/320/head 2025-12-04T09:20:35.7344586Z * [new branch] gh/soulitzer/320/orig -> origin/gh/soulitzer/320/orig 2025-12-04T09:20:35.7346964Z * [new branch] gh/soulitzer/336/base -> origin/gh/soulitzer/336/base 2025-12-04T09:20:35.7348677Z * [new branch] gh/soulitzer/336/head -> origin/gh/soulitzer/336/head 2025-12-04T09:20:35.7350363Z * [new branch] gh/soulitzer/336/orig -> origin/gh/soulitzer/336/orig 2025-12-04T09:20:35.7352670Z * [new branch] gh/soulitzer/347/base -> origin/gh/soulitzer/347/base 2025-12-04T09:20:35.7354416Z * [new branch] gh/soulitzer/347/head -> origin/gh/soulitzer/347/head 2025-12-04T09:20:35.7356078Z * [new branch] gh/soulitzer/347/orig -> origin/gh/soulitzer/347/orig 2025-12-04T09:20:35.7358548Z * [new branch] gh/soulitzer/349/base -> origin/gh/soulitzer/349/base 2025-12-04T09:20:35.7360283Z * [new branch] gh/soulitzer/349/head -> origin/gh/soulitzer/349/head 2025-12-04T09:20:35.7362074Z * [new branch] gh/soulitzer/349/orig -> origin/gh/soulitzer/349/orig 2025-12-04T09:20:35.7364562Z * [new branch] gh/soulitzer/350/base -> origin/gh/soulitzer/350/base 2025-12-04T09:20:35.7366288Z * [new branch] gh/soulitzer/350/head -> origin/gh/soulitzer/350/head 2025-12-04T09:20:35.7368032Z * [new branch] gh/soulitzer/350/orig -> origin/gh/soulitzer/350/orig 2025-12-04T09:20:35.7370378Z * [new branch] gh/soulitzer/351/base -> origin/gh/soulitzer/351/base 2025-12-04T09:20:35.7372347Z * [new branch] gh/soulitzer/351/head -> origin/gh/soulitzer/351/head 2025-12-04T09:20:35.7374045Z * [new branch] gh/soulitzer/351/orig -> origin/gh/soulitzer/351/orig 2025-12-04T09:20:35.7376341Z * [new branch] gh/soulitzer/353/base -> origin/gh/soulitzer/353/base 2025-12-04T09:20:35.7378192Z * [new branch] gh/soulitzer/353/head -> origin/gh/soulitzer/353/head 2025-12-04T09:20:35.7380226Z * [new branch] gh/soulitzer/353/orig -> origin/gh/soulitzer/353/orig 2025-12-04T09:20:35.7383063Z * [new branch] gh/soulitzer/358/base -> origin/gh/soulitzer/358/base 2025-12-04T09:20:35.7384940Z * [new branch] gh/soulitzer/358/head -> origin/gh/soulitzer/358/head 2025-12-04T09:20:35.7386684Z * [new branch] gh/soulitzer/358/orig -> origin/gh/soulitzer/358/orig 2025-12-04T09:20:35.7389662Z * [new branch] gh/soulitzer/359/base -> origin/gh/soulitzer/359/base 2025-12-04T09:20:35.7391414Z * [new branch] gh/soulitzer/359/head -> origin/gh/soulitzer/359/head 2025-12-04T09:20:35.7393129Z * [new branch] gh/soulitzer/359/orig -> origin/gh/soulitzer/359/orig 2025-12-04T09:20:35.7395636Z * [new branch] gh/soulitzer/374/base -> origin/gh/soulitzer/374/base 2025-12-04T09:20:35.7397393Z * [new branch] gh/soulitzer/374/head -> origin/gh/soulitzer/374/head 2025-12-04T09:20:35.7399595Z * [new branch] gh/soulitzer/374/orig -> origin/gh/soulitzer/374/orig 2025-12-04T09:20:35.7401921Z * [new branch] gh/soulitzer/375/base -> origin/gh/soulitzer/375/base 2025-12-04T09:20:35.7403942Z * [new branch] gh/soulitzer/375/head -> origin/gh/soulitzer/375/head 2025-12-04T09:20:35.7405371Z * [new branch] gh/soulitzer/375/orig -> origin/gh/soulitzer/375/orig 2025-12-04T09:20:35.7407736Z * [new branch] gh/soulitzer/380/base -> origin/gh/soulitzer/380/base 2025-12-04T09:20:35.7409899Z * [new branch] gh/soulitzer/380/head -> origin/gh/soulitzer/380/head 2025-12-04T09:20:35.7411233Z * [new branch] gh/soulitzer/380/orig -> origin/gh/soulitzer/380/orig 2025-12-04T09:20:35.7413805Z * [new branch] gh/soulitzer/385/base -> origin/gh/soulitzer/385/base 2025-12-04T09:20:35.7415602Z * [new branch] gh/soulitzer/385/head -> origin/gh/soulitzer/385/head 2025-12-04T09:20:35.7417430Z * [new branch] gh/soulitzer/385/orig -> origin/gh/soulitzer/385/orig 2025-12-04T09:20:35.7420489Z * [new branch] gh/soulitzer/386/base -> origin/gh/soulitzer/386/base 2025-12-04T09:20:35.7421837Z * [new branch] gh/soulitzer/386/head -> origin/gh/soulitzer/386/head 2025-12-04T09:20:35.7423554Z * [new branch] gh/soulitzer/386/orig -> origin/gh/soulitzer/386/orig 2025-12-04T09:20:35.7426404Z * [new branch] gh/soulitzer/387/base -> origin/gh/soulitzer/387/base 2025-12-04T09:20:35.7428126Z * [new branch] gh/soulitzer/387/head -> origin/gh/soulitzer/387/head 2025-12-04T09:20:35.7430263Z * [new branch] gh/soulitzer/387/orig -> origin/gh/soulitzer/387/orig 2025-12-04T09:20:35.7432813Z * [new branch] gh/soulitzer/388/base -> origin/gh/soulitzer/388/base 2025-12-04T09:20:35.7434200Z * [new branch] gh/soulitzer/388/head -> origin/gh/soulitzer/388/head 2025-12-04T09:20:35.7435946Z * [new branch] gh/soulitzer/388/orig -> origin/gh/soulitzer/388/orig 2025-12-04T09:20:35.7438756Z * [new branch] gh/soulitzer/389/base -> origin/gh/soulitzer/389/base 2025-12-04T09:20:35.7440133Z * [new branch] gh/soulitzer/389/head -> origin/gh/soulitzer/389/head 2025-12-04T09:20:35.7441906Z * [new branch] gh/soulitzer/389/orig -> origin/gh/soulitzer/389/orig 2025-12-04T09:20:35.7444655Z * [new branch] gh/soulitzer/390/base -> origin/gh/soulitzer/390/base 2025-12-04T09:20:35.7446094Z * [new branch] gh/soulitzer/390/head -> origin/gh/soulitzer/390/head 2025-12-04T09:20:35.7447716Z * [new branch] gh/soulitzer/390/orig -> origin/gh/soulitzer/390/orig 2025-12-04T09:20:35.7450392Z * [new branch] gh/soulitzer/391/base -> origin/gh/soulitzer/391/base 2025-12-04T09:20:35.7451805Z * [new branch] gh/soulitzer/391/head -> origin/gh/soulitzer/391/head 2025-12-04T09:20:35.7453534Z * [new branch] gh/soulitzer/391/orig -> origin/gh/soulitzer/391/orig 2025-12-04T09:20:35.7456239Z * [new branch] gh/soulitzer/392/base -> origin/gh/soulitzer/392/base 2025-12-04T09:20:35.7457727Z * [new branch] gh/soulitzer/392/head -> origin/gh/soulitzer/392/head 2025-12-04T09:20:35.7459541Z * [new branch] gh/soulitzer/392/orig -> origin/gh/soulitzer/392/orig 2025-12-04T09:20:35.7462985Z * [new branch] gh/swolchok/728/next -> origin/gh/swolchok/728/next 2025-12-04T09:20:35.7465485Z * [new branch] gh/swolchok/819/base -> origin/gh/swolchok/819/base 2025-12-04T09:20:35.7466863Z * [new branch] gh/swolchok/819/head -> origin/gh/swolchok/819/head 2025-12-04T09:20:35.7468587Z * [new branch] gh/swolchok/819/orig -> origin/gh/swolchok/819/orig 2025-12-04T09:20:35.7471369Z * [new branch] gh/swolchok/824/base -> origin/gh/swolchok/824/base 2025-12-04T09:20:35.7474778Z * [new branch] gh/swolchok/824/head -> origin/gh/swolchok/824/head 2025-12-04T09:20:35.7475994Z * [new branch] gh/swolchok/824/orig -> origin/gh/swolchok/824/orig 2025-12-04T09:20:35.7478724Z * [new branch] gh/swolchok/829/base -> origin/gh/swolchok/829/base 2025-12-04T09:20:35.7480135Z * [new branch] gh/swolchok/829/head -> origin/gh/swolchok/829/head 2025-12-04T09:20:35.7481893Z * [new branch] gh/swolchok/829/orig -> origin/gh/swolchok/829/orig 2025-12-04T09:20:35.7484667Z * [new branch] gh/swolchok/839/base -> origin/gh/swolchok/839/base 2025-12-04T09:20:35.7486025Z * [new branch] gh/swolchok/839/head -> origin/gh/swolchok/839/head 2025-12-04T09:20:35.7487761Z * [new branch] gh/swolchok/839/orig -> origin/gh/swolchok/839/orig 2025-12-04T09:20:35.7490410Z * [new branch] gh/swolchok/841/base -> origin/gh/swolchok/841/base 2025-12-04T09:20:35.7491871Z * [new branch] gh/swolchok/841/head -> origin/gh/swolchok/841/head 2025-12-04T09:20:35.7493602Z * [new branch] gh/swolchok/841/orig -> origin/gh/swolchok/841/orig 2025-12-04T09:20:35.7496259Z * [new branch] gh/swolchok/842/base -> origin/gh/swolchok/842/base 2025-12-04T09:20:35.7497746Z * [new branch] gh/swolchok/842/head -> origin/gh/swolchok/842/head 2025-12-04T09:20:35.7499843Z * [new branch] gh/swolchok/842/orig -> origin/gh/swolchok/842/orig 2025-12-04T09:20:35.7502217Z * [new branch] gh/swolchok/845/base -> origin/gh/swolchok/845/base 2025-12-04T09:20:35.7503704Z * [new branch] gh/swolchok/845/head -> origin/gh/swolchok/845/head 2025-12-04T09:20:35.7505538Z * [new branch] gh/swolchok/845/orig -> origin/gh/swolchok/845/orig 2025-12-04T09:20:35.7508142Z * [new branch] gh/swolchok/848/base -> origin/gh/swolchok/848/base 2025-12-04T09:20:35.7509976Z * [new branch] gh/swolchok/848/head -> origin/gh/swolchok/848/head 2025-12-04T09:20:35.7511556Z * [new branch] gh/swolchok/848/orig -> origin/gh/swolchok/848/orig 2025-12-04T09:20:35.7514209Z * [new branch] gh/swolchok/856/base -> origin/gh/swolchok/856/base 2025-12-04T09:20:35.7515621Z * [new branch] gh/swolchok/856/head -> origin/gh/swolchok/856/head 2025-12-04T09:20:35.7517591Z * [new branch] gh/swolchok/856/orig -> origin/gh/swolchok/856/orig 2025-12-04T09:20:35.7520186Z * [new branch] gh/swolchok/860/base -> origin/gh/swolchok/860/base 2025-12-04T09:20:35.7521614Z * [new branch] gh/swolchok/860/head -> origin/gh/swolchok/860/head 2025-12-04T09:20:35.7523621Z * [new branch] gh/swolchok/860/orig -> origin/gh/swolchok/860/orig 2025-12-04T09:20:35.7526559Z * [new branch] gh/swolchok/861/base -> origin/gh/swolchok/861/base 2025-12-04T09:20:35.7528089Z * [new branch] gh/swolchok/861/head -> origin/gh/swolchok/861/head 2025-12-04T09:20:35.7530110Z * [new branch] gh/swolchok/861/orig -> origin/gh/swolchok/861/orig 2025-12-04T09:20:35.7532638Z * [new branch] gh/swolchok/862/base -> origin/gh/swolchok/862/base 2025-12-04T09:20:35.7534114Z * [new branch] gh/swolchok/862/head -> origin/gh/swolchok/862/head 2025-12-04T09:20:35.7535791Z * [new branch] gh/swolchok/862/orig -> origin/gh/swolchok/862/orig 2025-12-04T09:20:35.7538555Z * [new branch] gh/swolchok/863/base -> origin/gh/swolchok/863/base 2025-12-04T09:20:35.7540167Z * [new branch] gh/swolchok/863/head -> origin/gh/swolchok/863/head 2025-12-04T09:20:35.7542356Z * [new branch] gh/swolchok/863/orig -> origin/gh/swolchok/863/orig 2025-12-04T09:20:35.7544889Z * [new branch] gh/swolchok/864/base -> origin/gh/swolchok/864/base 2025-12-04T09:20:35.7546813Z * [new branch] gh/swolchok/864/head -> origin/gh/swolchok/864/head 2025-12-04T09:20:35.7548639Z * [new branch] gh/swolchok/864/orig -> origin/gh/swolchok/864/orig 2025-12-04T09:20:35.7551203Z * [new branch] gh/swolchok/865/base -> origin/gh/swolchok/865/base 2025-12-04T09:20:35.7552892Z * [new branch] gh/swolchok/865/head -> origin/gh/swolchok/865/head 2025-12-04T09:20:35.7554562Z * [new branch] gh/swolchok/865/orig -> origin/gh/swolchok/865/orig 2025-12-04T09:20:35.7558043Z * [new branch] gh/swolchok/866/base -> origin/gh/swolchok/866/base 2025-12-04T09:20:35.7559539Z * [new branch] gh/swolchok/866/head -> origin/gh/swolchok/866/head 2025-12-04T09:20:35.7561910Z * [new branch] gh/swolchok/866/orig -> origin/gh/swolchok/866/orig 2025-12-04T09:20:35.7564262Z * [new branch] gh/swolchok/867/base -> origin/gh/swolchok/867/base 2025-12-04T09:20:35.7565799Z * [new branch] gh/swolchok/867/head -> origin/gh/swolchok/867/head 2025-12-04T09:20:35.7567575Z * [new branch] gh/swolchok/867/orig -> origin/gh/swolchok/867/orig 2025-12-04T09:20:35.7570256Z * [new branch] gh/swolchok/868/base -> origin/gh/swolchok/868/base 2025-12-04T09:20:35.7571961Z * [new branch] gh/swolchok/868/head -> origin/gh/swolchok/868/head 2025-12-04T09:20:35.7573608Z * [new branch] gh/swolchok/868/orig -> origin/gh/swolchok/868/orig 2025-12-04T09:20:35.7576359Z * [new branch] gh/swolchok/869/base -> origin/gh/swolchok/869/base 2025-12-04T09:20:35.7577802Z * [new branch] gh/swolchok/869/head -> origin/gh/swolchok/869/head 2025-12-04T09:20:35.7579658Z * [new branch] gh/swolchok/869/orig -> origin/gh/swolchok/869/orig 2025-12-04T09:20:35.7582557Z * [new branch] gh/swolchok/870/base -> origin/gh/swolchok/870/base 2025-12-04T09:20:35.7583871Z * [new branch] gh/swolchok/870/head -> origin/gh/swolchok/870/head 2025-12-04T09:20:35.7585643Z * [new branch] gh/swolchok/870/orig -> origin/gh/swolchok/870/orig 2025-12-04T09:20:35.7588337Z * [new branch] gh/swolchok/871/base -> origin/gh/swolchok/871/base 2025-12-04T09:20:35.7589940Z * [new branch] gh/swolchok/871/head -> origin/gh/swolchok/871/head 2025-12-04T09:20:35.7592354Z * [new branch] gh/swolchok/871/orig -> origin/gh/swolchok/871/orig 2025-12-04T09:20:35.7595297Z * [new branch] gh/teja-rao/4/base -> origin/gh/teja-rao/4/base 2025-12-04T09:20:35.7596746Z * [new branch] gh/teja-rao/4/head -> origin/gh/teja-rao/4/head 2025-12-04T09:20:35.7598471Z * [new branch] gh/teja-rao/4/orig -> origin/gh/teja-rao/4/orig 2025-12-04T09:20:35.7601673Z * [new branch] gh/tianyu-l/2/base -> origin/gh/tianyu-l/2/base 2025-12-04T09:20:35.7603102Z * [new branch] gh/tianyu-l/2/head -> origin/gh/tianyu-l/2/head 2025-12-04T09:20:35.7604839Z * [new branch] gh/tianyu-l/2/orig -> origin/gh/tianyu-l/2/orig 2025-12-04T09:20:35.7607530Z * [new branch] gh/tianyu-l/3/base -> origin/gh/tianyu-l/3/base 2025-12-04T09:20:35.7609134Z * [new branch] gh/tianyu-l/3/orig -> origin/gh/tianyu-l/3/orig 2025-12-04T09:20:35.7611847Z * [new branch] gh/tianyu-l/4/base -> origin/gh/tianyu-l/4/base 2025-12-04T09:20:35.7613292Z * [new branch] gh/tianyu-l/4/head -> origin/gh/tianyu-l/4/head 2025-12-04T09:20:35.7614966Z * [new branch] gh/tianyu-l/4/orig -> origin/gh/tianyu-l/4/orig 2025-12-04T09:20:35.7618696Z * [new branch] gh/tugsbayasgalan/10/base -> origin/gh/tugsbayasgalan/10/base 2025-12-04T09:20:35.7619996Z * [new branch] gh/tugsbayasgalan/10/head -> origin/gh/tugsbayasgalan/10/head 2025-12-04T09:20:35.7621711Z * [new branch] gh/tugsbayasgalan/10/orig -> origin/gh/tugsbayasgalan/10/orig 2025-12-04T09:20:35.7624418Z * [new branch] gh/tugsbayasgalan/13/base -> origin/gh/tugsbayasgalan/13/base 2025-12-04T09:20:35.7633618Z * [new branch] gh/tugsbayasgalan/13/head -> origin/gh/tugsbayasgalan/13/head 2025-12-04T09:20:35.7634544Z * [new branch] gh/tugsbayasgalan/13/orig -> origin/gh/tugsbayasgalan/13/orig 2025-12-04T09:20:35.7635358Z * [new branch] gh/tugsbayasgalan/17/base -> origin/gh/tugsbayasgalan/17/base 2025-12-04T09:20:35.7636436Z * [new branch] gh/tugsbayasgalan/17/head -> origin/gh/tugsbayasgalan/17/head 2025-12-04T09:20:35.7637250Z * [new branch] gh/tugsbayasgalan/17/orig -> origin/gh/tugsbayasgalan/17/orig 2025-12-04T09:20:35.7638141Z * [new branch] gh/tugsbayasgalan/2/base -> origin/gh/tugsbayasgalan/2/base 2025-12-04T09:20:35.7638961Z * [new branch] gh/tugsbayasgalan/2/head -> origin/gh/tugsbayasgalan/2/head 2025-12-04T09:20:35.7639780Z * [new branch] gh/tugsbayasgalan/2/orig -> origin/gh/tugsbayasgalan/2/orig 2025-12-04T09:20:35.7642510Z * [new branch] gh/tugsbayasgalan/28/base -> origin/gh/tugsbayasgalan/28/base 2025-12-04T09:20:35.7644019Z * [new branch] gh/tugsbayasgalan/28/head -> origin/gh/tugsbayasgalan/28/head 2025-12-04T09:20:35.7645807Z * [new branch] gh/tugsbayasgalan/28/orig -> origin/gh/tugsbayasgalan/28/orig 2025-12-04T09:20:35.7648450Z * [new branch] gh/tugsbayasgalan/32/base -> origin/gh/tugsbayasgalan/32/base 2025-12-04T09:20:35.7649931Z * [new branch] gh/tugsbayasgalan/32/head -> origin/gh/tugsbayasgalan/32/head 2025-12-04T09:20:35.7651854Z * [new branch] gh/tugsbayasgalan/32/orig -> origin/gh/tugsbayasgalan/32/orig 2025-12-04T09:20:35.7654481Z * [new branch] gh/tugsbayasgalan/35/base -> origin/gh/tugsbayasgalan/35/base 2025-12-04T09:20:35.7655910Z * [new branch] gh/tugsbayasgalan/35/head -> origin/gh/tugsbayasgalan/35/head 2025-12-04T09:20:35.7657650Z * [new branch] gh/tugsbayasgalan/35/orig -> origin/gh/tugsbayasgalan/35/orig 2025-12-04T09:20:35.7660439Z * [new branch] gh/tugsbayasgalan/36/base -> origin/gh/tugsbayasgalan/36/base 2025-12-04T09:20:35.7661938Z * [new branch] gh/tugsbayasgalan/36/head -> origin/gh/tugsbayasgalan/36/head 2025-12-04T09:20:35.7663610Z * [new branch] gh/tugsbayasgalan/36/orig -> origin/gh/tugsbayasgalan/36/orig 2025-12-04T09:20:35.7666248Z * [new branch] gh/tugsbayasgalan/37/base -> origin/gh/tugsbayasgalan/37/base 2025-12-04T09:20:35.7667786Z * [new branch] gh/tugsbayasgalan/37/head -> origin/gh/tugsbayasgalan/37/head 2025-12-04T09:20:35.7669483Z * [new branch] gh/tugsbayasgalan/37/orig -> origin/gh/tugsbayasgalan/37/orig 2025-12-04T09:20:35.7672327Z * [new branch] gh/tugsbayasgalan/43/base -> origin/gh/tugsbayasgalan/43/base 2025-12-04T09:20:35.7673857Z * [new branch] gh/tugsbayasgalan/43/head -> origin/gh/tugsbayasgalan/43/head 2025-12-04T09:20:35.7675632Z * [new branch] gh/tugsbayasgalan/43/orig -> origin/gh/tugsbayasgalan/43/orig 2025-12-04T09:20:35.7678117Z * [new branch] gh/tugsbayasgalan/48/base -> origin/gh/tugsbayasgalan/48/base 2025-12-04T09:20:35.7679447Z * [new branch] gh/tugsbayasgalan/48/head -> origin/gh/tugsbayasgalan/48/head 2025-12-04T09:20:35.7681246Z * [new branch] gh/tugsbayasgalan/48/orig -> origin/gh/tugsbayasgalan/48/orig 2025-12-04T09:20:35.7683587Z * [new branch] gh/tugsbayasgalan/51/base -> origin/gh/tugsbayasgalan/51/base 2025-12-04T09:20:35.7685473Z * [new branch] gh/tugsbayasgalan/51/head -> origin/gh/tugsbayasgalan/51/head 2025-12-04T09:20:35.7687047Z * [new branch] gh/tugsbayasgalan/51/orig -> origin/gh/tugsbayasgalan/51/orig 2025-12-04T09:20:35.7689244Z * [new branch] gh/tugsbayasgalan/52/base -> origin/gh/tugsbayasgalan/52/base 2025-12-04T09:20:35.7691063Z * [new branch] gh/tugsbayasgalan/52/head -> origin/gh/tugsbayasgalan/52/head 2025-12-04T09:20:35.7692770Z * [new branch] gh/tugsbayasgalan/52/orig -> origin/gh/tugsbayasgalan/52/orig 2025-12-04T09:20:35.7695183Z * [new branch] gh/tugsbayasgalan/53/base -> origin/gh/tugsbayasgalan/53/base 2025-12-04T09:20:35.7696956Z * [new branch] gh/tugsbayasgalan/53/head -> origin/gh/tugsbayasgalan/53/head 2025-12-04T09:20:35.7698771Z * [new branch] gh/tugsbayasgalan/53/orig -> origin/gh/tugsbayasgalan/53/orig 2025-12-04T09:20:35.7701323Z * [new branch] gh/tugsbayasgalan/55/base -> origin/gh/tugsbayasgalan/55/base 2025-12-04T09:20:35.7703196Z * [new branch] gh/tugsbayasgalan/55/head -> origin/gh/tugsbayasgalan/55/head 2025-12-04T09:20:35.7705019Z * [new branch] gh/tugsbayasgalan/55/orig -> origin/gh/tugsbayasgalan/55/orig 2025-12-04T09:20:35.7707581Z * [new branch] gh/tugsbayasgalan/59/base -> origin/gh/tugsbayasgalan/59/base 2025-12-04T09:20:35.7709300Z * [new branch] gh/tugsbayasgalan/59/head -> origin/gh/tugsbayasgalan/59/head 2025-12-04T09:20:35.7711042Z * [new branch] gh/tugsbayasgalan/59/orig -> origin/gh/tugsbayasgalan/59/orig 2025-12-04T09:20:35.7713367Z * [new branch] gh/tugsbayasgalan/6/base -> origin/gh/tugsbayasgalan/6/base 2025-12-04T09:20:35.7715096Z * [new branch] gh/tugsbayasgalan/6/head -> origin/gh/tugsbayasgalan/6/head 2025-12-04T09:20:35.7716769Z * [new branch] gh/tugsbayasgalan/6/orig -> origin/gh/tugsbayasgalan/6/orig 2025-12-04T09:20:35.7718967Z * [new branch] gh/tugsbayasgalan/60/base -> origin/gh/tugsbayasgalan/60/base 2025-12-04T09:20:35.7720710Z * [new branch] gh/tugsbayasgalan/60/head -> origin/gh/tugsbayasgalan/60/head 2025-12-04T09:20:35.7722528Z * [new branch] gh/tugsbayasgalan/60/orig -> origin/gh/tugsbayasgalan/60/orig 2025-12-04T09:20:35.7725419Z * [new branch] gh/tugsbayasgalan/61/base -> origin/gh/tugsbayasgalan/61/base 2025-12-04T09:20:35.7727125Z * [new branch] gh/tugsbayasgalan/61/head -> origin/gh/tugsbayasgalan/61/head 2025-12-04T09:20:35.7728867Z * [new branch] gh/tugsbayasgalan/61/orig -> origin/gh/tugsbayasgalan/61/orig 2025-12-04T09:20:35.7731377Z * [new branch] gh/tugsbayasgalan/63/base -> origin/gh/tugsbayasgalan/63/base 2025-12-04T09:20:35.7733106Z * [new branch] gh/tugsbayasgalan/63/head -> origin/gh/tugsbayasgalan/63/head 2025-12-04T09:20:35.7734878Z * [new branch] gh/tugsbayasgalan/63/orig -> origin/gh/tugsbayasgalan/63/orig 2025-12-04T09:20:35.7737315Z * [new branch] gh/tugsbayasgalan/67/base -> origin/gh/tugsbayasgalan/67/base 2025-12-04T09:20:35.7739059Z * [new branch] gh/tugsbayasgalan/67/head -> origin/gh/tugsbayasgalan/67/head 2025-12-04T09:20:35.7740945Z * [new branch] gh/tugsbayasgalan/67/orig -> origin/gh/tugsbayasgalan/67/orig 2025-12-04T09:20:35.7743448Z * [new branch] gh/tugsbayasgalan/68/base -> origin/gh/tugsbayasgalan/68/base 2025-12-04T09:20:35.7745181Z * [new branch] gh/tugsbayasgalan/68/head -> origin/gh/tugsbayasgalan/68/head 2025-12-04T09:20:35.7747018Z * [new branch] gh/tugsbayasgalan/68/orig -> origin/gh/tugsbayasgalan/68/orig 2025-12-04T09:20:35.7749446Z * [new branch] gh/tugsbayasgalan/7/base -> origin/gh/tugsbayasgalan/7/base 2025-12-04T09:20:35.7751183Z * [new branch] gh/tugsbayasgalan/7/head -> origin/gh/tugsbayasgalan/7/head 2025-12-04T09:20:35.7752999Z * [new branch] gh/tugsbayasgalan/7/orig -> origin/gh/tugsbayasgalan/7/orig 2025-12-04T09:20:35.7755636Z * [new branch] gh/tugsbayasgalan/70/base -> origin/gh/tugsbayasgalan/70/base 2025-12-04T09:20:35.7757511Z * [new branch] gh/tugsbayasgalan/70/head -> origin/gh/tugsbayasgalan/70/head 2025-12-04T09:20:35.7759338Z * [new branch] gh/tugsbayasgalan/70/orig -> origin/gh/tugsbayasgalan/70/orig 2025-12-04T09:20:35.7761976Z * [new branch] gh/tugsbayasgalan/71/base -> origin/gh/tugsbayasgalan/71/base 2025-12-04T09:20:35.7763802Z * [new branch] gh/tugsbayasgalan/71/head -> origin/gh/tugsbayasgalan/71/head 2025-12-04T09:20:35.7765600Z * [new branch] gh/tugsbayasgalan/71/orig -> origin/gh/tugsbayasgalan/71/orig 2025-12-04T09:20:35.7768145Z * [new branch] gh/tugsbayasgalan/72/base -> origin/gh/tugsbayasgalan/72/base 2025-12-04T09:20:35.7769883Z * [new branch] gh/tugsbayasgalan/72/head -> origin/gh/tugsbayasgalan/72/head 2025-12-04T09:20:35.7771839Z * [new branch] gh/tugsbayasgalan/72/orig -> origin/gh/tugsbayasgalan/72/orig 2025-12-04T09:20:35.7774351Z * [new branch] gh/tugsbayasgalan/73/base -> origin/gh/tugsbayasgalan/73/base 2025-12-04T09:20:35.7776160Z * [new branch] gh/tugsbayasgalan/73/head -> origin/gh/tugsbayasgalan/73/head 2025-12-04T09:20:35.7777928Z * [new branch] gh/tugsbayasgalan/73/orig -> origin/gh/tugsbayasgalan/73/orig 2025-12-04T09:20:35.7780569Z * [new branch] gh/tugsbayasgalan/74/base -> origin/gh/tugsbayasgalan/74/base 2025-12-04T09:20:35.7782417Z * [new branch] gh/tugsbayasgalan/74/head -> origin/gh/tugsbayasgalan/74/head 2025-12-04T09:20:35.7784177Z * [new branch] gh/tugsbayasgalan/74/orig -> origin/gh/tugsbayasgalan/74/orig 2025-12-04T09:20:35.7786675Z * [new branch] gh/tugsbayasgalan/75/base -> origin/gh/tugsbayasgalan/75/base 2025-12-04T09:20:35.7788364Z * [new branch] gh/tugsbayasgalan/75/head -> origin/gh/tugsbayasgalan/75/head 2025-12-04T09:20:35.7790075Z * [new branch] gh/tugsbayasgalan/75/orig -> origin/gh/tugsbayasgalan/75/orig 2025-12-04T09:20:35.7792762Z * [new branch] gh/tugsbayasgalan/76/base -> origin/gh/tugsbayasgalan/76/base 2025-12-04T09:20:35.7794586Z * [new branch] gh/tugsbayasgalan/76/head -> origin/gh/tugsbayasgalan/76/head 2025-12-04T09:20:35.7796392Z * [new branch] gh/tugsbayasgalan/76/orig -> origin/gh/tugsbayasgalan/76/orig 2025-12-04T09:20:35.7799093Z * [new branch] gh/tugsbayasgalan/77/base -> origin/gh/tugsbayasgalan/77/base 2025-12-04T09:20:35.7800781Z * [new branch] gh/tugsbayasgalan/77/head -> origin/gh/tugsbayasgalan/77/head 2025-12-04T09:20:35.7802622Z * [new branch] gh/tugsbayasgalan/77/orig -> origin/gh/tugsbayasgalan/77/orig 2025-12-04T09:20:35.7804978Z * [new branch] gh/tugsbayasgalan/78/base -> origin/gh/tugsbayasgalan/78/base 2025-12-04T09:20:35.7806842Z * [new branch] gh/tugsbayasgalan/78/head -> origin/gh/tugsbayasgalan/78/head 2025-12-04T09:20:35.7808576Z * [new branch] gh/tugsbayasgalan/78/orig -> origin/gh/tugsbayasgalan/78/orig 2025-12-04T09:20:35.7811036Z * [new branch] gh/tugsbayasgalan/79/base -> origin/gh/tugsbayasgalan/79/base 2025-12-04T09:20:35.7812867Z * [new branch] gh/tugsbayasgalan/79/head -> origin/gh/tugsbayasgalan/79/head 2025-12-04T09:20:35.7814575Z * [new branch] gh/tugsbayasgalan/79/orig -> origin/gh/tugsbayasgalan/79/orig 2025-12-04T09:20:35.7817003Z * [new branch] gh/tugsbayasgalan/8/base -> origin/gh/tugsbayasgalan/8/base 2025-12-04T09:20:35.7818696Z * [new branch] gh/tugsbayasgalan/8/head -> origin/gh/tugsbayasgalan/8/head 2025-12-04T09:20:35.7820826Z * [new branch] gh/tugsbayasgalan/8/orig -> origin/gh/tugsbayasgalan/8/orig 2025-12-04T09:20:35.7823022Z * [new branch] gh/tugsbayasgalan/80/base -> origin/gh/tugsbayasgalan/80/base 2025-12-04T09:20:35.7824712Z * [new branch] gh/tugsbayasgalan/80/head -> origin/gh/tugsbayasgalan/80/head 2025-12-04T09:20:35.7826393Z * [new branch] gh/tugsbayasgalan/80/orig -> origin/gh/tugsbayasgalan/80/orig 2025-12-04T09:20:35.7828873Z * [new branch] gh/tugsbayasgalan/81/base -> origin/gh/tugsbayasgalan/81/base 2025-12-04T09:20:35.7830572Z * [new branch] gh/tugsbayasgalan/81/head -> origin/gh/tugsbayasgalan/81/head 2025-12-04T09:20:35.7832350Z * [new branch] gh/tugsbayasgalan/81/orig -> origin/gh/tugsbayasgalan/81/orig 2025-12-04T09:20:35.7835200Z * [new branch] gh/tugsbayasgalan/82/base -> origin/gh/tugsbayasgalan/82/base 2025-12-04T09:20:35.7837048Z * [new branch] gh/tugsbayasgalan/82/head -> origin/gh/tugsbayasgalan/82/head 2025-12-04T09:20:35.7838830Z * [new branch] gh/tugsbayasgalan/82/orig -> origin/gh/tugsbayasgalan/82/orig 2025-12-04T09:20:35.7841147Z * [new branch] gh/tugsbayasgalan/83/base -> origin/gh/tugsbayasgalan/83/base 2025-12-04T09:20:35.7842862Z * [new branch] gh/tugsbayasgalan/83/head -> origin/gh/tugsbayasgalan/83/head 2025-12-04T09:20:35.7844672Z * [new branch] gh/tugsbayasgalan/83/orig -> origin/gh/tugsbayasgalan/83/orig 2025-12-04T09:20:35.7846984Z * [new branch] gh/tugsbayasgalan/84/base -> origin/gh/tugsbayasgalan/84/base 2025-12-04T09:20:35.7848772Z * [new branch] gh/tugsbayasgalan/84/head -> origin/gh/tugsbayasgalan/84/head 2025-12-04T09:20:35.7850528Z * [new branch] gh/tugsbayasgalan/84/orig -> origin/gh/tugsbayasgalan/84/orig 2025-12-04T09:20:35.7852829Z * [new branch] gh/tugsbayasgalan/85/base -> origin/gh/tugsbayasgalan/85/base 2025-12-04T09:20:35.7854606Z * [new branch] gh/tugsbayasgalan/85/head -> origin/gh/tugsbayasgalan/85/head 2025-12-04T09:20:35.7856321Z * [new branch] gh/tugsbayasgalan/85/orig -> origin/gh/tugsbayasgalan/85/orig 2025-12-04T09:20:35.7858761Z * [new branch] gh/tugsbayasgalan/86/base -> origin/gh/tugsbayasgalan/86/base 2025-12-04T09:20:35.7860676Z * [new branch] gh/tugsbayasgalan/86/head -> origin/gh/tugsbayasgalan/86/head 2025-12-04T09:20:35.7862392Z * [new branch] gh/tugsbayasgalan/86/orig -> origin/gh/tugsbayasgalan/86/orig 2025-12-04T09:20:35.7865158Z * [new branch] gh/tugsbayasgalan/87/base -> origin/gh/tugsbayasgalan/87/base 2025-12-04T09:20:35.7866874Z * [new branch] gh/tugsbayasgalan/87/head -> origin/gh/tugsbayasgalan/87/head 2025-12-04T09:20:35.7868659Z * [new branch] gh/tugsbayasgalan/87/orig -> origin/gh/tugsbayasgalan/87/orig 2025-12-04T09:20:35.7871434Z * [new branch] gh/tugsbayasgalan/88/base -> origin/gh/tugsbayasgalan/88/base 2025-12-04T09:20:35.7874692Z * [new branch] gh/tugsbayasgalan/88/head -> origin/gh/tugsbayasgalan/88/head 2025-12-04T09:20:35.7876391Z * [new branch] gh/tugsbayasgalan/88/orig -> origin/gh/tugsbayasgalan/88/orig 2025-12-04T09:20:35.7878949Z * [new branch] gh/tugsbayasgalan/89/base -> origin/gh/tugsbayasgalan/89/base 2025-12-04T09:20:35.7880714Z * [new branch] gh/tugsbayasgalan/89/head -> origin/gh/tugsbayasgalan/89/head 2025-12-04T09:20:35.7882479Z * [new branch] gh/tugsbayasgalan/89/orig -> origin/gh/tugsbayasgalan/89/orig 2025-12-04T09:20:35.7884878Z * [new branch] gh/tugsbayasgalan/9/base -> origin/gh/tugsbayasgalan/9/base 2025-12-04T09:20:35.7886479Z * [new branch] gh/tugsbayasgalan/9/head -> origin/gh/tugsbayasgalan/9/head 2025-12-04T09:20:35.7888250Z * [new branch] gh/tugsbayasgalan/9/orig -> origin/gh/tugsbayasgalan/9/orig 2025-12-04T09:20:35.7890959Z * [new branch] gh/tugsbayasgalan/90/base -> origin/gh/tugsbayasgalan/90/base 2025-12-04T09:20:35.7892537Z * [new branch] gh/tugsbayasgalan/90/head -> origin/gh/tugsbayasgalan/90/head 2025-12-04T09:20:35.7894341Z * [new branch] gh/tugsbayasgalan/90/orig -> origin/gh/tugsbayasgalan/90/orig 2025-12-04T09:20:35.7896905Z * [new branch] gh/tugsbayasgalan/91/base -> origin/gh/tugsbayasgalan/91/base 2025-12-04T09:20:35.7898592Z * [new branch] gh/tugsbayasgalan/91/head -> origin/gh/tugsbayasgalan/91/head 2025-12-04T09:20:35.7900454Z * [new branch] gh/tugsbayasgalan/91/orig -> origin/gh/tugsbayasgalan/91/orig 2025-12-04T09:20:35.7902754Z * [new branch] gh/tugsbayasgalan/92/base -> origin/gh/tugsbayasgalan/92/base 2025-12-04T09:20:35.7904588Z * [new branch] gh/tugsbayasgalan/92/head -> origin/gh/tugsbayasgalan/92/head 2025-12-04T09:20:35.7906320Z * [new branch] gh/tugsbayasgalan/92/orig -> origin/gh/tugsbayasgalan/92/orig 2025-12-04T09:20:35.7908947Z * [new branch] gh/tugsbayasgalan/93/base -> origin/gh/tugsbayasgalan/93/base 2025-12-04T09:20:35.7910710Z * [new branch] gh/tugsbayasgalan/93/head -> origin/gh/tugsbayasgalan/93/head 2025-12-04T09:20:35.7912576Z * [new branch] gh/tugsbayasgalan/93/orig -> origin/gh/tugsbayasgalan/93/orig 2025-12-04T09:20:35.7915477Z * [new branch] gh/v0i0/14/base -> origin/gh/v0i0/14/base 2025-12-04T09:20:35.7917209Z * [new branch] gh/v0i0/14/head -> origin/gh/v0i0/14/head 2025-12-04T09:20:35.7918936Z * [new branch] gh/v0i0/14/orig -> origin/gh/v0i0/14/orig 2025-12-04T09:20:35.7921619Z * [new branch] gh/v0i0/15/base -> origin/gh/v0i0/15/base 2025-12-04T09:20:35.7923506Z * [new branch] gh/v0i0/15/head -> origin/gh/v0i0/15/head 2025-12-04T09:20:35.7925248Z * [new branch] gh/v0i0/15/orig -> origin/gh/v0i0/15/orig 2025-12-04T09:20:35.7927623Z * [new branch] gh/v0i0/16/base -> origin/gh/v0i0/16/base 2025-12-04T09:20:35.7929349Z * [new branch] gh/v0i0/16/head -> origin/gh/v0i0/16/head 2025-12-04T09:20:35.7931129Z * [new branch] gh/v0i0/16/orig -> origin/gh/v0i0/16/orig 2025-12-04T09:20:35.7933417Z * [new branch] gh/v0i0/17/base -> origin/gh/v0i0/17/base 2025-12-04T09:20:35.7935131Z * [new branch] gh/v0i0/17/head -> origin/gh/v0i0/17/head 2025-12-04T09:20:35.7936865Z * [new branch] gh/v0i0/17/orig -> origin/gh/v0i0/17/orig 2025-12-04T09:20:35.7939373Z * [new branch] gh/v0i0/18/base -> origin/gh/v0i0/18/base 2025-12-04T09:20:35.7941260Z * [new branch] gh/v0i0/18/head -> origin/gh/v0i0/18/head 2025-12-04T09:20:35.7943035Z * [new branch] gh/v0i0/18/orig -> origin/gh/v0i0/18/orig 2025-12-04T09:20:35.7945360Z * [new branch] gh/v0i0/19/base -> origin/gh/v0i0/19/base 2025-12-04T09:20:35.7947094Z * [new branch] gh/v0i0/19/head -> origin/gh/v0i0/19/head 2025-12-04T09:20:35.7948806Z * [new branch] gh/v0i0/19/orig -> origin/gh/v0i0/19/orig 2025-12-04T09:20:35.7952337Z * [new branch] gh/vishal9-team/1/base -> origin/gh/vishal9-team/1/base 2025-12-04T09:20:35.7954087Z * [new branch] gh/vishal9-team/1/head -> origin/gh/vishal9-team/1/head 2025-12-04T09:20:35.7956319Z * [new branch] gh/vishal9-team/2/base -> origin/gh/vishal9-team/2/base 2025-12-04T09:20:35.7958080Z * [new branch] gh/vishal9-team/2/head -> origin/gh/vishal9-team/2/head 2025-12-04T09:20:35.7959856Z * [new branch] gh/vishal9-team/2/orig -> origin/gh/vishal9-team/2/orig 2025-12-04T09:20:35.7962329Z * [new branch] gh/vishal9-team/3/base -> origin/gh/vishal9-team/3/base 2025-12-04T09:20:35.7964065Z * [new branch] gh/vishal9-team/3/head -> origin/gh/vishal9-team/3/head 2025-12-04T09:20:35.7965800Z * [new branch] gh/vishal9-team/3/orig -> origin/gh/vishal9-team/3/orig 2025-12-04T09:20:35.7968030Z * [new branch] gh/vishal9-team/4/base -> origin/gh/vishal9-team/4/base 2025-12-04T09:20:35.7969803Z * [new branch] gh/vishal9-team/4/head -> origin/gh/vishal9-team/4/head 2025-12-04T09:20:35.7971687Z * [new branch] gh/vishal9-team/4/orig -> origin/gh/vishal9-team/4/orig 2025-12-04T09:20:35.7974494Z * [new branch] gh/vkuzo/1/next -> origin/gh/vkuzo/1/next 2025-12-04T09:20:35.7976787Z * [new branch] gh/vkuzo/2/next -> origin/gh/vkuzo/2/next 2025-12-04T09:20:35.7979194Z * [new branch] gh/vkuzo/3/next -> origin/gh/vkuzo/3/next 2025-12-04T09:20:35.7982191Z * [new branch] gh/wconstab/424/base -> origin/gh/wconstab/424/base 2025-12-04T09:20:35.7984424Z * [new branch] gh/wconstab/424/head -> origin/gh/wconstab/424/head 2025-12-04T09:20:35.7986314Z * [new branch] gh/wconstab/424/orig -> origin/gh/wconstab/424/orig 2025-12-04T09:20:35.7989173Z * [new branch] gh/wconstab/435/base -> origin/gh/wconstab/435/base 2025-12-04T09:20:35.7990917Z * [new branch] gh/wconstab/435/head -> origin/gh/wconstab/435/head 2025-12-04T09:20:35.7992660Z * [new branch] gh/wconstab/435/orig -> origin/gh/wconstab/435/orig 2025-12-04T09:20:35.7995012Z * [new branch] gh/wconstab/444/base -> origin/gh/wconstab/444/base 2025-12-04T09:20:35.7996832Z * [new branch] gh/wconstab/444/head -> origin/gh/wconstab/444/head 2025-12-04T09:20:35.7998589Z * [new branch] gh/wconstab/444/orig -> origin/gh/wconstab/444/orig 2025-12-04T09:20:35.8000950Z * [new branch] gh/wconstab/447/base -> origin/gh/wconstab/447/base 2025-12-04T09:20:35.8002679Z * [new branch] gh/wconstab/447/head -> origin/gh/wconstab/447/head 2025-12-04T09:20:35.8004450Z * [new branch] gh/wconstab/447/orig -> origin/gh/wconstab/447/orig 2025-12-04T09:20:35.8006741Z * [new branch] gh/wconstab/448/base -> origin/gh/wconstab/448/base 2025-12-04T09:20:35.8008555Z * [new branch] gh/wconstab/448/head -> origin/gh/wconstab/448/head 2025-12-04T09:20:35.8010528Z * [new branch] gh/wconstab/448/orig -> origin/gh/wconstab/448/orig 2025-12-04T09:20:35.8012737Z * [new branch] gh/wconstab/449/base -> origin/gh/wconstab/449/base 2025-12-04T09:20:35.8014507Z * [new branch] gh/wconstab/449/head -> origin/gh/wconstab/449/head 2025-12-04T09:20:35.8016306Z * [new branch] gh/wconstab/449/orig -> origin/gh/wconstab/449/orig 2025-12-04T09:20:35.8018508Z * [new branch] gh/wconstab/450/base -> origin/gh/wconstab/450/base 2025-12-04T09:20:35.8020424Z * [new branch] gh/wconstab/450/head -> origin/gh/wconstab/450/head 2025-12-04T09:20:35.8022116Z * [new branch] gh/wconstab/450/orig -> origin/gh/wconstab/450/orig 2025-12-04T09:20:35.8024347Z * [new branch] gh/wconstab/451/base -> origin/gh/wconstab/451/base 2025-12-04T09:20:35.8026210Z * [new branch] gh/wconstab/451/head -> origin/gh/wconstab/451/head 2025-12-04T09:20:35.8027921Z * [new branch] gh/wconstab/451/orig -> origin/gh/wconstab/451/orig 2025-12-04T09:20:35.8030325Z * [new branch] gh/wconstab/452/base -> origin/gh/wconstab/452/base 2025-12-04T09:20:35.8032090Z * [new branch] gh/wconstab/452/head -> origin/gh/wconstab/452/head 2025-12-04T09:20:35.8033977Z * [new branch] gh/wconstab/452/orig -> origin/gh/wconstab/452/orig 2025-12-04T09:20:35.8036153Z * [new branch] gh/wconstab/453/base -> origin/gh/wconstab/453/base 2025-12-04T09:20:35.8039315Z * [new branch] gh/wconstab/453/head -> origin/gh/wconstab/453/head 2025-12-04T09:20:35.8040173Z * [new branch] gh/wconstab/453/orig -> origin/gh/wconstab/453/orig 2025-12-04T09:20:35.8041889Z * [new branch] gh/wconstab/454/base -> origin/gh/wconstab/454/base 2025-12-04T09:20:35.8043623Z * [new branch] gh/wconstab/454/head -> origin/gh/wconstab/454/head 2025-12-04T09:20:35.8045382Z * [new branch] gh/wconstab/454/orig -> origin/gh/wconstab/454/orig 2025-12-04T09:20:35.8047748Z * [new branch] gh/wconstab/455/base -> origin/gh/wconstab/455/base 2025-12-04T09:20:35.8049506Z * [new branch] gh/wconstab/455/head -> origin/gh/wconstab/455/head 2025-12-04T09:20:35.8051264Z * [new branch] gh/wconstab/455/orig -> origin/gh/wconstab/455/orig 2025-12-04T09:20:35.8053899Z * [new branch] gh/wconstab/456/base -> origin/gh/wconstab/456/base 2025-12-04T09:20:35.8055957Z * [new branch] gh/wconstab/456/head -> origin/gh/wconstab/456/head 2025-12-04T09:20:35.8057770Z * [new branch] gh/wconstab/456/orig -> origin/gh/wconstab/456/orig 2025-12-04T09:20:35.8060299Z * [new branch] gh/wconstab/457/base -> origin/gh/wconstab/457/base 2025-12-04T09:20:35.8062079Z * [new branch] gh/wconstab/457/head -> origin/gh/wconstab/457/head 2025-12-04T09:20:35.8063767Z * [new branch] gh/wconstab/457/orig -> origin/gh/wconstab/457/orig 2025-12-04T09:20:35.8066163Z * [new branch] gh/wconstab/458/base -> origin/gh/wconstab/458/base 2025-12-04T09:20:35.8067996Z * [new branch] gh/wconstab/458/head -> origin/gh/wconstab/458/head 2025-12-04T09:20:35.8069670Z * [new branch] gh/wconstab/458/orig -> origin/gh/wconstab/458/orig 2025-12-04T09:20:35.8072024Z * [new branch] gh/wconstab/459/base -> origin/gh/wconstab/459/base 2025-12-04T09:20:35.8073821Z * [new branch] gh/wconstab/459/head -> origin/gh/wconstab/459/head 2025-12-04T09:20:35.8075602Z * [new branch] gh/wconstab/459/orig -> origin/gh/wconstab/459/orig 2025-12-04T09:20:35.8078500Z * [new branch] gh/wconstab/460/base -> origin/gh/wconstab/460/base 2025-12-04T09:20:35.8080494Z * [new branch] gh/wconstab/460/head -> origin/gh/wconstab/460/head 2025-12-04T09:20:35.8082397Z * [new branch] gh/wconstab/460/orig -> origin/gh/wconstab/460/orig 2025-12-04T09:20:35.8084971Z * [new branch] gh/wconstab/461/base -> origin/gh/wconstab/461/base 2025-12-04T09:20:35.8086667Z * [new branch] gh/wconstab/461/head -> origin/gh/wconstab/461/head 2025-12-04T09:20:35.8088424Z * [new branch] gh/wconstab/461/orig -> origin/gh/wconstab/461/orig 2025-12-04T09:20:35.8090732Z * [new branch] gh/wconstab/462/base -> origin/gh/wconstab/462/base 2025-12-04T09:20:35.8092631Z * [new branch] gh/wconstab/462/head -> origin/gh/wconstab/462/head 2025-12-04T09:20:35.8094384Z * [new branch] gh/wconstab/462/orig -> origin/gh/wconstab/462/orig 2025-12-04T09:20:35.8096830Z * [new branch] gh/wconstab/463/base -> origin/gh/wconstab/463/base 2025-12-04T09:20:35.8098607Z * [new branch] gh/wconstab/463/head -> origin/gh/wconstab/463/head 2025-12-04T09:20:35.8100506Z * [new branch] gh/wconstab/463/orig -> origin/gh/wconstab/463/orig 2025-12-04T09:20:35.8102911Z * [new branch] gh/wconstab/464/base -> origin/gh/wconstab/464/base 2025-12-04T09:20:35.8104907Z * [new branch] gh/wconstab/464/head -> origin/gh/wconstab/464/head 2025-12-04T09:20:35.8106549Z * [new branch] gh/wconstab/464/orig -> origin/gh/wconstab/464/orig 2025-12-04T09:20:35.8108871Z * [new branch] gh/wconstab/465/base -> origin/gh/wconstab/465/base 2025-12-04T09:20:35.8110653Z * [new branch] gh/wconstab/465/head -> origin/gh/wconstab/465/head 2025-12-04T09:20:35.8112315Z * [new branch] gh/wconstab/465/orig -> origin/gh/wconstab/465/orig 2025-12-04T09:20:35.8114827Z * [new branch] gh/wconstab/466/base -> origin/gh/wconstab/466/base 2025-12-04T09:20:35.8116536Z * [new branch] gh/wconstab/466/head -> origin/gh/wconstab/466/head 2025-12-04T09:20:35.8118157Z * [new branch] gh/wconstab/466/orig -> origin/gh/wconstab/466/orig 2025-12-04T09:20:35.8120937Z * [new branch] gh/wconstab/467/base -> origin/gh/wconstab/467/base 2025-12-04T09:20:35.8122699Z * [new branch] gh/wconstab/467/head -> origin/gh/wconstab/467/head 2025-12-04T09:20:35.8124479Z * [new branch] gh/wconstab/467/orig -> origin/gh/wconstab/467/orig 2025-12-04T09:20:35.8126759Z * [new branch] gh/wconstab/468/base -> origin/gh/wconstab/468/base 2025-12-04T09:20:35.8128573Z * [new branch] gh/wconstab/468/head -> origin/gh/wconstab/468/head 2025-12-04T09:20:35.8130261Z * [new branch] gh/wconstab/468/orig -> origin/gh/wconstab/468/orig 2025-12-04T09:20:35.8133144Z * [new branch] gh/weifengpy/39/base -> origin/gh/weifengpy/39/base 2025-12-04T09:20:35.8134846Z * [new branch] gh/weifengpy/39/head -> origin/gh/weifengpy/39/head 2025-12-04T09:20:35.8136663Z * [new branch] gh/weifengpy/39/orig -> origin/gh/weifengpy/39/orig 2025-12-04T09:20:35.8139173Z * [new branch] gh/weifengpy/40/base -> origin/gh/weifengpy/40/base 2025-12-04T09:20:35.8140976Z * [new branch] gh/weifengpy/40/head -> origin/gh/weifengpy/40/head 2025-12-04T09:20:35.8142808Z * [new branch] gh/weifengpy/40/orig -> origin/gh/weifengpy/40/orig 2025-12-04T09:20:35.8145211Z * [new branch] gh/weifengpy/41/base -> origin/gh/weifengpy/41/base 2025-12-04T09:20:35.8147062Z * [new branch] gh/weifengpy/41/head -> origin/gh/weifengpy/41/head 2025-12-04T09:20:35.8148876Z * [new branch] gh/weifengpy/41/orig -> origin/gh/weifengpy/41/orig 2025-12-04T09:20:35.8151935Z * [new branch] gh/williamwen42/250/base -> origin/gh/williamwen42/250/base 2025-12-04T09:20:35.8153719Z * [new branch] gh/williamwen42/250/head -> origin/gh/williamwen42/250/head 2025-12-04T09:20:35.8155455Z * [new branch] gh/williamwen42/250/orig -> origin/gh/williamwen42/250/orig 2025-12-04T09:20:35.8158008Z * [new branch] gh/williamwen42/279/base -> origin/gh/williamwen42/279/base 2025-12-04T09:20:35.8159831Z * [new branch] gh/williamwen42/279/head -> origin/gh/williamwen42/279/head 2025-12-04T09:20:35.8161550Z * [new branch] gh/williamwen42/279/orig -> origin/gh/williamwen42/279/orig 2025-12-04T09:20:35.8163903Z * [new branch] gh/williamwen42/282/base -> origin/gh/williamwen42/282/base 2025-12-04T09:20:35.8165649Z * [new branch] gh/williamwen42/282/head -> origin/gh/williamwen42/282/head 2025-12-04T09:20:35.8167356Z * [new branch] gh/williamwen42/282/orig -> origin/gh/williamwen42/282/orig 2025-12-04T09:20:35.8169744Z * [new branch] gh/williamwen42/287/base -> origin/gh/williamwen42/287/base 2025-12-04T09:20:35.8171549Z * [new branch] gh/williamwen42/287/head -> origin/gh/williamwen42/287/head 2025-12-04T09:20:35.8173477Z * [new branch] gh/williamwen42/287/orig -> origin/gh/williamwen42/287/orig 2025-12-04T09:20:35.8175989Z * [new branch] gh/williamwen42/288/base -> origin/gh/williamwen42/288/base 2025-12-04T09:20:35.8177613Z * [new branch] gh/williamwen42/288/head -> origin/gh/williamwen42/288/head 2025-12-04T09:20:35.8179524Z * [new branch] gh/williamwen42/288/orig -> origin/gh/williamwen42/288/orig 2025-12-04T09:20:35.8182074Z * [new branch] gh/williamwen42/296/base -> origin/gh/williamwen42/296/base 2025-12-04T09:20:35.8183937Z * [new branch] gh/williamwen42/296/head -> origin/gh/williamwen42/296/head 2025-12-04T09:20:35.8185722Z * [new branch] gh/williamwen42/296/orig -> origin/gh/williamwen42/296/orig 2025-12-04T09:20:35.8188058Z * [new branch] gh/williamwen42/297/base -> origin/gh/williamwen42/297/base 2025-12-04T09:20:35.8189809Z * [new branch] gh/williamwen42/297/head -> origin/gh/williamwen42/297/head 2025-12-04T09:20:35.8191579Z * [new branch] gh/williamwen42/297/orig -> origin/gh/williamwen42/297/orig 2025-12-04T09:20:35.8194285Z * [new branch] gh/williamwen42/306/base -> origin/gh/williamwen42/306/base 2025-12-04T09:20:35.8196220Z * [new branch] gh/williamwen42/306/head -> origin/gh/williamwen42/306/head 2025-12-04T09:20:35.8197931Z * [new branch] gh/williamwen42/306/orig -> origin/gh/williamwen42/306/orig 2025-12-04T09:20:35.8200397Z * [new branch] gh/williamwen42/309/base -> origin/gh/williamwen42/309/base 2025-12-04T09:20:35.8202166Z * [new branch] gh/williamwen42/309/head -> origin/gh/williamwen42/309/head 2025-12-04T09:20:35.8203995Z * [new branch] gh/williamwen42/309/orig -> origin/gh/williamwen42/309/orig 2025-12-04T09:20:35.8206274Z * [new branch] gh/williamwen42/310/base -> origin/gh/williamwen42/310/base 2025-12-04T09:20:35.8208047Z * [new branch] gh/williamwen42/310/head -> origin/gh/williamwen42/310/head 2025-12-04T09:20:35.8209882Z * [new branch] gh/williamwen42/310/orig -> origin/gh/williamwen42/310/orig 2025-12-04T09:20:35.8213232Z * [new branch] gh/williamwen42/311/base -> origin/gh/williamwen42/311/base 2025-12-04T09:20:35.8214951Z * [new branch] gh/williamwen42/311/head -> origin/gh/williamwen42/311/head 2025-12-04T09:20:35.8216859Z * [new branch] gh/williamwen42/311/orig -> origin/gh/williamwen42/311/orig 2025-12-04T09:20:35.8219082Z * [new branch] gh/williamwen42/319/base -> origin/gh/williamwen42/319/base 2025-12-04T09:20:35.8221003Z * [new branch] gh/williamwen42/319/head -> origin/gh/williamwen42/319/head 2025-12-04T09:20:35.8222700Z * [new branch] gh/williamwen42/319/orig -> origin/gh/williamwen42/319/orig 2025-12-04T09:20:35.8225136Z * [new branch] gh/williamwen42/325/base -> origin/gh/williamwen42/325/base 2025-12-04T09:20:35.8226979Z * [new branch] gh/williamwen42/325/head -> origin/gh/williamwen42/325/head 2025-12-04T09:20:35.8228692Z * [new branch] gh/williamwen42/325/orig -> origin/gh/williamwen42/325/orig 2025-12-04T09:20:35.8231005Z * [new branch] gh/williamwen42/326/base -> origin/gh/williamwen42/326/base 2025-12-04T09:20:35.8232849Z * [new branch] gh/williamwen42/326/head -> origin/gh/williamwen42/326/head 2025-12-04T09:20:35.8234637Z * [new branch] gh/williamwen42/326/orig -> origin/gh/williamwen42/326/orig 2025-12-04T09:20:35.8237021Z * [new branch] gh/williamwen42/327/base -> origin/gh/williamwen42/327/base 2025-12-04T09:20:35.8238705Z * [new branch] gh/williamwen42/327/head -> origin/gh/williamwen42/327/head 2025-12-04T09:20:35.8240845Z * [new branch] gh/williamwen42/327/orig -> origin/gh/williamwen42/327/orig 2025-12-04T09:20:35.8243291Z * [new branch] gh/williamwen42/328/base -> origin/gh/williamwen42/328/base 2025-12-04T09:20:35.8245157Z * [new branch] gh/williamwen42/328/head -> origin/gh/williamwen42/328/head 2025-12-04T09:20:35.8246817Z * [new branch] gh/williamwen42/328/orig -> origin/gh/williamwen42/328/orig 2025-12-04T09:20:35.8250005Z * [new branch] gh/williamwen42/329/base -> origin/gh/williamwen42/329/base 2025-12-04T09:20:35.8251870Z * [new branch] gh/williamwen42/329/head -> origin/gh/williamwen42/329/head 2025-12-04T09:20:35.8253506Z * [new branch] gh/williamwen42/329/orig -> origin/gh/williamwen42/329/orig 2025-12-04T09:20:35.8255908Z * [new branch] gh/williamwen42/330/base -> origin/gh/williamwen42/330/base 2025-12-04T09:20:35.8257675Z * [new branch] gh/williamwen42/330/head -> origin/gh/williamwen42/330/head 2025-12-04T09:20:35.8259487Z * [new branch] gh/williamwen42/330/orig -> origin/gh/williamwen42/330/orig 2025-12-04T09:20:35.8261994Z * [new branch] gh/williamwen42/331/base -> origin/gh/williamwen42/331/base 2025-12-04T09:20:35.8263775Z * [new branch] gh/williamwen42/331/head -> origin/gh/williamwen42/331/head 2025-12-04T09:20:35.8265534Z * [new branch] gh/williamwen42/331/orig -> origin/gh/williamwen42/331/orig 2025-12-04T09:20:35.8267883Z * [new branch] gh/williamwen42/332/base -> origin/gh/williamwen42/332/base 2025-12-04T09:20:35.8269589Z * [new branch] gh/williamwen42/332/head -> origin/gh/williamwen42/332/head 2025-12-04T09:20:35.8271440Z * [new branch] gh/williamwen42/332/orig -> origin/gh/williamwen42/332/orig 2025-12-04T09:20:35.8275333Z * [new branch] gh/williamwen42/333/base -> origin/gh/williamwen42/333/base 2025-12-04T09:20:35.8277084Z * [new branch] gh/williamwen42/333/head -> origin/gh/williamwen42/333/head 2025-12-04T09:20:35.8278818Z * [new branch] gh/williamwen42/333/orig -> origin/gh/williamwen42/333/orig 2025-12-04T09:20:35.8281225Z * [new branch] gh/williamwen42/334/base -> origin/gh/williamwen42/334/base 2025-12-04T09:20:35.8283335Z * [new branch] gh/williamwen42/334/head -> origin/gh/williamwen42/334/head 2025-12-04T09:20:35.8285186Z * [new branch] gh/williamwen42/334/orig -> origin/gh/williamwen42/334/orig 2025-12-04T09:20:35.8290612Z * [new branch] gh/williamwen42/335/base -> origin/gh/williamwen42/335/base 2025-12-04T09:20:35.8292389Z * [new branch] gh/williamwen42/335/head -> origin/gh/williamwen42/335/head 2025-12-04T09:20:35.8294098Z * [new branch] gh/williamwen42/335/orig -> origin/gh/williamwen42/335/orig 2025-12-04T09:20:35.8297015Z * [new branch] gh/williamwen42/336/base -> origin/gh/williamwen42/336/base 2025-12-04T09:20:35.8298751Z * [new branch] gh/williamwen42/336/head -> origin/gh/williamwen42/336/head 2025-12-04T09:20:35.8300599Z * [new branch] gh/williamwen42/336/orig -> origin/gh/williamwen42/336/orig 2025-12-04T09:20:35.8303506Z * [new branch] gh/williamwen42/337/base -> origin/gh/williamwen42/337/base 2025-12-04T09:20:35.8305267Z * [new branch] gh/williamwen42/337/head -> origin/gh/williamwen42/337/head 2025-12-04T09:20:35.8307010Z * [new branch] gh/williamwen42/337/orig -> origin/gh/williamwen42/337/orig 2025-12-04T09:20:35.8309497Z * [new branch] gh/williamwen42/338/base -> origin/gh/williamwen42/338/base 2025-12-04T09:20:35.8311254Z * [new branch] gh/williamwen42/338/head -> origin/gh/williamwen42/338/head 2025-12-04T09:20:35.8312989Z * [new branch] gh/williamwen42/338/orig -> origin/gh/williamwen42/338/orig 2025-12-04T09:20:35.8315504Z * [new branch] gh/williamwen42/339/base -> origin/gh/williamwen42/339/base 2025-12-04T09:20:35.8317377Z * [new branch] gh/williamwen42/339/head -> origin/gh/williamwen42/339/head 2025-12-04T09:20:35.8318997Z * [new branch] gh/williamwen42/339/orig -> origin/gh/williamwen42/339/orig 2025-12-04T09:20:35.8321510Z * [new branch] gh/williamwen42/340/base -> origin/gh/williamwen42/340/base 2025-12-04T09:20:35.8323230Z * [new branch] gh/williamwen42/340/head -> origin/gh/williamwen42/340/head 2025-12-04T09:20:35.8324904Z * [new branch] gh/williamwen42/340/orig -> origin/gh/williamwen42/340/orig 2025-12-04T09:20:35.8327452Z * [new branch] gh/williamwen42/341/base -> origin/gh/williamwen42/341/base 2025-12-04T09:20:35.8329232Z * [new branch] gh/williamwen42/341/head -> origin/gh/williamwen42/341/head 2025-12-04T09:20:35.8331006Z * [new branch] gh/williamwen42/341/orig -> origin/gh/williamwen42/341/orig 2025-12-04T09:20:35.8333348Z * [new branch] gh/williamwen42/342/base -> origin/gh/williamwen42/342/base 2025-12-04T09:20:35.8335129Z * [new branch] gh/williamwen42/342/head -> origin/gh/williamwen42/342/head 2025-12-04T09:20:35.8336871Z * [new branch] gh/williamwen42/342/orig -> origin/gh/williamwen42/342/orig 2025-12-04T09:20:35.8339356Z * [new branch] gh/williamwen42/343/base -> origin/gh/williamwen42/343/base 2025-12-04T09:20:35.8341198Z * [new branch] gh/williamwen42/343/head -> origin/gh/williamwen42/343/head 2025-12-04T09:20:35.8342943Z * [new branch] gh/williamwen42/343/orig -> origin/gh/williamwen42/343/orig 2025-12-04T09:20:35.8345327Z * [new branch] gh/williamwen42/344/base -> origin/gh/williamwen42/344/base 2025-12-04T09:20:35.8347059Z * [new branch] gh/williamwen42/344/head -> origin/gh/williamwen42/344/head 2025-12-04T09:20:35.8348770Z * [new branch] gh/williamwen42/344/orig -> origin/gh/williamwen42/344/orig 2025-12-04T09:20:35.8351398Z * [new branch] gh/williamwen42/345/base -> origin/gh/williamwen42/345/base 2025-12-04T09:20:35.8353105Z * [new branch] gh/williamwen42/345/head -> origin/gh/williamwen42/345/head 2025-12-04T09:20:35.8354821Z * [new branch] gh/williamwen42/345/orig -> origin/gh/williamwen42/345/orig 2025-12-04T09:20:35.8357238Z * [new branch] gh/williamwen42/346/base -> origin/gh/williamwen42/346/base 2025-12-04T09:20:35.8359071Z * [new branch] gh/williamwen42/346/head -> origin/gh/williamwen42/346/head 2025-12-04T09:20:35.8360795Z * [new branch] gh/williamwen42/346/orig -> origin/gh/williamwen42/346/orig 2025-12-04T09:20:35.8363278Z * [new branch] gh/williamwen42/347/base -> origin/gh/williamwen42/347/base 2025-12-04T09:20:35.8364981Z * [new branch] gh/williamwen42/347/head -> origin/gh/williamwen42/347/head 2025-12-04T09:20:35.8366753Z * [new branch] gh/williamwen42/347/orig -> origin/gh/williamwen42/347/orig 2025-12-04T09:20:35.8369057Z * [new branch] gh/williamwen42/348/base -> origin/gh/williamwen42/348/base 2025-12-04T09:20:35.8370716Z * [new branch] gh/williamwen42/348/head -> origin/gh/williamwen42/348/head 2025-12-04T09:20:35.8372653Z * [new branch] gh/williamwen42/348/orig -> origin/gh/williamwen42/348/orig 2025-12-04T09:20:35.8375004Z * [new branch] gh/williamwen42/349/base -> origin/gh/williamwen42/349/base 2025-12-04T09:20:35.8376772Z * [new branch] gh/williamwen42/349/head -> origin/gh/williamwen42/349/head 2025-12-04T09:20:35.8378462Z * [new branch] gh/williamwen42/349/orig -> origin/gh/williamwen42/349/orig 2025-12-04T09:20:35.8381103Z * [new branch] gh/williamwen42/350/base -> origin/gh/williamwen42/350/base 2025-12-04T09:20:35.8382836Z * [new branch] gh/williamwen42/350/head -> origin/gh/williamwen42/350/head 2025-12-04T09:20:35.8384700Z * [new branch] gh/williamwen42/350/orig -> origin/gh/williamwen42/350/orig 2025-12-04T09:20:35.8387063Z * [new branch] gh/williamwen42/351/base -> origin/gh/williamwen42/351/base 2025-12-04T09:20:35.8388940Z * [new branch] gh/williamwen42/351/head -> origin/gh/williamwen42/351/head 2025-12-04T09:20:35.8390709Z * [new branch] gh/williamwen42/351/orig -> origin/gh/williamwen42/351/orig 2025-12-04T09:20:35.8393123Z * [new branch] gh/williamwen42/352/base -> origin/gh/williamwen42/352/base 2025-12-04T09:20:35.8394926Z * [new branch] gh/williamwen42/352/head -> origin/gh/williamwen42/352/head 2025-12-04T09:20:35.8397190Z * [new branch] gh/williamwen42/352/orig -> origin/gh/williamwen42/352/orig 2025-12-04T09:20:35.8399738Z * [new branch] gh/williamwen42/353/base -> origin/gh/williamwen42/353/base 2025-12-04T09:20:35.8401577Z * [new branch] gh/williamwen42/353/head -> origin/gh/williamwen42/353/head 2025-12-04T09:20:35.8403239Z * [new branch] gh/williamwen42/353/orig -> origin/gh/williamwen42/353/orig 2025-12-04T09:20:35.8405619Z * [new branch] gh/williamwen42/354/base -> origin/gh/williamwen42/354/base 2025-12-04T09:20:35.8407426Z * [new branch] gh/williamwen42/354/head -> origin/gh/williamwen42/354/head 2025-12-04T09:20:35.8409174Z * [new branch] gh/williamwen42/354/orig -> origin/gh/williamwen42/354/orig 2025-12-04T09:20:35.8411564Z * [new branch] gh/williamwen42/355/base -> origin/gh/williamwen42/355/base 2025-12-04T09:20:35.8413368Z * [new branch] gh/williamwen42/355/head -> origin/gh/williamwen42/355/head 2025-12-04T09:20:35.8415116Z * [new branch] gh/williamwen42/355/orig -> origin/gh/williamwen42/355/orig 2025-12-04T09:20:35.8417586Z * [new branch] gh/williamwen42/356/base -> origin/gh/williamwen42/356/base 2025-12-04T09:20:35.8419402Z * [new branch] gh/williamwen42/356/head -> origin/gh/williamwen42/356/head 2025-12-04T09:20:35.8421154Z * [new branch] gh/williamwen42/356/orig -> origin/gh/williamwen42/356/orig 2025-12-04T09:20:35.8424103Z * [new branch] gh/williamwen42/357/base -> origin/gh/williamwen42/357/base 2025-12-04T09:20:35.8425887Z * [new branch] gh/williamwen42/357/head -> origin/gh/williamwen42/357/head 2025-12-04T09:20:35.8427620Z * [new branch] gh/williamwen42/357/orig -> origin/gh/williamwen42/357/orig 2025-12-04T09:20:35.8430202Z * [new branch] gh/williamwen42/358/base -> origin/gh/williamwen42/358/base 2025-12-04T09:20:35.8431924Z * [new branch] gh/williamwen42/358/head -> origin/gh/williamwen42/358/head 2025-12-04T09:20:35.8433923Z * [new branch] gh/williamwen42/358/orig -> origin/gh/williamwen42/358/orig 2025-12-04T09:20:35.8436712Z * [new branch] gh/xmfan/169/base -> origin/gh/xmfan/169/base 2025-12-04T09:20:35.8438493Z * [new branch] gh/xmfan/169/head -> origin/gh/xmfan/169/head 2025-12-04T09:20:35.8440686Z * [new branch] gh/xmfan/170/base -> origin/gh/xmfan/170/base 2025-12-04T09:20:35.8442412Z * [new branch] gh/xmfan/170/head -> origin/gh/xmfan/170/head 2025-12-04T09:20:35.8444790Z * [new branch] gh/xmfan/274/base -> origin/gh/xmfan/274/base 2025-12-04T09:20:35.8446574Z * [new branch] gh/xmfan/274/head -> origin/gh/xmfan/274/head 2025-12-04T09:20:35.8448244Z * [new branch] gh/xmfan/274/orig -> origin/gh/xmfan/274/orig 2025-12-04T09:20:35.8450563Z * [new branch] gh/xmfan/277/base -> origin/gh/xmfan/277/base 2025-12-04T09:20:35.8452212Z * [new branch] gh/xmfan/277/head -> origin/gh/xmfan/277/head 2025-12-04T09:20:35.8454074Z * [new branch] gh/xmfan/277/orig -> origin/gh/xmfan/277/orig 2025-12-04T09:20:35.8456452Z * [new branch] gh/xmfan/301/base -> origin/gh/xmfan/301/base 2025-12-04T09:20:35.8458072Z * [new branch] gh/xmfan/301/head -> origin/gh/xmfan/301/head 2025-12-04T09:20:35.8459815Z * [new branch] gh/xmfan/301/orig -> origin/gh/xmfan/301/orig 2025-12-04T09:20:35.8462103Z * [new branch] gh/xmfan/304/base -> origin/gh/xmfan/304/base 2025-12-04T09:20:35.8463734Z * [new branch] gh/xmfan/304/head -> origin/gh/xmfan/304/head 2025-12-04T09:20:35.8465444Z * [new branch] gh/xmfan/304/orig -> origin/gh/xmfan/304/orig 2025-12-04T09:20:35.8467868Z * [new branch] gh/xmfan/309/base -> origin/gh/xmfan/309/base 2025-12-04T09:20:35.8469590Z * [new branch] gh/xmfan/309/head -> origin/gh/xmfan/309/head 2025-12-04T09:20:35.8471422Z * [new branch] gh/xmfan/309/orig -> origin/gh/xmfan/309/orig 2025-12-04T09:20:35.8473826Z * [new branch] gh/xmfan/310/base -> origin/gh/xmfan/310/base 2025-12-04T09:20:35.8475528Z * [new branch] gh/xmfan/310/head -> origin/gh/xmfan/310/head 2025-12-04T09:20:35.8477282Z * [new branch] gh/xmfan/310/orig -> origin/gh/xmfan/310/orig 2025-12-04T09:20:35.8479587Z * [new branch] gh/xmfan/311/base -> origin/gh/xmfan/311/base 2025-12-04T09:20:35.8481317Z * [new branch] gh/xmfan/311/head -> origin/gh/xmfan/311/head 2025-12-04T09:20:35.8483012Z * [new branch] gh/xmfan/311/orig -> origin/gh/xmfan/311/orig 2025-12-04T09:20:35.8485427Z * [new branch] gh/xmfan/312/base -> origin/gh/xmfan/312/base 2025-12-04T09:20:35.8487173Z * [new branch] gh/xmfan/312/head -> origin/gh/xmfan/312/head 2025-12-04T09:20:35.8488938Z * [new branch] gh/xmfan/312/orig -> origin/gh/xmfan/312/orig 2025-12-04T09:20:35.8491356Z * [new branch] gh/xmfan/313/base -> origin/gh/xmfan/313/base 2025-12-04T09:20:35.8493058Z * [new branch] gh/xmfan/313/head -> origin/gh/xmfan/313/head 2025-12-04T09:20:35.8495285Z * [new branch] gh/xmfan/313/orig -> origin/gh/xmfan/313/orig 2025-12-04T09:20:35.8498194Z * [new branch] gh/xuanzhang816/27/base -> origin/gh/xuanzhang816/27/base 2025-12-04T09:20:35.8500020Z * [new branch] gh/xuanzhang816/27/head -> origin/gh/xuanzhang816/27/head 2025-12-04T09:20:35.8501825Z * [new branch] gh/xuanzhang816/27/orig -> origin/gh/xuanzhang816/27/orig 2025-12-04T09:20:35.8504093Z * [new branch] gh/xuanzhang816/32/base -> origin/gh/xuanzhang816/32/base 2025-12-04T09:20:35.8505828Z * [new branch] gh/xuanzhang816/32/head -> origin/gh/xuanzhang816/32/head 2025-12-04T09:20:35.8507548Z * [new branch] gh/xuanzhang816/32/orig -> origin/gh/xuanzhang816/32/orig 2025-12-04T09:20:35.8510022Z * [new branch] gh/xuanzhang816/33/base -> origin/gh/xuanzhang816/33/base 2025-12-04T09:20:35.8511717Z * [new branch] gh/xuanzhang816/33/head -> origin/gh/xuanzhang816/33/head 2025-12-04T09:20:35.8513450Z * [new branch] gh/xuanzhang816/33/orig -> origin/gh/xuanzhang816/33/orig 2025-12-04T09:20:35.8516146Z * [new branch] gh/xuanzhang816/34/base -> origin/gh/xuanzhang816/34/base 2025-12-04T09:20:35.8517825Z * [new branch] gh/xuanzhang816/34/head -> origin/gh/xuanzhang816/34/head 2025-12-04T09:20:35.8519553Z * [new branch] gh/xuanzhang816/34/orig -> origin/gh/xuanzhang816/34/orig 2025-12-04T09:20:35.8522140Z * [new branch] gh/xuanzhang816/35/base -> origin/gh/xuanzhang816/35/base 2025-12-04T09:20:35.8523902Z * [new branch] gh/xuanzhang816/35/head -> origin/gh/xuanzhang816/35/head 2025-12-04T09:20:35.8525750Z * [new branch] gh/xuanzhang816/35/orig -> origin/gh/xuanzhang816/35/orig 2025-12-04T09:20:35.8528403Z * [new branch] gh/yanbing-j/11/base -> origin/gh/yanbing-j/11/base 2025-12-04T09:20:35.8530102Z * [new branch] gh/yanbing-j/11/head -> origin/gh/yanbing-j/11/head 2025-12-04T09:20:35.8531862Z * [new branch] gh/yanbing-j/11/orig -> origin/gh/yanbing-j/11/orig 2025-12-04T09:20:35.8534175Z * [new branch] gh/yanbing-j/12/base -> origin/gh/yanbing-j/12/base 2025-12-04T09:20:35.8535880Z * [new branch] gh/yanbing-j/12/head -> origin/gh/yanbing-j/12/head 2025-12-04T09:20:35.8537769Z * [new branch] gh/yanbing-j/12/orig -> origin/gh/yanbing-j/12/orig 2025-12-04T09:20:35.8540267Z * [new branch] gh/yanbing-j/13/base -> origin/gh/yanbing-j/13/base 2025-12-04T09:20:35.8541852Z * [new branch] gh/yanbing-j/13/head -> origin/gh/yanbing-j/13/head 2025-12-04T09:20:35.8544208Z * [new branch] gh/yanbing-j/13/orig -> origin/gh/yanbing-j/13/orig 2025-12-04T09:20:35.8546482Z * [new branch] gh/yanbing-j/14/base -> origin/gh/yanbing-j/14/base 2025-12-04T09:20:35.8548675Z * [new branch] gh/yanbing-j/14/head -> origin/gh/yanbing-j/14/head 2025-12-04T09:20:35.8550371Z * [new branch] gh/yanbing-j/14/orig -> origin/gh/yanbing-j/14/orig 2025-12-04T09:20:35.8552778Z * [new branch] gh/yanbing-j/15/base -> origin/gh/yanbing-j/15/base 2025-12-04T09:20:35.8554624Z * [new branch] gh/yanbing-j/15/head -> origin/gh/yanbing-j/15/head 2025-12-04T09:20:35.8556318Z * [new branch] gh/yanbing-j/15/orig -> origin/gh/yanbing-j/15/orig 2025-12-04T09:20:35.8558553Z * [new branch] gh/yanbing-j/18/base -> origin/gh/yanbing-j/18/base 2025-12-04T09:20:35.8560268Z * [new branch] gh/yanbing-j/18/head -> origin/gh/yanbing-j/18/head 2025-12-04T09:20:35.8562071Z * [new branch] gh/yanbing-j/18/orig -> origin/gh/yanbing-j/18/orig 2025-12-04T09:20:35.8564379Z * [new branch] gh/yanbing-j/19/base -> origin/gh/yanbing-j/19/base 2025-12-04T09:20:35.8566108Z * [new branch] gh/yanbing-j/19/head -> origin/gh/yanbing-j/19/head 2025-12-04T09:20:35.8567804Z * [new branch] gh/yanbing-j/19/orig -> origin/gh/yanbing-j/19/orig 2025-12-04T09:20:35.8570161Z * [new branch] gh/yanbing-j/20/base -> origin/gh/yanbing-j/20/base 2025-12-04T09:20:35.8572282Z * [new branch] gh/yanbing-j/20/head -> origin/gh/yanbing-j/20/head 2025-12-04T09:20:35.8574390Z * [new branch] gh/yanbing-j/20/orig -> origin/gh/yanbing-j/20/orig 2025-12-04T09:20:35.8576716Z * [new branch] gh/yanbing-j/21/base -> origin/gh/yanbing-j/21/base 2025-12-04T09:20:35.8578450Z * [new branch] gh/yanbing-j/21/head -> origin/gh/yanbing-j/21/head 2025-12-04T09:20:35.8580852Z * [new branch] gh/yanbing-j/22/base -> origin/gh/yanbing-j/22/base 2025-12-04T09:20:35.8583144Z * [new branch] gh/yanbing-j/22/head -> origin/gh/yanbing-j/22/head 2025-12-04T09:20:35.8585368Z * [new branch] gh/yanbing-j/22/orig -> origin/gh/yanbing-j/22/orig 2025-12-04T09:20:35.8587850Z * [new branch] gh/yanbing-j/23/base -> origin/gh/yanbing-j/23/base 2025-12-04T09:20:35.8589659Z * [new branch] gh/yanbing-j/23/head -> origin/gh/yanbing-j/23/head 2025-12-04T09:20:35.8591388Z * [new branch] gh/yanbing-j/23/orig -> origin/gh/yanbing-j/23/orig 2025-12-04T09:20:35.8593657Z * [new branch] gh/yanbing-j/24/base -> origin/gh/yanbing-j/24/base 2025-12-04T09:20:35.8595412Z * [new branch] gh/yanbing-j/24/head -> origin/gh/yanbing-j/24/head 2025-12-04T09:20:35.8597317Z * [new branch] gh/yanbing-j/24/orig -> origin/gh/yanbing-j/24/orig 2025-12-04T09:20:35.8599550Z * [new branch] gh/yanbing-j/25/base -> origin/gh/yanbing-j/25/base 2025-12-04T09:20:35.8601244Z * [new branch] gh/yanbing-j/25/head -> origin/gh/yanbing-j/25/head 2025-12-04T09:20:35.8602978Z * [new branch] gh/yanbing-j/25/orig -> origin/gh/yanbing-j/25/orig 2025-12-04T09:20:35.8605307Z * [new branch] gh/yanbing-j/26/base -> origin/gh/yanbing-j/26/base 2025-12-04T09:20:35.8607127Z * [new branch] gh/yanbing-j/26/head -> origin/gh/yanbing-j/26/head 2025-12-04T09:20:35.8608852Z * [new branch] gh/yanbing-j/26/orig -> origin/gh/yanbing-j/26/orig 2025-12-04T09:20:35.8611897Z * [new branch] gh/yang-yu-hang/1/base -> origin/gh/yang-yu-hang/1/base 2025-12-04T09:20:35.8613778Z * [new branch] gh/yang-yu-hang/1/head -> origin/gh/yang-yu-hang/1/head 2025-12-04T09:20:35.8615620Z * [new branch] gh/yang-yu-hang/1/orig -> origin/gh/yang-yu-hang/1/orig 2025-12-04T09:20:35.8617971Z * [new branch] gh/yang-yu-hang/2/base -> origin/gh/yang-yu-hang/2/base 2025-12-04T09:20:35.8620006Z * [new branch] gh/yang-yu-hang/2/head -> origin/gh/yang-yu-hang/2/head 2025-12-04T09:20:35.8622146Z * [new branch] gh/yang-yu-hang/2/orig -> origin/gh/yang-yu-hang/2/orig 2025-12-04T09:20:35.8624360Z * [new branch] gh/yang-yu-hang/3/base -> origin/gh/yang-yu-hang/3/base 2025-12-04T09:20:35.8626122Z * [new branch] gh/yang-yu-hang/3/head -> origin/gh/yang-yu-hang/3/head 2025-12-04T09:20:35.8627919Z * [new branch] gh/yang-yu-hang/3/orig -> origin/gh/yang-yu-hang/3/orig 2025-12-04T09:20:35.8630772Z * [new branch] gh/yangw-dev/12/base -> origin/gh/yangw-dev/12/base 2025-12-04T09:20:35.8632644Z * [new branch] gh/yangw-dev/12/head -> origin/gh/yangw-dev/12/head 2025-12-04T09:20:35.8634334Z * [new branch] gh/yangw-dev/12/orig -> origin/gh/yangw-dev/12/orig 2025-12-04T09:20:35.8636740Z * [new branch] gh/yangw-dev/13/base -> origin/gh/yangw-dev/13/base 2025-12-04T09:20:35.8638652Z * [new branch] gh/yangw-dev/13/head -> origin/gh/yangw-dev/13/head 2025-12-04T09:20:35.8640346Z * [new branch] gh/yangw-dev/13/orig -> origin/gh/yangw-dev/13/orig 2025-12-04T09:20:35.8642641Z * [new branch] gh/yangw-dev/14/base -> origin/gh/yangw-dev/14/base 2025-12-04T09:20:35.8644430Z * [new branch] gh/yangw-dev/14/head -> origin/gh/yangw-dev/14/head 2025-12-04T09:20:35.8646166Z * [new branch] gh/yangw-dev/14/orig -> origin/gh/yangw-dev/14/orig 2025-12-04T09:20:35.8648493Z * [new branch] gh/yangw-dev/15/base -> origin/gh/yangw-dev/15/base 2025-12-04T09:20:35.8650227Z * [new branch] gh/yangw-dev/15/head -> origin/gh/yangw-dev/15/head 2025-12-04T09:20:35.8651995Z * [new branch] gh/yangw-dev/15/orig -> origin/gh/yangw-dev/15/orig 2025-12-04T09:20:35.8654835Z * [new branch] gh/yangw-dev/19/base -> origin/gh/yangw-dev/19/base 2025-12-04T09:20:35.8656614Z * [new branch] gh/yangw-dev/19/head -> origin/gh/yangw-dev/19/head 2025-12-04T09:20:35.8658355Z * [new branch] gh/yangw-dev/19/orig -> origin/gh/yangw-dev/19/orig 2025-12-04T09:20:35.8660917Z * [new branch] gh/yangw-dev/26/base -> origin/gh/yangw-dev/26/base 2025-12-04T09:20:35.8662542Z * [new branch] gh/yangw-dev/26/head -> origin/gh/yangw-dev/26/head 2025-12-04T09:20:35.8664266Z * [new branch] gh/yangw-dev/26/orig -> origin/gh/yangw-dev/26/orig 2025-12-04T09:20:35.8666574Z * [new branch] gh/yangw-dev/27/base -> origin/gh/yangw-dev/27/base 2025-12-04T09:20:35.8668554Z * [new branch] gh/yangw-dev/27/head -> origin/gh/yangw-dev/27/head 2025-12-04T09:20:35.8670122Z * [new branch] gh/yangw-dev/27/orig -> origin/gh/yangw-dev/27/orig 2025-12-04T09:20:35.8674709Z * [new branch] gh/ydwu4/292/base -> origin/gh/ydwu4/292/base 2025-12-04T09:20:35.8676341Z * [new branch] gh/ydwu4/292/head -> origin/gh/ydwu4/292/head 2025-12-04T09:20:35.8678049Z * [new branch] gh/ydwu4/292/orig -> origin/gh/ydwu4/292/orig 2025-12-04T09:20:35.8680547Z * [new branch] gh/ydwu4/294/base -> origin/gh/ydwu4/294/base 2025-12-04T09:20:35.8682360Z * [new branch] gh/ydwu4/294/head -> origin/gh/ydwu4/294/head 2025-12-04T09:20:35.8684202Z * [new branch] gh/ydwu4/294/orig -> origin/gh/ydwu4/294/orig 2025-12-04T09:20:35.8686726Z * [new branch] gh/ydwu4/295/base -> origin/gh/ydwu4/295/base 2025-12-04T09:20:35.8688536Z * [new branch] gh/ydwu4/295/head -> origin/gh/ydwu4/295/head 2025-12-04T09:20:35.8690339Z * [new branch] gh/ydwu4/295/orig -> origin/gh/ydwu4/295/orig 2025-12-04T09:20:35.8692562Z * [new branch] gh/ydwu4/296/base -> origin/gh/ydwu4/296/base 2025-12-04T09:20:35.8694312Z * [new branch] gh/ydwu4/296/head -> origin/gh/ydwu4/296/head 2025-12-04T09:20:35.8696035Z * [new branch] gh/ydwu4/296/orig -> origin/gh/ydwu4/296/orig 2025-12-04T09:20:35.8698466Z * [new branch] gh/ydwu4/306/base -> origin/gh/ydwu4/306/base 2025-12-04T09:20:35.8700403Z * [new branch] gh/ydwu4/306/head -> origin/gh/ydwu4/306/head 2025-12-04T09:20:35.8702160Z * [new branch] gh/ydwu4/306/orig -> origin/gh/ydwu4/306/orig 2025-12-04T09:20:35.8704585Z * [new branch] gh/ydwu4/312/base -> origin/gh/ydwu4/312/base 2025-12-04T09:20:35.8706867Z * [new branch] gh/ydwu4/312/head -> origin/gh/ydwu4/312/head 2025-12-04T09:20:35.8708588Z * [new branch] gh/ydwu4/312/orig -> origin/gh/ydwu4/312/orig 2025-12-04T09:20:35.8710912Z * [new branch] gh/ydwu4/322/base -> origin/gh/ydwu4/322/base 2025-12-04T09:20:35.8712595Z * [new branch] gh/ydwu4/322/head -> origin/gh/ydwu4/322/head 2025-12-04T09:20:35.8714451Z * [new branch] gh/ydwu4/322/orig -> origin/gh/ydwu4/322/orig 2025-12-04T09:20:35.8716800Z * [new branch] gh/ydwu4/327/base -> origin/gh/ydwu4/327/base 2025-12-04T09:20:35.8718551Z * [new branch] gh/ydwu4/327/head -> origin/gh/ydwu4/327/head 2025-12-04T09:20:35.8720277Z * [new branch] gh/ydwu4/327/orig -> origin/gh/ydwu4/327/orig 2025-12-04T09:20:35.8722738Z * [new branch] gh/ydwu4/328/base -> origin/gh/ydwu4/328/base 2025-12-04T09:20:35.8724460Z * [new branch] gh/ydwu4/328/head -> origin/gh/ydwu4/328/head 2025-12-04T09:20:35.8726146Z * [new branch] gh/ydwu4/328/orig -> origin/gh/ydwu4/328/orig 2025-12-04T09:20:35.8728519Z * [new branch] gh/ydwu4/329/base -> origin/gh/ydwu4/329/base 2025-12-04T09:20:35.8730274Z * [new branch] gh/ydwu4/329/head -> origin/gh/ydwu4/329/head 2025-12-04T09:20:35.8731933Z * [new branch] gh/ydwu4/329/orig -> origin/gh/ydwu4/329/orig 2025-12-04T09:20:35.8734329Z * [new branch] gh/ydwu4/330/base -> origin/gh/ydwu4/330/base 2025-12-04T09:20:35.8735992Z * [new branch] gh/ydwu4/330/head -> origin/gh/ydwu4/330/head 2025-12-04T09:20:35.8737800Z * [new branch] gh/ydwu4/330/orig -> origin/gh/ydwu4/330/orig 2025-12-04T09:20:35.8740173Z * [new branch] gh/ydwu4/331/base -> origin/gh/ydwu4/331/base 2025-12-04T09:20:35.8742073Z * [new branch] gh/ydwu4/331/head -> origin/gh/ydwu4/331/head 2025-12-04T09:20:35.8743673Z * [new branch] gh/ydwu4/331/orig -> origin/gh/ydwu4/331/orig 2025-12-04T09:20:35.8745847Z * [new branch] gh/ydwu4/332/base -> origin/gh/ydwu4/332/base 2025-12-04T09:20:35.8747582Z * [new branch] gh/ydwu4/332/head -> origin/gh/ydwu4/332/head 2025-12-04T09:20:35.8749287Z * [new branch] gh/ydwu4/332/orig -> origin/gh/ydwu4/332/orig 2025-12-04T09:20:35.8751578Z * [new branch] gh/ydwu4/333/base -> origin/gh/ydwu4/333/base 2025-12-04T09:20:35.8753302Z * [new branch] gh/ydwu4/333/head -> origin/gh/ydwu4/333/head 2025-12-04T09:20:35.8755065Z * [new branch] gh/ydwu4/333/orig -> origin/gh/ydwu4/333/orig 2025-12-04T09:20:35.8757273Z * [new branch] gh/ydwu4/334/base -> origin/gh/ydwu4/334/base 2025-12-04T09:20:35.8758991Z * [new branch] gh/ydwu4/334/head -> origin/gh/ydwu4/334/head 2025-12-04T09:20:35.8760705Z * [new branch] gh/ydwu4/334/orig -> origin/gh/ydwu4/334/orig 2025-12-04T09:20:35.8762936Z * [new branch] gh/ydwu4/335/base -> origin/gh/ydwu4/335/base 2025-12-04T09:20:35.8764662Z * [new branch] gh/ydwu4/335/head -> origin/gh/ydwu4/335/head 2025-12-04T09:20:35.8766388Z * [new branch] gh/ydwu4/335/orig -> origin/gh/ydwu4/335/orig 2025-12-04T09:20:35.8769269Z * [new branch] gh/ydwu4/337/base -> origin/gh/ydwu4/337/base 2025-12-04T09:20:35.8771218Z * [new branch] gh/ydwu4/337/head -> origin/gh/ydwu4/337/head 2025-12-04T09:20:35.8773047Z * [new branch] gh/ydwu4/337/orig -> origin/gh/ydwu4/337/orig 2025-12-04T09:20:35.8775510Z * [new branch] gh/ydwu4/339/base -> origin/gh/ydwu4/339/base 2025-12-04T09:20:35.8777262Z * [new branch] gh/ydwu4/339/head -> origin/gh/ydwu4/339/head 2025-12-04T09:20:35.8779026Z * [new branch] gh/ydwu4/339/orig -> origin/gh/ydwu4/339/orig 2025-12-04T09:20:35.8782011Z * [new branch] gh/yf225/133/base -> origin/gh/yf225/133/base 2025-12-04T09:20:35.8783716Z * [new branch] gh/yf225/133/head -> origin/gh/yf225/133/head 2025-12-04T09:20:35.8786054Z * [new branch] gh/yf225/93/base -> origin/gh/yf225/93/base 2025-12-04T09:20:35.8787937Z * [new branch] gh/yf225/93/head -> origin/gh/yf225/93/head 2025-12-04T09:20:35.8791207Z * [new branch] gh/yifuwang/152/base -> origin/gh/yifuwang/152/base 2025-12-04T09:20:35.8793288Z * [new branch] gh/yifuwang/152/head -> origin/gh/yifuwang/152/head 2025-12-04T09:20:35.8795143Z * [new branch] gh/yifuwang/152/orig -> origin/gh/yifuwang/152/orig 2025-12-04T09:20:35.8797607Z * [new branch] gh/yifuwang/195/base -> origin/gh/yifuwang/195/base 2025-12-04T09:20:35.8799371Z * [new branch] gh/yifuwang/195/head -> origin/gh/yifuwang/195/head 2025-12-04T09:20:35.8801096Z * [new branch] gh/yifuwang/195/orig -> origin/gh/yifuwang/195/orig 2025-12-04T09:20:35.8803986Z * [new branch] gh/yiming0416/1/base -> origin/gh/yiming0416/1/base 2025-12-04T09:20:35.8805788Z * [new branch] gh/yiming0416/1/head -> origin/gh/yiming0416/1/head 2025-12-04T09:20:35.8807990Z * [new branch] gh/yiming0416/2/base -> origin/gh/yiming0416/2/base 2025-12-04T09:20:35.8809639Z * [new branch] gh/yiming0416/2/head -> origin/gh/yiming0416/2/head 2025-12-04T09:20:35.8812434Z * [new branch] gh/yushangdi/1/base -> origin/gh/yushangdi/1/base 2025-12-04T09:20:35.8814284Z * [new branch] gh/yushangdi/1/head -> origin/gh/yushangdi/1/head 2025-12-04T09:20:35.8816574Z * [new branch] gh/yushangdi/10/base -> origin/gh/yushangdi/10/base 2025-12-04T09:20:35.8818357Z * [new branch] gh/yushangdi/10/head -> origin/gh/yushangdi/10/head 2025-12-04T09:20:35.8820256Z * [new branch] gh/yushangdi/10/orig -> origin/gh/yushangdi/10/orig 2025-12-04T09:20:35.8823187Z * [new branch] gh/yushangdi/11/base -> origin/gh/yushangdi/11/base 2025-12-04T09:20:35.8824936Z * [new branch] gh/yushangdi/11/head -> origin/gh/yushangdi/11/head 2025-12-04T09:20:35.8826696Z * [new branch] gh/yushangdi/11/orig -> origin/gh/yushangdi/11/orig 2025-12-04T09:20:35.8828973Z * [new branch] gh/yushangdi/2/base -> origin/gh/yushangdi/2/base 2025-12-04T09:20:35.8830676Z * [new branch] gh/yushangdi/2/head -> origin/gh/yushangdi/2/head 2025-12-04T09:20:35.8832945Z * [new branch] gh/yushangdi/7/base -> origin/gh/yushangdi/7/base 2025-12-04T09:20:35.8834652Z * [new branch] gh/yushangdi/7/head -> origin/gh/yushangdi/7/head 2025-12-04T09:20:35.8836349Z * [new branch] gh/yushangdi/7/orig -> origin/gh/yushangdi/7/orig 2025-12-04T09:20:35.8838968Z * [new branch] gh/yushangdi/8/base -> origin/gh/yushangdi/8/base 2025-12-04T09:20:35.8840839Z * [new branch] gh/yushangdi/8/head -> origin/gh/yushangdi/8/head 2025-12-04T09:20:35.8842731Z * [new branch] gh/yushangdi/8/orig -> origin/gh/yushangdi/8/orig 2025-12-04T09:20:35.8844943Z * [new branch] gh/yushangdi/9/base -> origin/gh/yushangdi/9/base 2025-12-04T09:20:35.8846851Z * [new branch] gh/yushangdi/9/head -> origin/gh/yushangdi/9/head 2025-12-04T09:20:35.8848659Z * [new branch] gh/yushangdi/9/orig -> origin/gh/yushangdi/9/orig 2025-12-04T09:20:35.8852044Z * [new branch] gh/zklaus/19/base -> origin/gh/zklaus/19/base 2025-12-04T09:20:35.8853791Z * [new branch] gh/zklaus/19/head -> origin/gh/zklaus/19/head 2025-12-04T09:20:35.8855520Z * [new branch] gh/zklaus/19/orig -> origin/gh/zklaus/19/orig 2025-12-04T09:20:35.8857862Z * [new branch] gh/zklaus/20/base -> origin/gh/zklaus/20/base 2025-12-04T09:20:35.8859691Z * [new branch] gh/zklaus/20/head -> origin/gh/zklaus/20/head 2025-12-04T09:20:35.8861444Z * [new branch] gh/zklaus/20/orig -> origin/gh/zklaus/20/orig 2025-12-04T09:20:35.8863756Z * [new branch] gh/zklaus/21/base -> origin/gh/zklaus/21/base 2025-12-04T09:20:35.8865506Z * [new branch] gh/zklaus/21/head -> origin/gh/zklaus/21/head 2025-12-04T09:20:35.8867223Z * [new branch] gh/zklaus/21/orig -> origin/gh/zklaus/21/orig 2025-12-04T09:20:35.8869499Z * [new branch] gh/zklaus/22/base -> origin/gh/zklaus/22/base 2025-12-04T09:20:35.8871394Z * [new branch] gh/zklaus/22/head -> origin/gh/zklaus/22/head 2025-12-04T09:20:35.8873228Z * [new branch] gh/zklaus/22/orig -> origin/gh/zklaus/22/orig 2025-12-04T09:20:35.8875563Z * [new branch] gh/zklaus/23/base -> origin/gh/zklaus/23/base 2025-12-04T09:20:35.8877282Z * [new branch] gh/zklaus/23/head -> origin/gh/zklaus/23/head 2025-12-04T09:20:35.8879030Z * [new branch] gh/zklaus/23/orig -> origin/gh/zklaus/23/orig 2025-12-04T09:20:35.8881310Z * [new branch] gh/zklaus/24/base -> origin/gh/zklaus/24/base 2025-12-04T09:20:35.8883056Z * [new branch] gh/zklaus/24/head -> origin/gh/zklaus/24/head 2025-12-04T09:20:35.8884843Z * [new branch] gh/zklaus/24/orig -> origin/gh/zklaus/24/orig 2025-12-04T09:20:35.8887935Z * [new branch] gh/zou3519/1197/base -> origin/gh/zou3519/1197/base 2025-12-04T09:20:35.8889522Z * [new branch] gh/zou3519/1197/head -> origin/gh/zou3519/1197/head 2025-12-04T09:20:35.8891242Z * [new branch] gh/zou3519/1197/orig -> origin/gh/zou3519/1197/orig 2025-12-04T09:20:35.8893977Z * [new branch] gh/zou3519/1199/base -> origin/gh/zou3519/1199/base 2025-12-04T09:20:35.8895703Z * [new branch] gh/zou3519/1199/head -> origin/gh/zou3519/1199/head 2025-12-04T09:20:35.8897418Z * [new branch] gh/zou3519/1199/orig -> origin/gh/zou3519/1199/orig 2025-12-04T09:20:35.8900018Z * [new branch] gh/zou3519/1200/base -> origin/gh/zou3519/1200/base 2025-12-04T09:20:35.8901804Z * [new branch] gh/zou3519/1200/head -> origin/gh/zou3519/1200/head 2025-12-04T09:20:35.8903465Z * [new branch] gh/zou3519/1200/orig -> origin/gh/zou3519/1200/orig 2025-12-04T09:20:35.8905877Z * [new branch] gh/zou3519/1201/base -> origin/gh/zou3519/1201/base 2025-12-04T09:20:35.8907573Z * [new branch] gh/zou3519/1201/head -> origin/gh/zou3519/1201/head 2025-12-04T09:20:35.8909274Z * [new branch] gh/zou3519/1201/orig -> origin/gh/zou3519/1201/orig 2025-12-04T09:20:35.8911887Z * [new branch] gh/zou3519/1202/base -> origin/gh/zou3519/1202/base 2025-12-04T09:20:35.8913677Z * [new branch] gh/zou3519/1202/head -> origin/gh/zou3519/1202/head 2025-12-04T09:20:35.8915478Z * [new branch] gh/zou3519/1202/orig -> origin/gh/zou3519/1202/orig 2025-12-04T09:20:35.8918330Z * [new branch] gh/zpcore/1/base -> origin/gh/zpcore/1/base 2025-12-04T09:20:35.8920042Z * [new branch] gh/zpcore/1/head -> origin/gh/zpcore/1/head 2025-12-04T09:20:35.8922425Z * [new branch] gh/zpcore/11/base -> origin/gh/zpcore/11/base 2025-12-04T09:20:35.8924301Z * [new branch] gh/zpcore/11/head -> origin/gh/zpcore/11/head 2025-12-04T09:20:35.8926050Z * [new branch] gh/zpcore/11/orig -> origin/gh/zpcore/11/orig 2025-12-04T09:20:35.8928696Z * [new branch] gh/zpcore/12/base -> origin/gh/zpcore/12/base 2025-12-04T09:20:35.8930414Z * [new branch] gh/zpcore/12/head -> origin/gh/zpcore/12/head 2025-12-04T09:20:35.8932181Z * [new branch] gh/zpcore/12/orig -> origin/gh/zpcore/12/orig 2025-12-04T09:20:35.8935169Z * [new branch] gh/zpcore/13/base -> origin/gh/zpcore/13/base 2025-12-04T09:20:35.8936882Z * [new branch] gh/zpcore/13/head -> origin/gh/zpcore/13/head 2025-12-04T09:20:35.8938594Z * [new branch] gh/zpcore/13/orig -> origin/gh/zpcore/13/orig 2025-12-04T09:20:35.8941092Z * [new branch] gh/zpcore/14/base -> origin/gh/zpcore/14/base 2025-12-04T09:20:35.8942777Z * [new branch] gh/zpcore/14/head -> origin/gh/zpcore/14/head 2025-12-04T09:20:35.8944518Z * [new branch] gh/zpcore/14/orig -> origin/gh/zpcore/14/orig 2025-12-04T09:20:35.8947036Z * [new branch] gh/zpcore/15/base -> origin/gh/zpcore/15/base 2025-12-04T09:20:35.8948768Z * [new branch] gh/zpcore/15/head -> origin/gh/zpcore/15/head 2025-12-04T09:20:35.8950996Z * [new branch] gh/zpcore/15/orig -> origin/gh/zpcore/15/orig 2025-12-04T09:20:35.8953420Z * [new branch] gh/zpcore/2/base -> origin/gh/zpcore/2/base 2025-12-04T09:20:35.8955261Z * [new branch] gh/zpcore/2/head -> origin/gh/zpcore/2/head 2025-12-04T09:20:35.8957988Z * [new branch] gh/zpcore/21/base -> origin/gh/zpcore/21/base 2025-12-04T09:20:35.8960155Z * [new branch] gh/zpcore/21/head -> origin/gh/zpcore/21/head 2025-12-04T09:20:35.8961768Z * [new branch] gh/zpcore/21/orig -> origin/gh/zpcore/21/orig 2025-12-04T09:20:35.8964544Z * [new branch] gh/zpcore/22/base -> origin/gh/zpcore/22/base 2025-12-04T09:20:35.8966356Z * [new branch] gh/zpcore/22/head -> origin/gh/zpcore/22/head 2025-12-04T09:20:35.8968627Z * [new branch] gh/zpcore/22/orig -> origin/gh/zpcore/22/orig 2025-12-04T09:20:35.8971179Z * [new branch] gh/zpcore/23/base -> origin/gh/zpcore/23/base 2025-12-04T09:20:35.8973112Z * [new branch] gh/zpcore/23/head -> origin/gh/zpcore/23/head 2025-12-04T09:20:35.8974862Z * [new branch] gh/zpcore/23/orig -> origin/gh/zpcore/23/orig 2025-12-04T09:20:35.8977102Z * [new branch] gh/zpcore/24/base -> origin/gh/zpcore/24/base 2025-12-04T09:20:35.8978832Z * [new branch] gh/zpcore/24/head -> origin/gh/zpcore/24/head 2025-12-04T09:20:35.8980731Z * [new branch] gh/zpcore/24/orig -> origin/gh/zpcore/24/orig 2025-12-04T09:20:35.8983216Z * [new branch] gh/zpcore/25/base -> origin/gh/zpcore/25/base 2025-12-04T09:20:35.8985024Z * [new branch] gh/zpcore/25/head -> origin/gh/zpcore/25/head 2025-12-04T09:20:35.8986806Z * [new branch] gh/zpcore/25/orig -> origin/gh/zpcore/25/orig 2025-12-04T09:20:35.8989216Z * [new branch] gh/zpcore/26/base -> origin/gh/zpcore/26/base 2025-12-04T09:20:35.8991089Z * [new branch] gh/zpcore/26/head -> origin/gh/zpcore/26/head 2025-12-04T09:20:35.8992822Z * [new branch] gh/zpcore/26/orig -> origin/gh/zpcore/26/orig 2025-12-04T09:20:35.8995229Z * [new branch] gh/zpcore/27/base -> origin/gh/zpcore/27/base 2025-12-04T09:20:35.8996954Z * [new branch] gh/zpcore/27/head -> origin/gh/zpcore/27/head 2025-12-04T09:20:35.8998691Z * [new branch] gh/zpcore/27/orig -> origin/gh/zpcore/27/orig 2025-12-04T09:20:35.9001486Z * [new branch] gh/zpcore/28/base -> origin/gh/zpcore/28/base 2025-12-04T09:20:35.9003620Z * [new branch] gh/zpcore/28/head -> origin/gh/zpcore/28/head 2025-12-04T09:20:35.9005468Z * [new branch] gh/zpcore/28/orig -> origin/gh/zpcore/28/orig 2025-12-04T09:20:35.9008177Z * [new branch] gh/zpcore/3/base -> origin/gh/zpcore/3/base 2025-12-04T09:20:35.9009902Z * [new branch] gh/zpcore/3/head -> origin/gh/zpcore/3/head 2025-12-04T09:20:35.9012055Z * [new branch] gh/zpcore/4/base -> origin/gh/zpcore/4/base 2025-12-04T09:20:35.9013738Z * [new branch] gh/zpcore/4/head -> origin/gh/zpcore/4/head 2025-12-04T09:20:35.9015933Z * [new branch] gh/zpcore/5/base -> origin/gh/zpcore/5/base 2025-12-04T09:20:35.9017680Z * [new branch] gh/zpcore/5/head -> origin/gh/zpcore/5/head 2025-12-04T09:20:35.9020008Z * [new branch] gh/zpcore/6/base -> origin/gh/zpcore/6/base 2025-12-04T09:20:35.9021751Z * [new branch] gh/zpcore/6/head -> origin/gh/zpcore/6/head 2025-12-04T09:20:35.9024868Z * [new branch] gh/zpcore/7/base -> origin/gh/zpcore/7/base 2025-12-04T09:20:35.9026558Z * [new branch] gh/zpcore/7/head -> origin/gh/zpcore/7/head 2025-12-04T09:20:35.9028817Z * [new branch] gh/zpcore/8/base -> origin/gh/zpcore/8/base 2025-12-04T09:20:35.9030553Z * [new branch] gh/zpcore/8/head -> origin/gh/zpcore/8/head 2025-12-04T09:20:35.9032455Z * [new branch] google-main -> origin/google-main 2025-12-04T09:20:35.9034994Z * [new branch] guangyey/external_stream -> origin/guangyey/external_stream 2025-12-04T09:20:35.9036908Z * [new branch] guangyey/test_2025 -> origin/guangyey/test_2025 2025-12-04T09:20:35.9039426Z * [new branch] guilhermeleobas/cherry-pick-55d87d9dfd9 -> origin/guilhermeleobas/cherry-pick-55d87d9dfd9 2025-12-04T09:20:35.9041765Z * [new branch] hameerabbasi/complex_tensor_subclass -> origin/hameerabbasi/complex_tensor_subclass 2025-12-04T09:20:35.9043500Z * [new branch] hameerabbasi/fix-ctensor-gradcheck-tests -> origin/hameerabbasi/fix-ctensor-gradcheck-tests 2025-12-04T09:20:35.9045097Z * [new branch] hameerabbasi/gradcheck-allclose -> origin/hameerabbasi/gradcheck-allclose 2025-12-04T09:20:35.9046822Z * [new branch] hc_baseline -> origin/hc_baseline 2025-12-04T09:20:35.9048691Z * [new branch] hhh_rand -> origin/hhh_rand 2025-12-04T09:20:35.9051021Z * [new branch] huba/f1 -> origin/huba/f1 2025-12-04T09:20:35.9053305Z * [new branch] increase-timeout-linux-jammy-cuda12_8-py3_10-gcc11-test -> origin/increase-timeout-linux-jammy-cuda12_8-py3_10-gcc11-test 2025-12-04T09:20:35.9054891Z * [new branch] inlining -> origin/inlining 2025-12-04T09:20:35.9056897Z * [new branch] inlining-ezyang -> origin/inlining-ezyang 2025-12-04T09:20:35.9058791Z * [new branch] install-torchao-0.13.0 -> origin/install-torchao-0.13.0 2025-12-04T09:20:35.9060951Z * [new branch] instrument-trunk-pull-linux-with-job-test-filters -> origin/instrument-trunk-pull-linux-with-job-test-filters 2025-12-04T09:20:35.9062561Z * [new branch] invoke-subgraph -> origin/invoke-subgraph 2025-12-04T09:20:35.9064501Z * [new branch] issue#58739 -> origin/issue#58739 2025-12-04T09:20:35.9066430Z * [new branch] jainapurva-patch-1 -> origin/jainapurva-patch-1 2025-12-04T09:20:35.9068705Z * [new branch] jathu/o3 -> origin/jathu/o3 2025-12-04T09:20:35.9070410Z * [new branch] jathu/sve -> origin/jathu/sve 2025-12-04T09:20:35.9074310Z * [new branch] jcaip/test-cusparselt-version-0.6.2 -> origin/jcaip/test-cusparselt-version-0.6.2 2025-12-04T09:20:35.9075947Z * [new branch] jcaip/update-cusparselt-0.6.2 -> origin/jcaip/update-cusparselt-0.6.2 2025-12-04T09:20:35.9078363Z * [new branch] jiannanWang/memorysnapshot_filter -> origin/jiannanWang/memorysnapshot_filter 2025-12-04T09:20:35.9080122Z * [new branch] jiannanWang/profilerstepwarning -> origin/jiannanWang/profilerstepwarning 2025-12-04T09:20:35.9081938Z * [new branch] jithunnair-amd-patch-1 -> origin/jithunnair-amd-patch-1 2025-12-04T09:20:35.9083782Z * [new branch] jithunnair-amd-patch-10 -> origin/jithunnair-amd-patch-10 2025-12-04T09:20:35.9085600Z * [new branch] jithunnair-amd-patch-2 -> origin/jithunnair-amd-patch-2 2025-12-04T09:20:35.9087536Z * [new branch] jithunnair-amd-patch-3 -> origin/jithunnair-amd-patch-3 2025-12-04T09:20:35.9089314Z * [new branch] jithunnair-amd-patch-4 -> origin/jithunnair-amd-patch-4 2025-12-04T09:20:35.9091128Z * [new branch] jithunnair-amd-patch-5 -> origin/jithunnair-amd-patch-5 2025-12-04T09:20:35.9092969Z * [new branch] jithunnair-amd-patch-6 -> origin/jithunnair-amd-patch-6 2025-12-04T09:20:35.9094830Z * [new branch] jithunnair-amd-patch-7 -> origin/jithunnair-amd-patch-7 2025-12-04T09:20:35.9096645Z * [new branch] jithunnair-amd-patch-8 -> origin/jithunnair-amd-patch-8 2025-12-04T09:20:35.9098477Z * [new branch] jithunnair-amd-patch-9 -> origin/jithunnair-amd-patch-9 2025-12-04T09:20:35.9101085Z * [new branch] justinchu/native-qdq -> origin/justinchu/native-qdq 2025-12-04T09:20:35.9103570Z * [new branch] kainan666/xlf_debug -> origin/kainan666/xlf_debug 2025-12-04T09:20:35.9105234Z * [new branch] kainan_test -> origin/kainan_test 2025-12-04T09:20:35.9107135Z * [new branch] larryliu0820-patch-1 -> origin/larryliu0820-patch-1 2025-12-04T09:20:35.9109679Z * [new branch] leslie/test_group_gemm_epilogues -> origin/leslie/test_group_gemm_epilogues 2025-12-04T09:20:35.9111946Z * [new branch] lessw2020/fix_cutlass_cache_error -> origin/lessw2020/fix_cutlass_cache_error 2025-12-04T09:20:35.9114248Z * [new branch] liaoxuan/shm_all_reduce -> origin/liaoxuan/shm_all_reduce 2025-12-04T09:20:35.9115976Z * [new branch] liaoxuan/test_fa_disable_softmax -> origin/liaoxuan/test_fa_disable_softmax 2025-12-04T09:20:35.9117544Z * [new branch] liaoxuan/test_int8_sdpa -> origin/liaoxuan/test_int8_sdpa 2025-12-04T09:20:35.9119322Z * [new branch] llama4-stable -> origin/llama4-stable 2025-12-04T09:20:35.9122697Z * [new branch] lts/release/1.8 -> origin/lts/release/1.8 2025-12-04T09:20:35.9125211Z * [new branch] lucaskabela/#94773 -> origin/lucaskabela/#94773 2025-12-04T09:20:35.9126871Z * [new branch] lucaskabela/fix_164876 -> origin/lucaskabela/fix_164876 2025-12-04T09:20:35.9128595Z * [new branch] lucaskabela/flop_counter -> origin/lucaskabela/flop_counter 2025-12-04T09:20:35.9130279Z * [new branch] lucaskabela/func_under_decomp -> origin/lucaskabela/func_under_decomp 2025-12-04T09:20:35.9132147Z * [new branch] lucaskabela/functional_in_dynamo -> origin/lucaskabela/functional_in_dynamo 2025-12-04T09:20:35.9133872Z * [new branch] lucaskabela/install_params_as_graph_attr -> origin/lucaskabela/install_params_as_graph_attr 2025-12-04T09:20:35.9135890Z * [new branch] lucaskabela/parameters_as_graph_attr -> origin/lucaskabela/parameters_as_graph_attr 2025-12-04T09:20:35.9138152Z * [new branch] lucaskabela/remove_aot_dispatcher_metadata -> origin/lucaskabela/remove_aot_dispatcher_metadata 2025-12-04T09:20:35.9139974Z * [new branch] lucaskabela/rnn_decomp -> origin/lucaskabela/rnn_decomp 2025-12-04T09:20:35.9141719Z * [new branch] lucaskabela/typing_backends -> origin/lucaskabela/typing_backends 2025-12-04T09:20:35.9143387Z * [new branch] lucaskabela/typing_ctx_manager -> origin/lucaskabela/typing_ctx_manager 2025-12-04T09:20:35.9145273Z * [new branch] lucaskabela/typing_nn_module -> origin/lucaskabela/typing_nn_module 2025-12-04T09:20:35.9147478Z * [new branch] lucaskabela/typing_user_defined -> origin/lucaskabela/typing_user_defined 2025-12-04T09:20:35.9149222Z * [new branch] lucaskabela/typing_variables -> origin/lucaskabela/typing_variables 2025-12-04T09:20:35.9151007Z * [new branch] lucaskabela/typing_variables_dicts -> origin/lucaskabela/typing_variables_dicts 2025-12-04T09:20:35.9152750Z * [new branch] lucaskabela/typing_variables_functions -> origin/lucaskabela/typing_variables_functions 2025-12-04T09:20:35.9154448Z * [new branch] lucaskabela/typing_variables_lists -> origin/lucaskabela/typing_variables_lists 2025-12-04T09:20:35.9156803Z * [new branch] lw/torch_box_by_ref -> origin/lw/torch_box_by_ref 2025-12-04T09:20:35.9158595Z * [new branch] main -> origin/main 2025-12-04T09:20:35.9160605Z * [new branch] malfet-patch-1 -> origin/malfet-patch-1 2025-12-04T09:20:35.9162528Z * [new branch] malfet-patch-2 -> origin/malfet-patch-2 2025-12-04T09:20:35.9164532Z * [new branch] malfet-patch-3 -> origin/malfet-patch-3 2025-12-04T09:20:35.9166477Z * [new branch] malfet-patch-4 -> origin/malfet-patch-4 2025-12-04T09:20:35.9168234Z * [new branch] malfet-patch-5 -> origin/malfet-patch-5 2025-12-04T09:20:35.9170007Z * [new branch] malfet-patch-6 -> origin/malfet-patch-6 2025-12-04T09:20:35.9172164Z * [new branch] malfet-patch-7 -> origin/malfet-patch-7 2025-12-04T09:20:35.9173925Z * [new branch] malfet-patch-8 -> origin/malfet-patch-8 2025-12-04T09:20:35.9176316Z * [new branch] malfet/add-3.14-ci -> origin/malfet/add-3.14-ci 2025-12-04T09:20:35.9178242Z * [new branch] malfet/be-do-not-make-typos-in-build-artifacts -> origin/malfet/be-do-not-make-typos-in-build-artifacts 2025-12-04T09:20:35.9180165Z * [new branch] malfet/be-move-more-settings-to-checkout-pytorch -> origin/malfet/be-move-more-settings-to-checkout-pytorch 2025-12-04T09:20:35.9181954Z * [new branch] malfet/be-remove-misisng-neon-headers -> origin/malfet/be-remove-misisng-neon-headers 2025-12-04T09:20:35.9184298Z * [new branch] malfet/mps-implement-col2im -> origin/malfet/mps-implement-col2im 2025-12-04T09:20:35.9186894Z * [new branch] manuel/aoti_metal_shimify-thread_safe -> origin/manuel/aoti_metal_shimify-thread_safe 2025-12-04T09:20:35.9188574Z * [new branch] manuel/inductor_link_openmp -> origin/manuel/inductor_link_openmp 2025-12-04T09:20:35.9190825Z * [new branch] masnesral/metaconda -> origin/masnesral/metaconda 2025-12-04T09:20:35.9192752Z * [new branch] mem_profiler_flaky_fix -> origin/mem_profiler_flaky_fix 2025-12-04T09:20:35.9194594Z * [new branch] mem_profiler_stack_trace -> origin/mem_profiler_stack_trace 2025-12-04T09:20:35.9196432Z * [new branch] memory_profiler_stack -> origin/memory_profiler_stack 2025-12-04T09:20:35.9198266Z * [new branch] metascroy-patch-1 -> origin/metascroy-patch-1 2025-12-04T09:20:35.9200120Z * [new branch] mingw_posix -> origin/mingw_posix 2025-12-04T09:20:35.9202576Z * [new branch] mlazos/S429861-debug -> origin/mlazos/S429861-debug 2025-12-04T09:20:35.9204254Z * [new branch] mlazos/aa -> origin/mlazos/aa 2025-12-04T09:20:35.9205916Z * [new branch] mlazos/acts -> origin/mlazos/acts 2025-12-04T09:20:35.9207592Z * [new branch] mlazos/arg-renames -> origin/mlazos/arg-renames 2025-12-04T09:20:35.9209257Z * [new branch] mlazos/bad-cudagraphs -> origin/mlazos/bad-cudagraphs 2025-12-04T09:20:35.9210960Z * [new branch] mlazos/baseline-graph-breaks -> origin/mlazos/baseline-graph-breaks 2025-12-04T09:20:35.9212553Z * [new branch] mlazos/beta-tensor -> origin/mlazos/beta-tensor 2025-12-04T09:20:35.9214303Z * [new branch] mlazos/buffers -> origin/mlazos/buffers 2025-12-04T09:20:35.9215847Z * [new branch] mlazos/buffers2 -> origin/mlazos/buffers2 2025-12-04T09:20:35.9217849Z * [new branch] mlazos/buffers3 -> origin/mlazos/buffers3 2025-12-04T09:20:35.9219968Z * [new branch] mlazos/bwd -> origin/mlazos/bwd 2025-12-04T09:20:35.9221625Z * [new branch] mlazos/combo-test -> origin/mlazos/combo-test 2025-12-04T09:20:35.9223349Z * [new branch] mlazos/ctx-cleanup -> origin/mlazos/ctx-cleanup 2025-12-04T09:20:35.9225241Z * [new branch] mlazos/cuda-cmd-log -> origin/mlazos/cuda-cmd-log 2025-12-04T09:20:35.9227101Z * [new branch] mlazos/cudagraph-tests -> origin/mlazos/cudagraph-tests 2025-12-04T09:20:35.9228886Z * [new branch] mlazos/cudagraphs-measurement -> origin/mlazos/cudagraphs-measurement 2025-12-04T09:20:35.9230721Z * [new branch] mlazos/cutlass-test -> origin/mlazos/cutlass-test 2025-12-04T09:20:35.9232610Z * [new branch] mlazos/cutlass-topo-bug -> origin/mlazos/cutlass-topo-bug 2025-12-04T09:20:35.9234352Z * [new branch] mlazos/dataclass-proxy -> origin/mlazos/dataclass-proxy 2025-12-04T09:20:35.9236111Z * [new branch] mlazos/dc-attrs -> origin/mlazos/dc-attrs 2025-12-04T09:20:35.9237921Z * [new branch] mlazos/dc-helion -> origin/mlazos/dc-helion 2025-12-04T09:20:35.9239637Z * [new branch] mlazos/dict-fix -> origin/mlazos/dict-fix 2025-12-04T09:20:35.9241378Z * [new branch] mlazos/disable-tf -> origin/mlazos/disable-tf 2025-12-04T09:20:35.9243167Z * [new branch] mlazos/dupe-fix -> origin/mlazos/dupe-fix 2025-12-04T09:20:35.9245041Z * [new branch] mlazos/dyn-batch -> origin/mlazos/dyn-batch 2025-12-04T09:20:35.9246746Z * [new branch] mlazos/evt -> origin/mlazos/evt 2025-12-04T09:20:35.9248491Z * [new branch] mlazos/extract-examples -> origin/mlazos/extract-examples 2025-12-04T09:20:35.9250251Z * [new branch] mlazos/foreach-op -> origin/mlazos/foreach-op 2025-12-04T09:20:35.9252006Z * [new branch] mlazos/fp8 -> origin/mlazos/fp8 2025-12-04T09:20:35.9253824Z * [new branch] mlazos/fp8-bias -> origin/mlazos/fp8-bias 2025-12-04T09:20:35.9255607Z * [new branch] mlazos/fp8-bias-fusion -> origin/mlazos/fp8-bias-fusion 2025-12-04T09:20:35.9257445Z * [new branch] mlazos/fp8-fixes -> origin/mlazos/fp8-fixes 2025-12-04T09:20:35.9259283Z * [new branch] mlazos/freezing -> origin/mlazos/freezing 2025-12-04T09:20:35.9261121Z * [new branch] mlazos/h-comp -> origin/mlazos/h-comp 2025-12-04T09:20:35.9262893Z * [new branch] mlazos/h-comp2 -> origin/mlazos/h-comp2 2025-12-04T09:20:35.9264685Z * [new branch] mlazos/hash-hop -> origin/mlazos/hash-hop 2025-12-04T09:20:35.9266471Z * [new branch] mlazos/hc -> origin/mlazos/hc 2025-12-04T09:20:35.9268322Z * [new branch] mlazos/hc-cycles -> origin/mlazos/hc-cycles 2025-12-04T09:20:35.9270060Z * [new branch] mlazos/hc-fixes -> origin/mlazos/hc-fixes 2025-12-04T09:20:35.9272100Z * [new branch] mlazos/hc-fixes3 -> origin/mlazos/hc-fixes3 2025-12-04T09:20:35.9273851Z * [new branch] mlazos/hc-fixes4 -> origin/mlazos/hc-fixes4 2025-12-04T09:20:35.9275710Z * [new branch] mlazos/hc-hf -> origin/mlazos/hc-hf 2025-12-04T09:20:35.9277438Z * [new branch] mlazos/hc-mut -> origin/mlazos/hc-mut 2025-12-04T09:20:35.9279741Z * [new branch] mlazos/hc10 -> origin/mlazos/hc10 2025-12-04T09:20:35.9281517Z * [new branch] mlazos/hc11 -> origin/mlazos/hc11 2025-12-04T09:20:35.9283206Z * [new branch] mlazos/hc12 -> origin/mlazos/hc12 2025-12-04T09:20:35.9284934Z * [new branch] mlazos/hc13 -> origin/mlazos/hc13 2025-12-04T09:20:35.9286760Z * [new branch] mlazos/hc14 -> origin/mlazos/hc14 2025-12-04T09:20:35.9288535Z * [new branch] mlazos/hc15 -> origin/mlazos/hc15 2025-12-04T09:20:35.9290365Z * [new branch] mlazos/hc2 -> origin/mlazos/hc2 2025-12-04T09:20:35.9292097Z * [new branch] mlazos/hc4 -> origin/mlazos/hc4 2025-12-04T09:20:35.9293911Z * [new branch] mlazos/hc5 -> origin/mlazos/hc5 2025-12-04T09:20:35.9295653Z * [new branch] mlazos/hc6 -> origin/mlazos/hc6 2025-12-04T09:20:35.9297380Z * [new branch] mlazos/hc7 -> origin/mlazos/hc7 2025-12-04T09:20:35.9299286Z * [new branch] mlazos/hc8 -> origin/mlazos/hc8 2025-12-04T09:20:35.9301077Z * [new branch] mlazos/hc9 -> origin/mlazos/hc9 2025-12-04T09:20:35.9302763Z * [new branch] mlazos/hc_baseline2 -> origin/mlazos/hc_baseline2 2025-12-04T09:20:35.9304444Z * [new branch] mlazos/inductor-streams -> origin/mlazos/inductor-streams 2025-12-04T09:20:35.9306027Z * [new branch] mlazos/main -> origin/mlazos/main 2025-12-04T09:20:35.9307797Z * [new branch] mlazos/mcg2 -> origin/mlazos/mcg2 2025-12-04T09:20:35.9309629Z * [new branch] mlazos/meta-guards -> origin/mlazos/meta-guards 2025-12-04T09:20:35.9312014Z * [new branch] mlazos/mlazos/foreach-map-adam -> origin/mlazos/mlazos/foreach-map-adam 2025-12-04T09:20:35.9313772Z * [new branch] mlazos/mlazos/tf-mode-backup -> origin/mlazos/mlazos/tf-mode-backup 2025-12-04T09:20:35.9315482Z * [new branch] mlazos/mod-fix -> origin/mlazos/mod-fix 2025-12-04T09:20:35.9317813Z * [new branch] mlazos/mode-fix -> origin/mlazos/mode-fix 2025-12-04T09:20:35.9319544Z * [new branch] mlazos/offsets -> origin/mlazos/offsets 2025-12-04T09:20:35.9321216Z * [new branch] mlazos/overguarding -> origin/mlazos/overguarding 2025-12-04T09:20:35.9323042Z * [new branch] mlazos/proxy-ctors -> origin/mlazos/proxy-ctors 2025-12-04T09:20:35.9324828Z * [new branch] mlazos/quant-fix -> origin/mlazos/quant-fix 2025-12-04T09:20:35.9326594Z * [new branch] mlazos/resnet-fix -> origin/mlazos/resnet-fix 2025-12-04T09:20:35.9328462Z * [new branch] mlazos/rm-buf-names -> origin/mlazos/rm-buf-names 2025-12-04T09:20:35.9330202Z * [new branch] mlazos/rm-code -> origin/mlazos/rm-code 2025-12-04T09:20:35.9331942Z * [new branch] mlazos/rm-spam -> origin/mlazos/rm-spam 2025-12-04T09:20:35.9333891Z * [new branch] mlazos/rtp -> origin/mlazos/rtp 2025-12-04T09:20:35.9335679Z * [new branch] mlazos/static-idx-dbg -> origin/mlazos/static-idx-dbg 2025-12-04T09:20:35.9337467Z * [new branch] mlazos/static-inputs-log -> origin/mlazos/static-inputs-log 2025-12-04T09:20:35.9339104Z * [new branch] mlazos/stests -> origin/mlazos/stests 2025-12-04T09:20:35.9341128Z * [new branch] mlazos/stream-ops -> origin/mlazos/stream-ops 2025-12-04T09:20:35.9342797Z * [new branch] mlazos/td-fix2 -> origin/mlazos/td-fix2 2025-12-04T09:20:35.9344597Z * [new branch] mlazos/tensor-hasattr2 -> origin/mlazos/tensor-hasattr2 2025-12-04T09:20:35.9346367Z * [new branch] mlazos/test -> origin/mlazos/test 2025-12-04T09:20:35.9348140Z * [new branch] mlazos/tf-mode -> origin/mlazos/tf-mode 2025-12-04T09:20:35.9349933Z * [new branch] mlazos/tf-mode-backup2 -> origin/mlazos/tf-mode-backup2 2025-12-04T09:20:35.9351750Z * [new branch] mlazos/tf-mode-reland -> origin/mlazos/tf-mode-reland 2025-12-04T09:20:35.9353660Z * [new branch] mlazos/tf-mode-reland2 -> origin/mlazos/tf-mode-reland2 2025-12-04T09:20:35.9355369Z * [new branch] mlazos/tf-mode-reland3 -> origin/mlazos/tf-mode-reland3 2025-12-04T09:20:35.9357194Z * [new branch] mlazos/triton-no-epi -> origin/mlazos/triton-no-epi 2025-12-04T09:20:35.9359058Z * [new branch] mlazos/tune-proto -> origin/mlazos/tune-proto 2025-12-04T09:20:35.9360766Z * [new branch] mlazos/tuple-fixes -> origin/mlazos/tuple-fixes 2025-12-04T09:20:35.9362594Z * [new branch] mlazos/tuple-fixes2 -> origin/mlazos/tuple-fixes2 2025-12-04T09:20:35.9364351Z * [new branch] mlazos/tuple-handling -> origin/mlazos/tuple-handling 2025-12-04T09:20:35.9366198Z * [new branch] mlazos/user-stream-base -> origin/mlazos/user-stream-base 2025-12-04T09:20:35.9367904Z * [new branch] mlazos/user-streams -> origin/mlazos/user-streams 2025-12-04T09:20:35.9369687Z * [new branch] mlazos/user-streams-backup -> origin/mlazos/user-streams-backup 2025-12-04T09:20:35.9371663Z * [new branch] mlazos/user-streams-backup2 -> origin/mlazos/user-streams-backup2 2025-12-04T09:20:35.9373389Z * [new branch] mlazos/vary-beta -> origin/mlazos/vary-beta 2025-12-04T09:20:35.9375194Z * [new branch] mlazos/vary-beta2 -> origin/mlazos/vary-beta2 2025-12-04T09:20:35.9376917Z * [new branch] mlazos/weird-perf1 -> origin/mlazos/weird-perf1 2025-12-04T09:20:35.9378759Z * [new branch] mm_out_dtype_compile -> origin/mm_out_dtype_compile 2025-12-04T09:20:35.9380745Z * [new branch] module-shim -> origin/module-shim 2025-12-04T09:20:35.9382583Z * [new branch] move_config -> origin/move_config 2025-12-04T09:20:35.9385136Z * [new branch] msaroufim/reduce -> origin/msaroufim/reduce 2025-12-04T09:20:35.9387431Z * [new branch] mtia/basic-cmake -> origin/mtia/basic-cmake 2025-12-04T09:20:35.9389857Z * [new branch] mwizak/fix-triton-block-shape -> origin/mwizak/fix-triton-block-shape 2025-12-04T09:20:35.9391610Z * [new branch] my_varlen_backup -> origin/my_varlen_backup 2025-12-04T09:20:35.9393338Z * [new branch] nativert_num_outputs -> origin/nativert_num_outputs 2025-12-04T09:20:35.9395147Z * [new branch] new-codegen -> origin/new-codegen 2025-12-04T09:20:35.9397099Z * [new branch] newtest-base -> origin/newtest-base 2025-12-04T09:20:35.9399418Z * [new branch] ngimel/addmm_dtype -> origin/ngimel/addmm_dtype 2025-12-04T09:20:35.9401021Z * [new branch] ngimel/div_inv -> origin/ngimel/div_inv 2025-12-04T09:20:35.9402704Z * [new branch] ngimel/error_index_list -> origin/ngimel/error_index_list 2025-12-04T09:20:35.9404397Z * [new branch] ngimel/gather_grid -> origin/ngimel/gather_grid 2025-12-04T09:20:35.9406137Z * [new branch] ngimel/gather_grid_release -> origin/ngimel/gather_grid_release 2025-12-04T09:20:35.9407759Z * [new branch] ngimel/gg_new -> origin/ngimel/gg_new 2025-12-04T09:20:35.9409494Z * [new branch] ngimel/hostalloc -> origin/ngimel/hostalloc 2025-12-04T09:20:35.9411139Z * [new branch] ngimel/storage_id -> origin/ngimel/storage_id 2025-12-04T09:20:35.9412877Z * [new branch] nightly -> origin/nightly 2025-12-04T09:20:35.9415369Z * [new branch] nikitaved/addmm_1_rowcol_lt_path_check -> origin/nikitaved/addmm_1_rowcol_lt_path_check 2025-12-04T09:20:35.9417049Z * [new branch] nikitaved/addmm_epilogue_fusions_2d_bias -> origin/nikitaved/addmm_epilogue_fusions_2d_bias 2025-12-04T09:20:35.9418783Z * [new branch] nikitaved/addmm_epilogue_fusions_inductor -> origin/nikitaved/addmm_epilogue_fusions_inductor 2025-12-04T09:20:35.9420941Z * [new branch] nikitaved/addmm_epilogue_fusions_scratch -> origin/nikitaved/addmm_epilogue_fusions_scratch 2025-12-04T09:20:35.9422992Z * [new branch] nikitaved/grad_addmm_epilogue_fusions -> origin/nikitaved/grad_addmm_epilogue_fusions 2025-12-04T09:20:35.9425100Z * [new branch] nikitaved/simpler_can_use_32bit_index -> origin/nikitaved/simpler_can_use_32bit_index 2025-12-04T09:20:35.9426888Z * [new branch] nikitaved/test -> origin/nikitaved/test 2025-12-04T09:20:35.9428943Z * [new branch] nmacchioni-perf-test-async-autotune -> origin/nmacchioni-perf-test-async-autotune 2025-12-04T09:20:35.9430725Z * [new branch] no_distributed_log_spew -> origin/no_distributed_log_spew 2025-12-04T09:20:35.9432620Z * [new branch] nofun-hack -> origin/nofun-hack 2025-12-04T09:20:35.9434402Z * [new branch] norm_bench -> origin/norm_bench 2025-12-04T09:20:35.9436853Z * [new branch] nullplay/fuse_matmul -> origin/nullplay/fuse_matmul 2025-12-04T09:20:35.9438655Z * [new branch] nullplay_fuse_matmul -> origin/nullplay_fuse_matmul 2025-12-04T09:20:35.9440541Z * [new branch] optimizer_test -> origin/optimizer_test 2025-12-04T09:20:35.9443407Z * [new branch] orig/release/1.10 -> origin/orig/release/1.10 2025-12-04T09:20:35.9445754Z * [new branch] orig/release/1.11 -> origin/orig/release/1.11 2025-12-04T09:20:35.9447664Z * [new branch] orig/release/1.12 -> origin/orig/release/1.12 2025-12-04T09:20:35.9449607Z * [new branch] orig/release/1.13 -> origin/orig/release/1.13 2025-12-04T09:20:35.9451421Z * [new branch] orig/release/1.6 -> origin/orig/release/1.6 2025-12-04T09:20:35.9453309Z * [new branch] orig/release/1.7 -> origin/orig/release/1.7 2025-12-04T09:20:35.9455116Z * [new branch] orig/release/1.8 -> origin/orig/release/1.8 2025-12-04T09:20:35.9456964Z * [new branch] orig/release/1.9 -> origin/orig/release/1.9 2025-12-04T09:20:35.9458733Z * [new branch] orig/release/2.0 -> origin/orig/release/2.0 2025-12-04T09:20:35.9460602Z * [new branch] orig/release/2.1 -> origin/orig/release/2.1 2025-12-04T09:20:35.9462379Z * [new branch] orig/release/2.2 -> origin/orig/release/2.2 2025-12-04T09:20:35.9464275Z * [new branch] orig/release/2.3 -> origin/orig/release/2.3 2025-12-04T09:20:35.9466018Z * [new branch] orig/release/2.4 -> origin/orig/release/2.4 2025-12-04T09:20:35.9467705Z * [new branch] orig/release/2.5 -> origin/orig/release/2.5 2025-12-04T09:20:35.9469447Z * [new branch] orig/release/2.6 -> origin/orig/release/2.6 2025-12-04T09:20:35.9471660Z * [new branch] orig/release/2.7 -> origin/orig/release/2.7 2025-12-04T09:20:35.9475213Z * [new branch] orig/release/2.8 -> origin/orig/release/2.8 2025-12-04T09:20:35.9477017Z * [new branch] orig/release/2.9 -> origin/orig/release/2.9 2025-12-04T09:20:35.9480731Z * [new branch] origin/gh/fxdawnn/1/base -> origin/origin/gh/fxdawnn/1/base 2025-12-04T09:20:35.9482434Z * [new branch] origin/gh/fxdawnn/1/orig -> origin/origin/gh/fxdawnn/1/orig 2025-12-04T09:20:35.9485386Z * [new branch] origin/gh/zpcore/14/orig -> origin/origin/gh/zpcore/14/orig 2025-12-04T09:20:35.9487321Z * [new branch] oulgen-patch-1 -> origin/oulgen-patch-1 2025-12-04T09:20:35.9489662Z * [new branch] oulgen-patch-2 -> origin/oulgen-patch-2 2025-12-04T09:20:35.9491584Z * [new branch] oulgen-patch-3 -> origin/oulgen-patch-3 2025-12-04T09:20:35.9493401Z * [new branch] oulgen-patch-4 -> origin/oulgen-patch-4 2025-12-04T09:20:35.9495676Z * [new branch] padded-tensor -> origin/padded-tensor 2025-12-04T09:20:35.9497543Z * [new branch] pca2 -> origin/pca2 2025-12-04T09:20:35.9499555Z * [new branch] per_channel_backup -> origin/per_channel_backup 2025-12-04T09:20:35.9501872Z * [new branch] perf_ops -> origin/perf_ops 2025-12-04T09:20:35.9503345Z * [new branch] perf_ops_2_9 -> origin/perf_ops_2_9 2025-12-04T09:20:35.9505219Z * [new branch] pianpwk-patch-1 -> origin/pianpwk-patch-1 2025-12-04T09:20:35.9507595Z * [new branch] pianpwk/__draft_debug_mode -> origin/pianpwk/__draft_debug_mode 2025-12-04T09:20:35.9509413Z * [new branch] pianpwk/_debug_mode_for_triton_draft -> origin/pianpwk/_debug_mode_for_triton_draft 2025-12-04T09:20:35.9511109Z * [new branch] pianpwk/_debug_nn_module_compile -> origin/pianpwk/_debug_nn_module_compile 2025-12-04T09:20:35.9512844Z * [new branch] pianpwk/_draft_triton_11_3 -> origin/pianpwk/_draft_triton_11_3 2025-12-04T09:20:35.9514445Z * [new branch] pianpwk/_manual_bucket_draft -> origin/pianpwk/_manual_bucket_draft 2025-12-04T09:20:35.9516569Z * [new branch] pianpwk/_profile_w_dispatch_keys -> origin/pianpwk/_profile_w_dispatch_keys 2025-12-04T09:20:35.9518681Z * [new branch] pianpwk/_super_draft_debug_mode -> origin/pianpwk/_super_draft_debug_mode 2025-12-04T09:20:35.9520540Z * [new branch] pianpwk/_unbacked_local_shard_size -> origin/pianpwk/_unbacked_local_shard_size 2025-12-04T09:20:35.9522296Z * [new branch] pianpwk/anomaly_tb -> origin/pianpwk/anomaly_tb 2025-12-04T09:20:35.9524016Z * [new branch] pianpwk/auto_fx_annotate -> origin/pianpwk/auto_fx_annotate 2025-12-04T09:20:35.9525856Z * [new branch] pianpwk/backed_size_oblivious_export -> origin/pianpwk/backed_size_oblivious_export 2025-12-04T09:20:35.9527532Z * [new branch] pianpwk/bert_dynamic_perf -> origin/pianpwk/bert_dynamic_perf 2025-12-04T09:20:35.9529409Z * [new branch] pianpwk/debug_fwd_stack_traces -> origin/pianpwk/debug_fwd_stack_traces 2025-12-04T09:20:35.9531150Z * [new branch] pianpwk/debug_hash_tensor -> origin/pianpwk/debug_hash_tensor 2025-12-04T09:20:35.9532862Z * [new branch] pianpwk/debug_mode_annotate -> origin/pianpwk/debug_mode_annotate 2025-12-04T09:20:35.9534607Z * [new branch] pianpwk/debug_mode_defaults -> origin/pianpwk/debug_mode_defaults 2025-12-04T09:20:35.9536256Z * [new branch] pianpwk/debug_mode_hacks -> origin/pianpwk/debug_mode_hacks 2025-12-04T09:20:35.9538053Z * [new branch] pianpwk/debug_mode_opcall_refactor -> origin/pianpwk/debug_mode_opcall_refactor 2025-12-04T09:20:35.9539867Z * [new branch] pianpwk/debug_mode_show_ids -> origin/pianpwk/debug_mode_show_ids 2025-12-04T09:20:35.9541630Z * [new branch] pianpwk/debug_mode_triton -> origin/pianpwk/debug_mode_triton 2025-12-04T09:20:35.9543462Z * [new branch] pianpwk/debug_show_stack_trace -> origin/pianpwk/debug_show_stack_trace 2025-12-04T09:20:35.9545321Z * [new branch] pianpwk/debug_wait_on_collective -> origin/pianpwk/debug_wait_on_collective 2025-12-04T09:20:35.9547103Z * [new branch] pianpwk/debugmode_compile_tf -> origin/pianpwk/debugmode_compile_tf 2025-12-04T09:20:35.9548921Z * [new branch] pianpwk/dispatch_key_debugging_for_debug -> origin/pianpwk/dispatch_key_debugging_for_debug 2025-12-04T09:20:35.9551109Z * [new branch] pianpwk/draft_debug_mode_tfcompile -> origin/pianpwk/draft_debug_mode_tfcompile 2025-12-04T09:20:35.9552874Z * [new branch] pianpwk/draft_multikernel_nn -> origin/pianpwk/draft_multikernel_nn 2025-12-04T09:20:35.9554872Z * [new branch] pianpwk/draft_multikernel_status_10_5 -> origin/pianpwk/draft_multikernel_status_10_5 2025-12-04T09:20:35.9556698Z * [new branch] pianpwk/dtensor_custom_chunk -> origin/pianpwk/dtensor_custom_chunk 2025-12-04T09:20:35.9558509Z * [new branch] pianpwk/dtensor_unbacked_keypath -> origin/pianpwk/dtensor_unbacked_keypath 2025-12-04T09:20:35.9560386Z * [new branch] pianpwk/event_list_tree -> origin/pianpwk/event_list_tree 2025-12-04T09:20:35.9562154Z * [new branch] pianpwk/false_numel_refs -> origin/pianpwk/false_numel_refs 2025-12-04T09:20:35.9563882Z * [new branch] pianpwk/maybe_guard_rel -> origin/pianpwk/maybe_guard_rel 2025-12-04T09:20:35.9565662Z * [new branch] pianpwk/multikernel_hints_draft -> origin/pianpwk/multikernel_hints_draft 2025-12-04T09:20:35.9567472Z * [new branch] pianpwk/no_size_oblivious_slice_scat -> origin/pianpwk/no_size_oblivious_slice_scat 2025-12-04T09:20:35.9569220Z * [new branch] pianpwk/oblivious_reshape_view_better -> origin/pianpwk/oblivious_reshape_view_better 2025-12-04T09:20:35.9571165Z * [new branch] pianpwk/pre_forward_hook -> origin/pianpwk/pre_forward_hook 2025-12-04T09:20:35.9573066Z * [new branch] pianpwk/skip_python_keys_alternate -> origin/pianpwk/skip_python_keys_alternate 2025-12-04T09:20:35.9574910Z * [new branch] pianpwk/skip_python_keys_in_guards -> origin/pianpwk/skip_python_keys_in_guards 2025-12-04T09:20:35.9576593Z * [new branch] pianpwk/sym_tokens_draft -> origin/pianpwk/sym_tokens_draft 2025-12-04T09:20:35.9579500Z * [new branch] pianpwk/symint_one_hot -> origin/pianpwk/symint_one_hot 2025-12-04T09:20:35.9580596Z * [new branch] pianpwk/test_pointwise_guard_or_false -> origin/pianpwk/test_pointwise_guard_or_false 2025-12-04T09:20:35.9582114Z * [new branch] pianpwk/totally_draft_sym_wrap -> origin/pianpwk/totally_draft_sym_wrap 2025-12-04T09:20:35.9583918Z * [new branch] pianpwk/try_dumb_stuff -> origin/pianpwk/try_dumb_stuff 2025-12-04T09:20:35.9585585Z * [new branch] pianpwk/try_dumb_stuff_2 -> origin/pianpwk/try_dumb_stuff_2 2025-12-04T09:20:35.9587357Z * [new branch] pianpwk/unbacked_dtensor_mm -> origin/pianpwk/unbacked_dtensor_mm 2025-12-04T09:20:35.9589122Z * [new branch] pianpwk/unbacked_tracing_12_2 -> origin/pianpwk/unbacked_tracing_12_2 2025-12-04T09:20:35.9590902Z * [new branch] pianpwk/user_symints -> origin/pianpwk/user_symints 2025-12-04T09:20:35.9592567Z * [new branch] pianpwk/wan21_reshape -> origin/pianpwk/wan21_reshape 2025-12-04T09:20:35.9594913Z * [new branch] piz/fix_partial_backward_1112 -> origin/piz/fix_partial_backward_1112 2025-12-04T09:20:35.9596536Z * [new branch] piz/prop_cache_clean -> origin/piz/prop_cache_clean 2025-12-04T09:20:35.9598438Z * [new branch] pool-separate -> origin/pool-separate 2025-12-04T09:20:35.9600233Z * [new branch] pr-156087 -> origin/pr-156087 2025-12-04T09:20:35.9602586Z * [new branch] pr/131860 -> origin/pr/131860 2025-12-04T09:20:35.9604515Z * [new branch] predispatch_to -> origin/predispatch_to 2025-12-04T09:20:35.9606371Z * [new branch] protect-c17 -> origin/protect-c17 2025-12-04T09:20:35.9608135Z * [new branch] pt-opt-cuda3 -> origin/pt-opt-cuda3 2025-12-04T09:20:35.9610487Z * [new branch] python_compiled_autograd -> origin/python_compiled_autograd 2025-12-04T09:20:35.9613045Z * [new branch] q1l1/fix_device_moved_constant_type_unknown -> origin/q1l1/fix_device_moved_constant_type_unknown 2025-12-04T09:20:35.9614780Z * [new branch] q1l1/fix_wrong_default_type_for_kernel_call_args -> origin/q1l1/fix_wrong_default_type_for_kernel_call_args 2025-12-04T09:20:35.9617467Z * [new branch] qchip/export-D54134695 -> origin/qchip/export-D54134695 2025-12-04T09:20:35.9619539Z * [new branch] quote-pytest_cache -> origin/quote-pytest_cache 2025-12-04T09:20:35.9621941Z * [new branch] reland-accgrad-stream-warn -> origin/reland-accgrad-stream-warn 2025-12-04T09:20:35.9624677Z * [new branch] release/1.10 -> origin/release/1.10 2025-12-04T09:20:35.9626402Z * [new branch] release/1.11 -> origin/release/1.11 2025-12-04T09:20:35.9628458Z * [new branch] release/1.12 -> origin/release/1.12 2025-12-04T09:20:35.9630204Z * [new branch] release/1.13 -> origin/release/1.13 2025-12-04T09:20:35.9631962Z * [new branch] release/1.4 -> origin/release/1.4 2025-12-04T09:20:35.9633449Z * [new branch] release/1.4.1 -> origin/release/1.4.1 2025-12-04T09:20:35.9635180Z * [new branch] release/1.5 -> origin/release/1.5 2025-12-04T09:20:35.9637015Z * [new branch] release/1.6 -> origin/release/1.6 2025-12-04T09:20:35.9638735Z * [new branch] release/1.7 -> origin/release/1.7 2025-12-04T09:20:35.9640572Z * [new branch] release/1.8 -> origin/release/1.8 2025-12-04T09:20:35.9642364Z * [new branch] release/1.9 -> origin/release/1.9 2025-12-04T09:20:35.9644123Z * [new branch] release/2.0 -> origin/release/2.0 2025-12-04T09:20:35.9646105Z * [new branch] release/2.1 -> origin/release/2.1 2025-12-04T09:20:35.9648347Z * [new branch] release/2.2 -> origin/release/2.2 2025-12-04T09:20:35.9650401Z * [new branch] release/2.3 -> origin/release/2.3 2025-12-04T09:20:35.9652663Z * [new branch] release/2.4 -> origin/release/2.4 2025-12-04T09:20:35.9654977Z * [new branch] release/2.5 -> origin/release/2.5 2025-12-04T09:20:35.9656828Z * [new branch] release/2.6 -> origin/release/2.6 2025-12-04T09:20:35.9658888Z * [new branch] release/2.7 -> origin/release/2.7 2025-12-04T09:20:35.9660886Z * [new branch] release/2.8 -> origin/release/2.8 2025-12-04T09:20:35.9662651Z * [new branch] release/2.9 -> origin/release/2.9 2025-12-04T09:20:35.9664482Z * [new branch] release_notes -> origin/release_notes 2025-12-04T09:20:35.9666371Z * [new branch] remove_pyinterpreter -> origin/remove_pyinterpreter 2025-12-04T09:20:35.9668491Z * [new branch] replace-pytorch-labs-20250812-195836 -> origin/replace-pytorch-labs-20250812-195836 2025-12-04T09:20:35.9670214Z * [new branch] replace-pytorch-labs-20250812-200248 -> origin/replace-pytorch-labs-20250812-200248 2025-12-04T09:20:35.9672091Z * [new branch] replace-pytorch-labs-20250812-200324 -> origin/replace-pytorch-labs-20250812-200324 2025-12-04T09:20:35.9673813Z * [new branch] replace-pytorch-labs-20250812-204020 -> origin/replace-pytorch-labs-20250812-204020 2025-12-04T09:20:35.9677216Z * [new branch] revert-131069-gh/krzysztofjordan/1/head -> origin/revert-131069-gh/krzysztofjordan/1/head 2025-12-04T09:20:35.9681662Z * [new branch] revert-131469-gh/andrewor14/51/head -> origin/revert-131469-gh/andrewor14/51/head 2025-12-04T09:20:35.9685564Z * [new branch] revert-152361-gh/fadara01/1/head -> origin/revert-152361-gh/fadara01/1/head 2025-12-04T09:20:35.9688840Z * [new branch] revert-156870-gh/skarjala/3/head -> origin/revert-156870-gh/skarjala/3/head 2025-12-04T09:20:35.9690945Z * [new branch] revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ -> origin/revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ 2025-12-04T09:20:35.9692539Z * [new branch] revert-hoo-invoke-subgraph -> origin/revert-hoo-invoke-subgraph 2025-12-04T09:20:35.9694564Z * [new branch] revert_always_build_distributed -> origin/revert_always_build_distributed 2025-12-04T09:20:35.9696289Z * [new branch] rms_norm_patch -> origin/rms_norm_patch 2025-12-04T09:20:35.9698921Z * [new branch] ruisi/fix_all_to_all_estimation -> origin/ruisi/fix_all_to_all_estimation 2025-12-04T09:20:35.9700512Z * [new branch] ruisi/fix_comm_estimation -> origin/ruisi/fix_comm_estimation 2025-12-04T09:20:35.9709067Z * [new branch] ruisi/fix_dynamic_shape_estimation -> origin/ruisi/fix_dynamic_shape_estimation 2025-12-04T09:20:35.9709452Z * [new branch] ruisi/fix_llama3_autobucketing -> origin/ruisi/fix_llama3_autobucketing 2025-12-04T09:20:35.9709739Z * [new branch] ruisi/fix_manual_bucketing_ep_pass -> origin/ruisi/fix_manual_bucketing_ep_pass 2025-12-04T09:20:35.9710147Z * [new branch] ruisi/manual_bucket_pass -> origin/ruisi/manual_bucket_pass 2025-12-04T09:20:35.9710519Z * [new branch] ryanguo99/cleanup-dynamo-expected-failures -> origin/ryanguo99/cleanup-dynamo-expected-failures 2025-12-04T09:20:35.9711983Z * [new branch] ryanguo99/fix-closure-var -> origin/ryanguo99/fix-closure-var 2025-12-04T09:20:35.9714561Z * [new branch] rzou/faketensor_bench -> origin/rzou/faketensor_bench 2025-12-04T09:20:35.9716067Z * [new branch] rzou/njt -> origin/rzou/njt 2025-12-04T09:20:35.9717761Z * [new branch] rzou/pca -> origin/rzou/pca 2025-12-04T09:20:35.9719422Z * [new branch] rzou/realprop -> origin/rzou/realprop 2025-12-04T09:20:35.9721229Z * [new branch] samplevllm -> origin/samplevllm 2025-12-04T09:20:35.9723911Z * [new branch] sanchitintel/weird_thing_with_test_cpu_select_algorithm -> origin/sanchitintel/weird_thing_with_test_cpu_select_algorithm 2025-12-04T09:20:35.9725785Z * [new branch] sapling-pr-archive-SS-JIA -> origin/sapling-pr-archive-SS-JIA 2025-12-04T09:20:35.9727825Z * [new branch] sapling-pr-archive-tushar00jain -> origin/sapling-pr-archive-tushar00jain 2025-12-04T09:20:35.9729517Z * [new branch] save -> origin/save 2025-12-04T09:20:35.9731424Z * [new branch] scaled_mm -> origin/scaled_mm 2025-12-04T09:20:35.9733246Z * [new branch] scan_attempt -> origin/scan_attempt 2025-12-04T09:20:35.9735584Z * [new branch] sdym/2.5.1 -> origin/sdym/2.5.1 2025-12-04T09:20:35.9737486Z * [new branch] sekyondaMeta-dynamoconfig-fix -> origin/sekyondaMeta-dynamoconfig-fix 2025-12-04T09:20:35.9739896Z * [new branch] shengf/fx-xform-perf -> origin/shengf/fx-xform-perf 2025-12-04T09:20:35.9741804Z * [new branch] shoumikhin-patch-1 -> origin/shoumikhin-patch-1 2025-12-04T09:20:35.9743603Z * [new branch] solve-accuracy-fix -> origin/solve-accuracy-fix 2025-12-04T09:20:35.9745390Z * [new branch] some_rocm_inductor_skips -> origin/some_rocm_inductor_skips 2025-12-04T09:20:35.9747805Z * [new branch] soulitzer/stash-tls-ac -> origin/soulitzer/stash-tls-ac 2025-12-04T09:20:35.9749659Z * [new branch] sparse-mm-bf16-support -> origin/sparse-mm-bf16-support 2025-12-04T09:20:35.9751537Z * [new branch] starterTaskUpdate -> origin/starterTaskUpdate 2025-12-04T09:20:35.9753431Z * [new branch] suo -> origin/suo 2025-12-04T09:20:35.9755272Z * [new branch] sve-poc -> origin/sve-poc 2025-12-04T09:20:35.9757158Z * [new branch] switch-bn -> origin/switch-bn 2025-12-04T09:20:35.9758996Z * [new branch] sy_annotation_in_autograd_hop -> origin/sy_annotation_in_autograd_hop 2025-12-04T09:20:35.9760816Z * [new branch] sy_aot_eager_record -> origin/sy_aot_eager_record 2025-12-04T09:20:35.9762594Z * [new branch] sy_custom_bucketing -> origin/sy_custom_bucketing 2025-12-04T09:20:35.9764725Z * [new branch] sy_debug_mode_test -> origin/sy_debug_mode_test 2025-12-04T09:20:35.9766487Z * [new branch] sy_deserialize -> origin/sy_deserialize 2025-12-04T09:20:35.9768225Z * [new branch] sy_dump_gm_code -> origin/sy_dump_gm_code 2025-12-04T09:20:35.9770007Z * [new branch] sy_exp -> origin/sy_exp 2025-12-04T09:20:35.9772042Z * [new branch] sy_export_annotation -> origin/sy_export_annotation 2025-12-04T09:20:35.9773877Z * [new branch] sy_invoke_subgraph -> origin/sy_invoke_subgraph 2025-12-04T09:20:35.9775721Z * [new branch] sy_kernel_bw_name -> origin/sy_kernel_bw_name 2025-12-04T09:20:35.9777512Z * [new branch] sy_multi_arch -> origin/sy_multi_arch 2025-12-04T09:20:35.9779388Z * [new branch] sy_nn_module_stack -> origin/sy_nn_module_stack 2025-12-04T09:20:35.9781368Z * [new branch] sy_original_dtensor -> origin/sy_original_dtensor 2025-12-04T09:20:35.9783176Z * [new branch] sy_profiler_cia -> origin/sy_profiler_cia 2025-12-04T09:20:35.9784949Z * [new branch] symm_mem_sync -> origin/symm_mem_sync 2025-12-04T09:20:35.9786894Z * [new branch] sympy-bottleneck-repro -> origin/sympy-bottleneck-repro 2025-12-04T09:20:35.9788788Z * [new branch] tensordict_integration -> origin/tensordict_integration 2025-12-04T09:20:35.9790656Z * [new branch] test-move-conda-builds -> origin/test-move-conda-builds 2025-12-04T09:20:35.9792476Z * [new branch] test-old -> origin/test-old 2025-12-04T09:20:35.9794961Z * [new branch] test/bmm_heur -> origin/test/bmm_heur 2025-12-04T09:20:35.9797873Z * [new branch] tianren/customOp_autotune_fix -> origin/tianren/customOp_autotune_fix 2025-12-04T09:20:35.9799583Z * [new branch] tianren/customOp_enable_max_autotune -> origin/tianren/customOp_enable_max_autotune 2025-12-04T09:20:35.9801188Z * [new branch] tianren/customOp_fusion -> origin/tianren/customOp_fusion 2025-12-04T09:20:35.9802878Z * [new branch] tianren/customop_collectiveop_benchmark -> origin/tianren/customop_collectiveop_benchmark 2025-12-04T09:20:35.9804922Z * [new branch] tianren/customop_collectiveop_benchmark_fix -> origin/tianren/customop_collectiveop_benchmark_fix 2025-12-04T09:20:35.9806988Z * [new branch] tianren/customop_dynamic_config -> origin/tianren/customop_dynamic_config 2025-12-04T09:20:35.9808728Z * [new branch] tianren/dynamic_range_input -> origin/tianren/dynamic_range_input 2025-12-04T09:20:35.9810670Z * [new branch] tianren/dynamic_range_input_fix -> origin/tianren/dynamic_range_input_fix 2025-12-04T09:20:35.9812496Z * [new branch] tianren/dynamic_range_input_merge -> origin/tianren/dynamic_range_input_merge 2025-12-04T09:20:35.9814221Z * [new branch] tianren/flex_paged_attn_fix_temp -> origin/tianren/flex_paged_attn_fix_temp 2025-12-04T09:20:35.9816069Z * [new branch] tianren/fx_codegen_dump -> origin/tianren/fx_codegen_dump 2025-12-04T09:20:35.9817869Z * [new branch] tianren/symmetric_memory -> origin/tianren/symmetric_memory 2025-12-04T09:20:35.9819634Z * [new branch] tianren/test -> origin/tianren/test 2025-12-04T09:20:35.9821916Z * [new branch] tidy_performance_cyy -> origin/tidy_performance_cyy 2025-12-04T09:20:35.9823581Z * [new branch] tmp -> origin/tmp 2025-12-04T09:20:35.9825949Z * [new branch] torchtitan_ep -> origin/torchtitan_ep 2025-12-04T09:20:35.9827831Z * [new branch] torchtitan_integration -> origin/torchtitan_integration 2025-12-04T09:20:35.9829814Z * [new branch] trace_fsdp_torchtune_lora -> origin/trace_fsdp_torchtune_lora 2025-12-04T09:20:35.9831571Z * [new branch] traceable_fsdp_unit_tests -> origin/traceable_fsdp_unit_tests 2025-12-04T09:20:35.9833432Z * [new branch] tree_loop_vec_base -> origin/tree_loop_vec_base 2025-12-04T09:20:35.9835283Z * [new branch] triton_kernel -> origin/triton_kernel 2025-12-04T09:20:35.9837086Z * [new branch] tt_pkg_1908 -> origin/tt_pkg_1908 2025-12-04T09:20:35.9838932Z * [new branch] type_dec -> origin/type_dec 2025-12-04T09:20:35.9840900Z * [new branch] udate-sphinx-dependancies -> origin/udate-sphinx-dependancies 2025-12-04T09:20:35.9843919Z * [new branch] update-audio-commit-hash/17630256502-1803-1 -> origin/update-audio-commit-hash/17630256502-1803-1 2025-12-04T09:20:35.9845595Z * [new branch] update-audio-commit-hash/19087141161-1916-1 -> origin/update-audio-commit-hash/19087141161-1916-1 2025-12-04T09:20:35.9847308Z * [new branch] update-audio-commit-hash/19250643381-1929-1 -> origin/update-audio-commit-hash/19250643381-1929-1 2025-12-04T09:20:35.9849028Z * [new branch] update-audio-commit-hash/19397724337-1935-1 -> origin/update-audio-commit-hash/19397724337-1935-1 2025-12-04T09:20:35.9850869Z * [new branch] update-audio-commit-hash/19555670148-1941-1 -> origin/update-audio-commit-hash/19555670148-1941-1 2025-12-04T09:20:35.9852853Z * [new branch] update-audio-commit-hash/19750627930-1946-1 -> origin/update-audio-commit-hash/19750627930-1946-1 2025-12-04T09:20:35.9855229Z * [new branch] update-triton-commit-hash/13663274526-1487-2 -> origin/update-triton-commit-hash/13663274526-1487-2 2025-12-04T09:20:35.9857642Z * [new branch] update-vision-commit-hash/19087141161-1916-1 -> origin/update-vision-commit-hash/19087141161-1916-1 2025-12-04T09:20:35.9859371Z * [new branch] update-vision-commit-hash/19184897099-1925-1 -> origin/update-vision-commit-hash/19184897099-1925-1 2025-12-04T09:20:35.9861024Z * [new branch] update-vision-commit-hash/19250643381-1929-1 -> origin/update-vision-commit-hash/19250643381-1929-1 2025-12-04T09:20:35.9862738Z * [new branch] update-vision-commit-hash/19381328640-1934-1 -> origin/update-vision-commit-hash/19381328640-1934-1 2025-12-04T09:20:35.9864369Z * [new branch] update-vision-commit-hash/19485237164-1938-1 -> origin/update-vision-commit-hash/19485237164-1938-1 2025-12-04T09:20:35.9866914Z * [new branch] update-vllm-commit-hash/18451675449-1879-1 -> origin/update-vllm-commit-hash/18451675449-1879-1 2025-12-04T09:20:35.9868726Z * [new branch] update-vllm-dockerfile -> origin/update-vllm-dockerfile 2025-12-04T09:20:35.9871421Z * [new branch] update-xla-commit-hash/19224287370-211-1 -> origin/update-xla-commit-hash/19224287370-211-1 2025-12-04T09:20:35.9873255Z * [new branch] update-xla-commit-hash/19422028566-212-1 -> origin/update-xla-commit-hash/19422028566-212-1 2025-12-04T09:20:35.9874893Z * [new branch] update-xla-commit-hash/19626841311-213-1 -> origin/update-xla-commit-hash/19626841311-213-1 2025-12-04T09:20:35.9876864Z * [new branch] update_docs_torch_multinomial_issue#125388 -> origin/update_docs_torch_multinomial_issue#125388 2025-12-04T09:20:35.9878506Z * [new branch] update_operator_readme -> origin/update_operator_readme 2025-12-04T09:20:35.9880389Z * [new branch] update_slow_tests_1722488736 -> origin/update_slow_tests_1722488736 2025-12-04T09:20:35.9882152Z * [new branch] update_slow_tests_1722879173 -> origin/update_slow_tests_1722879173 2025-12-04T09:20:35.9884058Z * [new branch] update_slow_tests_1762155677 -> origin/update_slow_tests_1762155677 2025-12-04T09:20:35.9886119Z * [new branch] update_slow_tests_1763365283 -> origin/update_slow_tests_1763365283 2025-12-04T09:20:35.9887775Z * [new branch] update_submodule_FBGEMM -> origin/update_submodule_FBGEMM 2025-12-04T09:20:35.9889574Z * [new branch] update_submodule_kineto -> origin/update_submodule_kineto 2025-12-04T09:20:35.9891407Z * [new branch] update_submodule_tensorpipe -> origin/update_submodule_tensorpipe 2025-12-04T09:20:35.9893361Z * [new branch] upload-tests-for-autorevert -> origin/upload-tests-for-autorevert 2025-12-04T09:20:35.9895158Z * [new branch] v0.1.2 -> origin/v0.1.2 2025-12-04T09:20:35.9897184Z * [new branch] v1.0.1 -> origin/v1.0.1 2025-12-04T09:20:35.9899171Z * [new branch] v1.0.3 -> origin/v1.0.3 2025-12-04T09:20:35.9901407Z * [new branch] v1.1.0 -> origin/v1.1.0 2025-12-04T09:20:35.9903261Z * [new branch] v1.2.0 -> origin/v1.2.0 2025-12-04T09:20:35.9905127Z * [new branch] v1.3.0 -> origin/v1.3.0 2025-12-04T09:20:35.9907045Z * [new branch] v1.3.1 -> origin/v1.3.1 2025-12-04T09:20:35.9908956Z * [new branch] validate_fn -> origin/validate_fn 2025-12-04T09:20:35.9910886Z * [new branch] validations_2.6 -> origin/validations_2.6 2025-12-04T09:20:35.9912836Z * [new branch] validations_2.8 -> origin/validations_2.8 2025-12-04T09:20:35.9914772Z * [new branch] varlen-api -> origin/varlen-api 2025-12-04T09:20:35.9916652Z * [new branch] varlen-api-backup -> origin/varlen-api-backup 2025-12-04T09:20:35.9918428Z * [new branch] varlen_batch_invariance -> origin/varlen_batch_invariance 2025-12-04T09:20:35.9920607Z * [new branch] viable/strict -> origin/viable/strict 2025-12-04T09:20:35.9923242Z * [new branch] vishal9-team/dtensor_parallelism_toy -> origin/vishal9-team/dtensor_parallelism_toy 2025-12-04T09:20:35.9925045Z * [new branch] vllmbuildci -> origin/vllmbuildci 2025-12-04T09:20:35.9926883Z * [new branch] vllmpin -> origin/vllmpin 2025-12-04T09:20:35.9928838Z * [new branch] vscode-recommend-pyrefly -> origin/vscode-recommend-pyrefly 2025-12-04T09:20:35.9930653Z * [new branch] wdvr-patch-1 -> origin/wdvr-patch-1 2025-12-04T09:20:35.9932986Z * [new branch] wdvr/iss_145259 -> origin/wdvr/iss_145259 2025-12-04T09:20:35.9935392Z * [new branch] whc/pei -> origin/whc/pei 2025-12-04T09:20:35.9937144Z * [new branch] whc/pp_fix -> origin/whc/pp_fix 2025-12-04T09:20:35.9938892Z * [new branch] whc/sharding -> origin/whc/sharding 2025-12-04T09:20:35.9940740Z * [new branch] whc/sharding2 -> origin/whc/sharding2 2025-12-04T09:20:35.9942198Z * [new branch] whc/uneven -> origin/whc/uneven 2025-12-04T09:20:35.9944167Z * [new branch] whc/uneven-merge -> origin/whc/uneven-merge 2025-12-04T09:20:35.9946030Z * [new branch] win_warnings -> origin/win_warnings 2025-12-04T09:20:35.9947927Z * [new branch] windows_libtorch_free -> origin/windows_libtorch_free 2025-12-04T09:20:35.9949728Z * [new branch] xmfan-war -> origin/xmfan-war 2025-12-04T09:20:35.9952077Z * [new branch] xmfan/ca_0516 -> origin/xmfan/ca_0516 2025-12-04T09:20:35.9953786Z * [new branch] xmfan/ca_1051b93192 -> origin/xmfan/ca_1051b93192 2025-12-04T09:20:35.9955620Z * [new branch] xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 -> origin/xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 2025-12-04T09:20:35.9957001Z * [new branch] xmfan/ca_5a2be192d1 -> origin/xmfan/ca_5a2be192d1 2025-12-04T09:20:35.9958617Z * [new branch] xmfan/ca_9d59b516e9 -> origin/xmfan/ca_9d59b516e9 2025-12-04T09:20:35.9960251Z * [new branch] xmfan/ca_apr8 -> origin/xmfan/ca_apr8 2025-12-04T09:20:35.9961971Z * [new branch] xmfan/ca_base -> origin/xmfan/ca_base 2025-12-04T09:20:35.9963927Z * [new branch] xmfan/ca_dynamic -> origin/xmfan/ca_dynamic 2025-12-04T09:20:35.9966134Z * [new branch] xmfan/ca_fix_dyn -> origin/xmfan/ca_fix_dyn 2025-12-04T09:20:35.9967968Z * [new branch] xmfan/ca_fix_lowering -> origin/xmfan/ca_fix_lowering 2025-12-04T09:20:35.9969732Z * [new branch] xmfan/ca_fix_polyfills -> origin/xmfan/ca_fix_polyfills 2025-12-04T09:20:35.9971999Z * [new branch] xmfan/ca_jan3 -> origin/xmfan/ca_jan3 2025-12-04T09:20:35.9973777Z * [new branch] xmfan/ca_jun18 -> origin/xmfan/ca_jun18 2025-12-04T09:20:35.9975712Z * [new branch] xmfan/ca_jun24 -> origin/xmfan/ca_jun24 2025-12-04T09:20:35.9977471Z * [new branch] xmfan/ca_nested -> origin/xmfan/ca_nested 2025-12-04T09:20:35.9979291Z * [new branch] xmfan/ca_overhead -> origin/xmfan/ca_overhead 2025-12-04T09:20:35.9981153Z * [new branch] xmfan/ca_overhead_0eba7e5451 -> origin/xmfan/ca_overhead_0eba7e5451 2025-12-04T09:20:35.9982900Z * [new branch] xmfan/cacu_jun18 -> origin/xmfan/cacu_jun18 2025-12-04T09:20:35.9984613Z * [new branch] xmfan/cacu_jun19 -> origin/xmfan/cacu_jun19 2025-12-04T09:20:35.9986396Z * [new branch] xmfan/cacu_jun4 -> origin/xmfan/cacu_jun4 2025-12-04T09:20:35.9988180Z * [new branch] xmfan/disable_duck_shape -> origin/xmfan/disable_duck_shape 2025-12-04T09:20:35.9989943Z * [new branch] xmfan/fca_cpp_node_passthrough -> origin/xmfan/fca_cpp_node_passthrough 2025-12-04T09:20:35.9991826Z * [new branch] xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 2025-12-04T09:20:35.9993509Z * [new branch] xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 2025-12-04T09:20:35.9995096Z * [new branch] xmfan/single_step -> origin/xmfan/single_step 2025-12-04T09:20:35.9996946Z * [new branch] xmfan/sth_0829 -> origin/xmfan/sth_0829 2025-12-04T09:20:35.9998720Z * [new branch] xmfan/test -> origin/xmfan/test 2025-12-04T09:20:36.0001215Z * [new branch] yguo/debug-0226-constexpr -> origin/yguo/debug-0226-constexpr 2025-12-04T09:20:36.0002969Z * [new branch] yguo/new_latest_changes -> origin/yguo/new_latest_changes 2025-12-04T09:20:36.0004708Z * [new branch] yguo/patch_constexpr_changes -> origin/yguo/patch_constexpr_changes 2025-12-04T09:20:36.0007064Z * [new branch] yiming/bootcamp -> origin/yiming/bootcamp 2025-12-04T09:20:36.0008745Z * [new branch] yiming/run_with_start_end_rng_hop -> origin/yiming/run_with_start_end_rng_hop 2025-12-04T09:20:36.0010524Z * [new branch] yolo-llama3 -> origin/yolo-llama3 2025-12-04T09:20:36.0013014Z * [new branch] zainr/canary-test -> origin/zainr/canary-test 2025-12-04T09:20:36.0014826Z * [new branch] zainr/cleanup-gh-runners -> origin/zainr/cleanup-gh-runners 2025-12-04T09:20:36.0016469Z * [new branch] zainr/pull-migration-c -> origin/zainr/pull-migration-c 2025-12-04T09:20:36.0018058Z * [new branch] zainr/test2 -> origin/zainr/test2 2025-12-04T09:20:36.0020409Z * [new branch] zasdfgbnm-patch-3 -> origin/zasdfgbnm-patch-3 2025-12-04T09:20:36.0021953Z * [new branch] zb2p -> origin/zb2p 2025-12-04T09:20:36.0023760Z * [new branch] zeros-and-scatter-part2 -> origin/zeros-and-scatter-part2 2025-12-04T09:20:36.0026689Z * [new branch] zhxchen17/ci/vllm_lora_oom -> origin/zhxchen17/ci/vllm_lora_oom 2025-12-04T09:20:36.0028455Z * [new branch] zhxchen17/ci/vllm_multimodal_oom -> origin/zhxchen17/ci/vllm_multimodal_oom 2025-12-04T09:20:36.0030049Z * [new branch] zhxchen17/ci/vllm_pin -> origin/zhxchen17/ci/vllm_pin 2025-12-04T09:20:36.0032490Z * [new branch] zhxchen17/dynamo/unsafe_drop_all_guards -> origin/zhxchen17/dynamo/unsafe_drop_all_guards 2025-12-04T09:20:36.0034692Z * [new branch] zhxchen17/export/call_override -> origin/zhxchen17/export/call_override 2025-12-04T09:20:36.0036395Z * [new branch] zhxchen17/export/codemod1 -> origin/zhxchen17/export/codemod1 2025-12-04T09:20:36.0038163Z * [new branch] zhxchen17/export/ctx_return -> origin/zhxchen17/export/ctx_return 2025-12-04T09:20:36.0040007Z * [new branch] zhxchen17/export/disable_side_effect_warn -> origin/zhxchen17/export/disable_side_effect_warn 2025-12-04T09:20:36.0041616Z * [new branch] zhxchen17/export/pytree_check -> origin/zhxchen17/export/pytree_check 2025-12-04T09:20:36.0044122Z * [new branch] zhxchen17/precompile/aoti -> origin/zhxchen17/precompile/aoti 2025-12-04T09:20:36.0045819Z * [new branch] zhxchen17/precompile/globals -> origin/zhxchen17/precompile/globals 2025-12-04T09:20:36.0047698Z * [new branch] zhxchen17/precompile/inductor_guards -> origin/zhxchen17/precompile/inductor_guards 2025-12-04T09:20:36.0049896Z * [new branch] zhxchen17/scratch/0 -> origin/zhxchen17/scratch/0 2025-12-04T09:20:36.0051698Z * [new branch] zhxchen17/torch_export_api_update -> origin/zhxchen17/torch_export_api_update 2025-12-04T09:20:36.0053945Z * [new branch] zhxhcen17/moodycamel -> origin/zhxhcen17/moodycamel 2025-12-04T09:20:36.0056378Z * [new branch] zxiiro/build-times -> origin/zxiiro/build-times 2025-12-04T09:20:36.0058070Z * [new branch] zxiiro/c7i.2xlarge -> origin/zxiiro/c7i.2xlarge 2025-12-04T09:20:36.0059926Z * [new branch] zxiiro/c7i.2xlarge.h100 -> origin/zxiiro/c7i.2xlarge.h100 2025-12-04T09:20:36.0061625Z * [new branch] zxiiro/main -> origin/zxiiro/main 2025-12-04T09:20:36.0063339Z * [new branch] zxiiro/risc64 -> origin/zxiiro/risc64 2025-12-04T09:20:36.0065168Z * [new branch] zxiiro/test-multicloud-arc -> origin/zxiiro/test-multicloud-arc 2025-12-04T09:20:36.0066704Z * [new tag] bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug -> bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug 2025-12-04T09:20:36.0068243Z * [new tag] ci/binaries/77164 -> ci/binaries/77164 2025-12-04T09:20:36.0069855Z * [new tag] ciflow/b200/115316 -> ciflow/b200/115316 2025-12-04T09:20:36.0071258Z * [new tag] ciflow/b200/160685 -> ciflow/b200/160685 2025-12-04T09:20:36.0073866Z * [new tag] ciflow/b200/161607 -> ciflow/b200/161607 2025-12-04T09:20:36.0075049Z * [new tag] ciflow/b200/161938 -> ciflow/b200/161938 2025-12-04T09:20:36.0076447Z * [new tag] ciflow/b200/167207 -> ciflow/b200/167207 2025-12-04T09:20:36.0077698Z * [new tag] ciflow/b200/167989 -> ciflow/b200/167989 2025-12-04T09:20:36.0078941Z * [new tag] ciflow/b200/168096 -> ciflow/b200/168096 2025-12-04T09:20:36.0080241Z * [new tag] ciflow/b200/168175 -> ciflow/b200/168175 2025-12-04T09:20:36.0081579Z * [new tag] ciflow/b200/168195 -> ciflow/b200/168195 2025-12-04T09:20:36.0082796Z * [new tag] ciflow/b200/169200 -> ciflow/b200/169200 2025-12-04T09:20:36.0084066Z * [new tag] ciflow/b200/169216 -> ciflow/b200/169216 2025-12-04T09:20:36.0085742Z * [new tag] ciflow/b200/169380 -> ciflow/b200/169380 2025-12-04T09:20:36.0087553Z * [new tag] ciflow/b200/169412 -> ciflow/b200/169412 2025-12-04T09:20:36.0089159Z * [new tag] ciflow/b200/169470 -> ciflow/b200/169470 2025-12-04T09:20:36.0090447Z * [new tag] ciflow/b200/169471 -> ciflow/b200/169471 2025-12-04T09:20:36.0091727Z * [new tag] ciflow/b200/169472 -> ciflow/b200/169472 2025-12-04T09:20:36.0093257Z * [new tag] ciflow/b200/169514 -> ciflow/b200/169514 2025-12-04T09:20:36.0094518Z * [new tag] ciflow/b200/169517 -> ciflow/b200/169517 2025-12-04T09:20:36.0096101Z * [new tag] ciflow/binaries/165922 -> ciflow/binaries/165922 2025-12-04T09:20:36.0097309Z * [new tag] ciflow/binaries/169510 -> ciflow/binaries/169510 2025-12-04T09:20:36.0098826Z * [new tag] ciflow/binaries_wheel/157994 -> ciflow/binaries_wheel/157994 2025-12-04T09:20:36.0100240Z * [new tag] ciflow/binaries_wheel/166829 -> ciflow/binaries_wheel/166829 2025-12-04T09:20:36.0101375Z * [new tag] ciflow/binaries_wheel/167972 -> ciflow/binaries_wheel/167972 2025-12-04T09:20:36.0102696Z * [new tag] ciflow/binaries_wheel/167981 -> ciflow/binaries_wheel/167981 2025-12-04T09:20:36.0104090Z * [new tag] ciflow/dynamo/167695 -> ciflow/dynamo/167695 2025-12-04T09:20:36.0105270Z * [new tag] ciflow/dynamo/168096 -> ciflow/dynamo/168096 2025-12-04T09:20:36.0106576Z * [new tag] ciflow/dynamo/169525 -> ciflow/dynamo/169525 2025-12-04T09:20:36.0108086Z * [new tag] ciflow/h100-cutlass-backend/161938 -> ciflow/h100-cutlass-backend/161938 2025-12-04T09:20:36.0109274Z * [new tag] ciflow/h100-cutlass-backend/161940 -> ciflow/h100-cutlass-backend/161940 2025-12-04T09:20:36.0110751Z * [new tag] ciflow/h100-distributed/168923 -> ciflow/h100-distributed/168923 2025-12-04T09:20:36.0112236Z * [new tag] ciflow/h100-symm-mem/167552 -> ciflow/h100-symm-mem/167552 2025-12-04T09:20:36.0113468Z * [new tag] ciflow/h100-symm-mem/168129 -> ciflow/h100-symm-mem/168129 2025-12-04T09:20:36.0114657Z * [new tag] ciflow/h100-symm-mem/168917 -> ciflow/h100-symm-mem/168917 2025-12-04T09:20:36.0116123Z * [new tag] ciflow/h100-symm-mem/169156 -> ciflow/h100-symm-mem/169156 2025-12-04T09:20:36.0117275Z * [new tag] ciflow/h100-symm-mem/169200 -> ciflow/h100-symm-mem/169200 2025-12-04T09:20:36.0118469Z * [new tag] ciflow/h100-symm-mem/169216 -> ciflow/h100-symm-mem/169216 2025-12-04T09:20:36.0119656Z * [new tag] ciflow/h100-symm-mem/169338 -> ciflow/h100-symm-mem/169338 2025-12-04T09:20:36.0120937Z * [new tag] ciflow/h100-symm-mem/169355 -> ciflow/h100-symm-mem/169355 2025-12-04T09:20:36.0122041Z * [new tag] ciflow/h100-symm-mem/169543 -> ciflow/h100-symm-mem/169543 2025-12-04T09:20:36.0123414Z * [new tag] ciflow/h100/115316 -> ciflow/h100/115316 2025-12-04T09:20:36.0124630Z * [new tag] ciflow/h100/160685 -> ciflow/h100/160685 2025-12-04T09:20:36.0125796Z * [new tag] ciflow/h100/160729 -> ciflow/h100/160729 2025-12-04T09:20:36.0127080Z * [new tag] ciflow/h100/161607 -> ciflow/h100/161607 2025-12-04T09:20:36.0127996Z * [new tag] ciflow/h100/161938 -> ciflow/h100/161938 2025-12-04T09:20:36.0129401Z * [new tag] ciflow/h100/167207 -> ciflow/h100/167207 2025-12-04T09:20:36.0130504Z * [new tag] ciflow/h100/167989 -> ciflow/h100/167989 2025-12-04T09:20:36.0131621Z * [new tag] ciflow/h100/168096 -> ciflow/h100/168096 2025-12-04T09:20:36.0132777Z * [new tag] ciflow/h100/168175 -> ciflow/h100/168175 2025-12-04T09:20:36.0134352Z * [new tag] ciflow/h100/168195 -> ciflow/h100/168195 2025-12-04T09:20:36.0135566Z * [new tag] ciflow/h100/168980 -> ciflow/h100/168980 2025-12-04T09:20:36.0137047Z * [new tag] ciflow/h100/169200 -> ciflow/h100/169200 2025-12-04T09:20:36.0138612Z * [new tag] ciflow/h100/169216 -> ciflow/h100/169216 2025-12-04T09:20:36.0140168Z * [new tag] ciflow/h100/169380 -> ciflow/h100/169380 2025-12-04T09:20:36.0141462Z * [new tag] ciflow/h100/169412 -> ciflow/h100/169412 2025-12-04T09:20:36.0142731Z * [new tag] ciflow/h100/169470 -> ciflow/h100/169470 2025-12-04T09:20:36.0144004Z * [new tag] ciflow/h100/169471 -> ciflow/h100/169471 2025-12-04T09:20:36.0145291Z * [new tag] ciflow/h100/169472 -> ciflow/h100/169472 2025-12-04T09:20:36.0146480Z * [new tag] ciflow/h100/169514 -> ciflow/h100/169514 2025-12-04T09:20:36.0147942Z * [new tag] ciflow/inductor-cu126/168096 -> ciflow/inductor-cu126/168096 2025-12-04T09:20:36.0149787Z * [new tag] ciflow/inductor-micro-benchmark-cpu-x86/168096 -> ciflow/inductor-micro-benchmark-cpu-x86/168096 2025-12-04T09:20:36.0151195Z * [new tag] ciflow/inductor-micro-benchmark/166165 -> ciflow/inductor-micro-benchmark/166165 2025-12-04T09:20:36.0152416Z * [new tag] ciflow/inductor-micro-benchmark/168096 -> ciflow/inductor-micro-benchmark/168096 2025-12-04T09:20:36.0153802Z * [new tag] ciflow/inductor-perf-compare/168096 -> ciflow/inductor-perf-compare/168096 2025-12-04T09:20:36.0155547Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/168073 -> ciflow/inductor-perf-test-nightly-rocm-mi300/168073 2025-12-04T09:20:36.0156714Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/168096 -> ciflow/inductor-perf-test-nightly-rocm-mi300/168096 2025-12-04T09:20:36.0158005Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/169024 -> ciflow/inductor-perf-test-nightly-rocm-mi300/169024 2025-12-04T09:20:36.0159460Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi355/169024 -> ciflow/inductor-perf-test-nightly-rocm-mi355/169024 2025-12-04T09:20:36.0160833Z * [new tag] ciflow/inductor-perf-test-nightly/168096 -> ciflow/inductor-perf-test-nightly/168096 2025-12-04T09:20:36.0162313Z * [new tag] ciflow/inductor-periodic/168096 -> ciflow/inductor-periodic/168096 2025-12-04T09:20:36.0163476Z * [new tag] ciflow/inductor-periodic/169024 -> ciflow/inductor-periodic/169024 2025-12-04T09:20:36.0164772Z * [new tag] ciflow/inductor-periodic/169425 -> ciflow/inductor-periodic/169425 2025-12-04T09:20:36.0166336Z * [new tag] ciflow/inductor-rocm-mi200/165545 -> ciflow/inductor-rocm-mi200/165545 2025-12-04T09:20:36.0167627Z * [new tag] ciflow/inductor-rocm-mi200/165997 -> ciflow/inductor-rocm-mi200/165997 2025-12-04T09:20:36.0168777Z * [new tag] ciflow/inductor-rocm-mi200/168096 -> ciflow/inductor-rocm-mi200/168096 2025-12-04T09:20:36.0170060Z * [new tag] ciflow/inductor-rocm-mi200/169063 -> ciflow/inductor-rocm-mi200/169063 2025-12-04T09:20:36.0171248Z * [new tag] ciflow/inductor-rocm-mi200/169425 -> ciflow/inductor-rocm-mi200/169425 2025-12-04T09:20:36.0172992Z * [new tag] ciflow/inductor-rocm-mi300/165545 -> ciflow/inductor-rocm-mi300/165545 2025-12-04T09:20:36.0173855Z * [new tag] ciflow/inductor-rocm-mi300/168096 -> ciflow/inductor-rocm-mi300/168096 2025-12-04T09:20:36.0175138Z * [new tag] ciflow/inductor-rocm-mi300/169063 -> ciflow/inductor-rocm-mi300/169063 2025-12-04T09:20:36.0176307Z * [new tag] ciflow/inductor-rocm-mi300/169425 -> ciflow/inductor-rocm-mi300/169425 2025-12-04T09:20:36.0177891Z * [new tag] ciflow/inductor-rocm/162052 -> ciflow/inductor-rocm/162052 2025-12-04T09:20:36.0179080Z * [new tag] ciflow/inductor-rocm/168971 -> ciflow/inductor-rocm/168971 2025-12-04T09:20:36.0180780Z * [new tag] ciflow/inductor-windows/168096 -> ciflow/inductor-windows/168096 2025-12-04T09:20:36.0182169Z * [new tag] ciflow/inductor/144542 -> ciflow/inductor/144542 2025-12-04T09:20:36.0183407Z * [new tag] ciflow/inductor/146506 -> ciflow/inductor/146506 2025-12-04T09:20:36.0184537Z * [new tag] ciflow/inductor/147990 -> ciflow/inductor/147990 2025-12-04T09:20:36.0185848Z * [new tag] ciflow/inductor/148294 -> ciflow/inductor/148294 2025-12-04T09:20:36.0186997Z * [new tag] ciflow/inductor/148492 -> ciflow/inductor/148492 2025-12-04T09:20:36.0188596Z * [new tag] ciflow/inductor/157149 -> ciflow/inductor/157149 2025-12-04T09:20:36.0189816Z * [new tag] ciflow/inductor/157994 -> ciflow/inductor/157994 2025-12-04T09:20:36.0190995Z * [new tag] ciflow/inductor/160685 -> ciflow/inductor/160685 2025-12-04T09:20:36.0192153Z * [new tag] ciflow/inductor/160686 -> ciflow/inductor/160686 2025-12-04T09:20:36.0193317Z * [new tag] ciflow/inductor/160687 -> ciflow/inductor/160687 2025-12-04T09:20:36.0194510Z * [new tag] ciflow/inductor/160688 -> ciflow/inductor/160688 2025-12-04T09:20:36.0196004Z * [new tag] ciflow/inductor/160706 -> ciflow/inductor/160706 2025-12-04T09:20:36.0197671Z * [new tag] ciflow/inductor/160729 -> ciflow/inductor/160729 2025-12-04T09:20:36.0199124Z * [new tag] ciflow/inductor/161938 -> ciflow/inductor/161938 2025-12-04T09:20:36.0200398Z * [new tag] ciflow/inductor/161939 -> ciflow/inductor/161939 2025-12-04T09:20:36.0201655Z * [new tag] ciflow/inductor/161940 -> ciflow/inductor/161940 2025-12-04T09:20:36.0203030Z * [new tag] ciflow/inductor/162052 -> ciflow/inductor/162052 2025-12-04T09:20:36.0204243Z * [new tag] ciflow/inductor/162275 -> ciflow/inductor/162275 2025-12-04T09:20:36.0205491Z * [new tag] ciflow/inductor/162795 -> ciflow/inductor/162795 2025-12-04T09:20:36.0206902Z * [new tag] ciflow/inductor/163245 -> ciflow/inductor/163245 2025-12-04T09:20:36.0208205Z * [new tag] ciflow/inductor/163335 -> ciflow/inductor/163335 2025-12-04T09:20:36.0209392Z * [new tag] ciflow/inductor/163503 -> ciflow/inductor/163503 2025-12-04T09:20:36.0210628Z * [new tag] ciflow/inductor/163942 -> ciflow/inductor/163942 2025-12-04T09:20:36.0211999Z * [new tag] ciflow/inductor/165270 -> ciflow/inductor/165270 2025-12-04T09:20:36.0213260Z * [new tag] ciflow/inductor/165274 -> ciflow/inductor/165274 2025-12-04T09:20:36.0214445Z * [new tag] ciflow/inductor/165322 -> ciflow/inductor/165322 2025-12-04T09:20:36.0215707Z * [new tag] ciflow/inductor/165597 -> ciflow/inductor/165597 2025-12-04T09:20:36.0216948Z * [new tag] ciflow/inductor/166063 -> ciflow/inductor/166063 2025-12-04T09:20:36.0218286Z * [new tag] ciflow/inductor/166075 -> ciflow/inductor/166075 2025-12-04T09:20:36.0219719Z * [new tag] ciflow/inductor/166165 -> ciflow/inductor/166165 2025-12-04T09:20:36.0221142Z * [new tag] ciflow/inductor/166254 -> ciflow/inductor/166254 2025-12-04T09:20:36.0222376Z * [new tag] ciflow/inductor/166483 -> ciflow/inductor/166483 2025-12-04T09:20:36.0223640Z * [new tag] ciflow/inductor/166494 -> ciflow/inductor/166494 2025-12-04T09:20:36.0224865Z * [new tag] ciflow/inductor/166545 -> ciflow/inductor/166545 2025-12-04T09:20:36.0226108Z * [new tag] ciflow/inductor/166788 -> ciflow/inductor/166788 2025-12-04T09:20:36.0227479Z * [new tag] ciflow/inductor/166846 -> ciflow/inductor/166846 2025-12-04T09:20:36.0228735Z * [new tag] ciflow/inductor/167300 -> ciflow/inductor/167300 2025-12-04T09:20:36.0229966Z * [new tag] ciflow/inductor/167407 -> ciflow/inductor/167407 2025-12-04T09:20:36.0231323Z * [new tag] ciflow/inductor/167536 -> ciflow/inductor/167536 2025-12-04T09:20:36.0232562Z * [new tag] ciflow/inductor/167552 -> ciflow/inductor/167552 2025-12-04T09:20:36.0233971Z * [new tag] ciflow/inductor/167555 -> ciflow/inductor/167555 2025-12-04T09:20:36.0235183Z * [new tag] ciflow/inductor/167583 -> ciflow/inductor/167583 2025-12-04T09:20:36.0236381Z * [new tag] ciflow/inductor/167599 -> ciflow/inductor/167599 2025-12-04T09:20:36.0237635Z * [new tag] ciflow/inductor/167647 -> ciflow/inductor/167647 2025-12-04T09:20:36.0238900Z * [new tag] ciflow/inductor/167677 -> ciflow/inductor/167677 2025-12-04T09:20:36.0240158Z * [new tag] ciflow/inductor/167680 -> ciflow/inductor/167680 2025-12-04T09:20:36.0241409Z * [new tag] ciflow/inductor/167695 -> ciflow/inductor/167695 2025-12-04T09:20:36.0242610Z * [new tag] ciflow/inductor/167742 -> ciflow/inductor/167742 2025-12-04T09:20:36.0243871Z * [new tag] ciflow/inductor/167768 -> ciflow/inductor/167768 2025-12-04T09:20:36.0245349Z * [new tag] ciflow/inductor/167773 -> ciflow/inductor/167773 2025-12-04T09:20:36.0246660Z * [new tag] ciflow/inductor/167781 -> ciflow/inductor/167781 2025-12-04T09:20:36.0247920Z * [new tag] ciflow/inductor/167880 -> ciflow/inductor/167880 2025-12-04T09:20:36.0249173Z * [new tag] ciflow/inductor/167887 -> ciflow/inductor/167887 2025-12-04T09:20:36.0250406Z * [new tag] ciflow/inductor/167972 -> ciflow/inductor/167972 2025-12-04T09:20:36.0251661Z * [new tag] ciflow/inductor/167989 -> ciflow/inductor/167989 2025-12-04T09:20:36.0252990Z * [new tag] ciflow/inductor/168002 -> ciflow/inductor/168002 2025-12-04T09:20:36.0254252Z * [new tag] ciflow/inductor/168050 -> ciflow/inductor/168050 2025-12-04T09:20:36.0255493Z * [new tag] ciflow/inductor/168051 -> ciflow/inductor/168051 2025-12-04T09:20:36.0256743Z * [new tag] ciflow/inductor/168052 -> ciflow/inductor/168052 2025-12-04T09:20:36.0258036Z * [new tag] ciflow/inductor/168073 -> ciflow/inductor/168073 2025-12-04T09:20:36.0259323Z * [new tag] ciflow/inductor/168096 -> ciflow/inductor/168096 2025-12-04T09:20:36.0260746Z * [new tag] ciflow/inductor/168114 -> ciflow/inductor/168114 2025-12-04T09:20:36.0261966Z * [new tag] ciflow/inductor/168115 -> ciflow/inductor/168115 2025-12-04T09:20:36.0263204Z * [new tag] ciflow/inductor/168127 -> ciflow/inductor/168127 2025-12-04T09:20:36.0264401Z * [new tag] ciflow/inductor/168129 -> ciflow/inductor/168129 2025-12-04T09:20:36.0265677Z * [new tag] ciflow/inductor/168157 -> ciflow/inductor/168157 2025-12-04T09:20:36.0267038Z * [new tag] ciflow/inductor/168175 -> ciflow/inductor/168175 2025-12-04T09:20:36.0268221Z * [new tag] ciflow/inductor/168185 -> ciflow/inductor/168185 2025-12-04T09:20:36.0269456Z * [new tag] ciflow/inductor/168195 -> ciflow/inductor/168195 2025-12-04T09:20:36.0271169Z * [new tag] ciflow/inductor/168209 -> ciflow/inductor/168209 2025-12-04T09:20:36.0273924Z * [new tag] ciflow/inductor/168266 -> ciflow/inductor/168266 2025-12-04T09:20:36.0275208Z * [new tag] ciflow/inductor/168316 -> ciflow/inductor/168316 2025-12-04T09:20:36.0276583Z * [new tag] ciflow/inductor/168326 -> ciflow/inductor/168326 2025-12-04T09:20:36.0277819Z * [new tag] ciflow/inductor/168368 -> ciflow/inductor/168368 2025-12-04T09:20:36.0279050Z * [new tag] ciflow/inductor/168894 -> ciflow/inductor/168894 2025-12-04T09:20:36.0280327Z * [new tag] ciflow/inductor/168934 -> ciflow/inductor/168934 2025-12-04T09:20:36.0281541Z * [new tag] ciflow/inductor/168939 -> ciflow/inductor/168939 2025-12-04T09:20:36.0282791Z * [new tag] ciflow/inductor/168946 -> ciflow/inductor/168946 2025-12-04T09:20:36.0284056Z * [new tag] ciflow/inductor/168950 -> ciflow/inductor/168950 2025-12-04T09:20:36.0285342Z * [new tag] ciflow/inductor/168951 -> ciflow/inductor/168951 2025-12-04T09:20:36.0286594Z * [new tag] ciflow/inductor/168952 -> ciflow/inductor/168952 2025-12-04T09:20:36.0287817Z * [new tag] ciflow/inductor/168955 -> ciflow/inductor/168955 2025-12-04T09:20:36.0289090Z * [new tag] ciflow/inductor/168971 -> ciflow/inductor/168971 2025-12-04T09:20:36.0290364Z * [new tag] ciflow/inductor/168979 -> ciflow/inductor/168979 2025-12-04T09:20:36.0291599Z * [new tag] ciflow/inductor/168980 -> ciflow/inductor/168980 2025-12-04T09:20:36.0292993Z * [new tag] ciflow/inductor/168983 -> ciflow/inductor/168983 2025-12-04T09:20:36.0294237Z * [new tag] ciflow/inductor/169006 -> ciflow/inductor/169006 2025-12-04T09:20:36.0295600Z * [new tag] ciflow/inductor/169023 -> ciflow/inductor/169023 2025-12-04T09:20:36.0296819Z * [new tag] ciflow/inductor/169024 -> ciflow/inductor/169024 2025-12-04T09:20:36.0298151Z * [new tag] ciflow/inductor/169025 -> ciflow/inductor/169025 2025-12-04T09:20:36.0299483Z * [new tag] ciflow/inductor/169066 -> ciflow/inductor/169066 2025-12-04T09:20:36.0300797Z * [new tag] ciflow/inductor/169091 -> ciflow/inductor/169091 2025-12-04T09:20:36.0302012Z * [new tag] ciflow/inductor/169102 -> ciflow/inductor/169102 2025-12-04T09:20:36.0303333Z * [new tag] ciflow/inductor/169103 -> ciflow/inductor/169103 2025-12-04T09:20:36.0304598Z * [new tag] ciflow/inductor/169121 -> ciflow/inductor/169121 2025-12-04T09:20:36.0305851Z * [new tag] ciflow/inductor/169134 -> ciflow/inductor/169134 2025-12-04T09:20:36.0307064Z * [new tag] ciflow/inductor/169135 -> ciflow/inductor/169135 2025-12-04T09:20:36.0308447Z * [new tag] ciflow/inductor/169141 -> ciflow/inductor/169141 2025-12-04T09:20:36.0309722Z * [new tag] ciflow/inductor/169151 -> ciflow/inductor/169151 2025-12-04T09:20:36.0311125Z * [new tag] ciflow/inductor/169161 -> ciflow/inductor/169161 2025-12-04T09:20:36.0312371Z * [new tag] ciflow/inductor/169167 -> ciflow/inductor/169167 2025-12-04T09:20:36.0313782Z * [new tag] ciflow/inductor/169177 -> ciflow/inductor/169177 2025-12-04T09:20:36.0315292Z * [new tag] ciflow/inductor/169185 -> ciflow/inductor/169185 2025-12-04T09:20:36.0316467Z * [new tag] ciflow/inductor/169196 -> ciflow/inductor/169196 2025-12-04T09:20:36.0317658Z * [new tag] ciflow/inductor/169200 -> ciflow/inductor/169200 2025-12-04T09:20:36.0318915Z * [new tag] ciflow/inductor/169204 -> ciflow/inductor/169204 2025-12-04T09:20:36.0320164Z * [new tag] ciflow/inductor/169216 -> ciflow/inductor/169216 2025-12-04T09:20:36.0321400Z * [new tag] ciflow/inductor/169219 -> ciflow/inductor/169219 2025-12-04T09:20:36.0322639Z * [new tag] ciflow/inductor/169220 -> ciflow/inductor/169220 2025-12-04T09:20:36.0324034Z * [new tag] ciflow/inductor/169230 -> ciflow/inductor/169230 2025-12-04T09:20:36.0325290Z * [new tag] ciflow/inductor/169242 -> ciflow/inductor/169242 2025-12-04T09:20:36.0326535Z * [new tag] ciflow/inductor/169245 -> ciflow/inductor/169245 2025-12-04T09:20:36.0327917Z * [new tag] ciflow/inductor/169260 -> ciflow/inductor/169260 2025-12-04T09:20:36.0329158Z * [new tag] ciflow/inductor/169282 -> ciflow/inductor/169282 2025-12-04T09:20:36.0330435Z * [new tag] ciflow/inductor/169286 -> ciflow/inductor/169286 2025-12-04T09:20:36.0331697Z * [new tag] ciflow/inductor/169299 -> ciflow/inductor/169299 2025-12-04T09:20:36.0333089Z * [new tag] ciflow/inductor/169304 -> ciflow/inductor/169304 2025-12-04T09:20:36.0334690Z * [new tag] ciflow/inductor/169305 -> ciflow/inductor/169305 2025-12-04T09:20:36.0335962Z * [new tag] ciflow/inductor/169308 -> ciflow/inductor/169308 2025-12-04T09:20:36.0337241Z * [new tag] ciflow/inductor/169319 -> ciflow/inductor/169319 2025-12-04T09:20:36.0338490Z * [new tag] ciflow/inductor/169326 -> ciflow/inductor/169326 2025-12-04T09:20:36.0339886Z * [new tag] ciflow/inductor/169332 -> ciflow/inductor/169332 2025-12-04T09:20:36.0341132Z * [new tag] ciflow/inductor/169333 -> ciflow/inductor/169333 2025-12-04T09:20:36.0342545Z * [new tag] ciflow/inductor/169336 -> ciflow/inductor/169336 2025-12-04T09:20:36.0343774Z * [new tag] ciflow/inductor/169340 -> ciflow/inductor/169340 2025-12-04T09:20:36.0345164Z * [new tag] ciflow/inductor/169341 -> ciflow/inductor/169341 2025-12-04T09:20:36.0346446Z * [new tag] ciflow/inductor/169343 -> ciflow/inductor/169343 2025-12-04T09:20:36.0347710Z * [new tag] ciflow/inductor/169346 -> ciflow/inductor/169346 2025-12-04T09:20:36.0349050Z * [new tag] ciflow/inductor/169348 -> ciflow/inductor/169348 2025-12-04T09:20:36.0350416Z * [new tag] ciflow/inductor/169350 -> ciflow/inductor/169350 2025-12-04T09:20:36.0351653Z * [new tag] ciflow/inductor/169355 -> ciflow/inductor/169355 2025-12-04T09:20:36.0352960Z * [new tag] ciflow/inductor/169370 -> ciflow/inductor/169370 2025-12-04T09:20:36.0354448Z * [new tag] ciflow/inductor/169375 -> ciflow/inductor/169375 2025-12-04T09:20:36.0356176Z * [new tag] ciflow/inductor/169389 -> ciflow/inductor/169389 2025-12-04T09:20:36.0357458Z * [new tag] ciflow/inductor/169391 -> ciflow/inductor/169391 2025-12-04T09:20:36.0358680Z * [new tag] ciflow/inductor/169393 -> ciflow/inductor/169393 2025-12-04T09:20:36.0359937Z * [new tag] ciflow/inductor/169399 -> ciflow/inductor/169399 2025-12-04T09:20:36.0361319Z * [new tag] ciflow/inductor/169400 -> ciflow/inductor/169400 2025-12-04T09:20:36.0362643Z * [new tag] ciflow/inductor/169415 -> ciflow/inductor/169415 2025-12-04T09:20:36.0363978Z * [new tag] ciflow/inductor/169417 -> ciflow/inductor/169417 2025-12-04T09:20:36.0365150Z * [new tag] ciflow/inductor/169418 -> ciflow/inductor/169418 2025-12-04T09:20:36.0366670Z * [new tag] ciflow/inductor/169430 -> ciflow/inductor/169430 2025-12-04T09:20:36.0367949Z * [new tag] ciflow/inductor/169432 -> ciflow/inductor/169432 2025-12-04T09:20:36.0369164Z * [new tag] ciflow/inductor/169436 -> ciflow/inductor/169436 2025-12-04T09:20:36.0370539Z * [new tag] ciflow/inductor/169437 -> ciflow/inductor/169437 2025-12-04T09:20:36.0372099Z * [new tag] ciflow/inductor/169438 -> ciflow/inductor/169438 2025-12-04T09:20:36.0373384Z * [new tag] ciflow/inductor/169441 -> ciflow/inductor/169441 2025-12-04T09:20:36.0374658Z * [new tag] ciflow/inductor/169446 -> ciflow/inductor/169446 2025-12-04T09:20:36.0376028Z * [new tag] ciflow/inductor/169447 -> ciflow/inductor/169447 2025-12-04T09:20:36.0377278Z * [new tag] ciflow/inductor/169452 -> ciflow/inductor/169452 2025-12-04T09:20:36.0378734Z * [new tag] ciflow/inductor/169455 -> ciflow/inductor/169455 2025-12-04T09:20:36.0380075Z * [new tag] ciflow/inductor/169459 -> ciflow/inductor/169459 2025-12-04T09:20:36.0381400Z * [new tag] ciflow/inductor/169463 -> ciflow/inductor/169463 2025-12-04T09:20:36.0382740Z * [new tag] ciflow/inductor/169476 -> ciflow/inductor/169476 2025-12-04T09:20:36.0384037Z * [new tag] ciflow/inductor/169485 -> ciflow/inductor/169485 2025-12-04T09:20:36.0385291Z * [new tag] ciflow/inductor/169493 -> ciflow/inductor/169493 2025-12-04T09:20:36.0386539Z * [new tag] ciflow/inductor/169496 -> ciflow/inductor/169496 2025-12-04T09:20:36.0387889Z * [new tag] ciflow/inductor/169497 -> ciflow/inductor/169497 2025-12-04T09:20:36.0389156Z * [new tag] ciflow/inductor/169503 -> ciflow/inductor/169503 2025-12-04T09:20:36.0390391Z * [new tag] ciflow/inductor/169504 -> ciflow/inductor/169504 2025-12-04T09:20:36.0391949Z * [new tag] ciflow/inductor/169505 -> ciflow/inductor/169505 2025-12-04T09:20:36.0393613Z * [new tag] ciflow/inductor/169508 -> ciflow/inductor/169508 2025-12-04T09:20:36.0394879Z * [new tag] ciflow/inductor/169509 -> ciflow/inductor/169509 2025-12-04T09:20:36.0396220Z * [new tag] ciflow/inductor/169513 -> ciflow/inductor/169513 2025-12-04T09:20:36.0397476Z * [new tag] ciflow/inductor/169514 -> ciflow/inductor/169514 2025-12-04T09:20:36.0398741Z * [new tag] ciflow/inductor/169515 -> ciflow/inductor/169515 2025-12-04T09:20:36.0400033Z * [new tag] ciflow/inductor/169517 -> ciflow/inductor/169517 2025-12-04T09:20:36.0401275Z * [new tag] ciflow/inductor/169519 -> ciflow/inductor/169519 2025-12-04T09:20:36.0402531Z * [new tag] ciflow/inductor/169520 -> ciflow/inductor/169520 2025-12-04T09:20:36.0403799Z * [new tag] ciflow/inductor/169521 -> ciflow/inductor/169521 2025-12-04T09:20:36.0405057Z * [new tag] ciflow/inductor/169524 -> ciflow/inductor/169524 2025-12-04T09:20:36.0406302Z * [new tag] ciflow/inductor/169527 -> ciflow/inductor/169527 2025-12-04T09:20:36.0407548Z * [new tag] ciflow/inductor/169528 -> ciflow/inductor/169528 2025-12-04T09:20:36.0408919Z * [new tag] ciflow/inductor/169532 -> ciflow/inductor/169532 2025-12-04T09:20:36.0410179Z * [new tag] ciflow/inductor/169535 -> ciflow/inductor/169535 2025-12-04T09:20:36.0411436Z * [new tag] ciflow/inductor/169536 -> ciflow/inductor/169536 2025-12-04T09:20:36.0412887Z * [new tag] ciflow/inductor/169547 -> ciflow/inductor/169547 2025-12-04T09:20:36.0414085Z * [new tag] ciflow/inductor/169548 -> ciflow/inductor/169548 2025-12-04T09:20:36.0415320Z * [new tag] ciflow/inductor/169549 -> ciflow/inductor/169549 2025-12-04T09:20:36.0416563Z * [new tag] ciflow/inductor/169551 -> ciflow/inductor/169551 2025-12-04T09:20:36.0417814Z * [new tag] ciflow/inductor/169552 -> ciflow/inductor/169552 2025-12-04T09:20:36.0419180Z * [new tag] ciflow/inductor/169553 -> ciflow/inductor/169553 2025-12-04T09:20:36.0420507Z * [new tag] ciflow/inductor/169557 -> ciflow/inductor/169557 2025-12-04T09:20:36.0421938Z * [new tag] ciflow/inductor/3b9a386 -> ciflow/inductor/3b9a386 2025-12-04T09:20:36.0423394Z * [new tag] ciflow/inductor/3d4b92b -> ciflow/inductor/3d4b92b 2025-12-04T09:20:36.0424821Z * [new tag] ciflow/inductor/d224ac7 -> ciflow/inductor/d224ac7 2025-12-04T09:20:36.0426268Z * [new tag] ciflow/linux-aarch64/157994 -> ciflow/linux-aarch64/157994 2025-12-04T09:20:36.0427456Z * [new tag] ciflow/linux-aarch64/166075 -> ciflow/linux-aarch64/166075 2025-12-04T09:20:36.0428722Z * [new tag] ciflow/linux-aarch64/166876 -> ciflow/linux-aarch64/166876 2025-12-04T09:20:36.0429921Z * [new tag] ciflow/linux-aarch64/167981 -> ciflow/linux-aarch64/167981 2025-12-04T09:20:36.0431317Z * [new tag] ciflow/mps/166254 -> ciflow/mps/166254 2025-12-04T09:20:36.0432905Z * [new tag] ciflow/mps/169017 -> ciflow/mps/169017 2025-12-04T09:20:36.0434223Z * [new tag] ciflow/mps/169372 -> ciflow/mps/169372 2025-12-04T09:20:36.0435462Z * [new tag] ciflow/mps/169478 -> ciflow/mps/169478 2025-12-04T09:20:36.0436912Z * [new tag] ciflow/op-benchmark/157994 -> ciflow/op-benchmark/157994 2025-12-04T09:20:36.0438063Z * [new tag] ciflow/op-benchmark/166075 -> ciflow/op-benchmark/166075 2025-12-04T09:20:36.0439269Z * [new tag] ciflow/op-benchmark/169544 -> ciflow/op-benchmark/169544 2025-12-04T09:20:36.0440802Z * [new tag] ciflow/periodic-rocm-mi200/165997 -> ciflow/periodic-rocm-mi200/165997 2025-12-04T09:20:36.0442092Z * [new tag] ciflow/periodic-rocm-mi200/166517 -> ciflow/periodic-rocm-mi200/166517 2025-12-04T09:20:36.0443268Z * [new tag] ciflow/periodic-rocm-mi200/169063 -> ciflow/periodic-rocm-mi200/169063 2025-12-04T09:20:36.0444488Z * [new tag] ciflow/periodic-rocm-mi200/169425 -> ciflow/periodic-rocm-mi200/169425 2025-12-04T09:20:36.0445947Z * [new tag] ciflow/periodic-rocm-mi300/166517 -> ciflow/periodic-rocm-mi300/166517 2025-12-04T09:20:36.0447093Z * [new tag] ciflow/periodic-rocm-mi300/169063 -> ciflow/periodic-rocm-mi300/169063 2025-12-04T09:20:36.0448269Z * [new tag] ciflow/periodic-rocm-mi300/169425 -> ciflow/periodic-rocm-mi300/169425 2025-12-04T09:20:36.0449921Z * [new tag] ciflow/periodic/054a2fd -> ciflow/periodic/054a2fd 2025-12-04T09:20:36.0451180Z * [new tag] ciflow/periodic/167207 -> ciflow/periodic/167207 2025-12-04T09:20:36.0452430Z * [new tag] ciflow/periodic/167978 -> ciflow/periodic/167978 2025-12-04T09:20:36.0453583Z * [new tag] ciflow/periodic/168096 -> ciflow/periodic/168096 2025-12-04T09:20:36.0454768Z * [new tag] ciflow/periodic/169286 -> ciflow/periodic/169286 2025-12-04T09:20:36.0456170Z * [new tag] ciflow/periodic/2a6d37d -> ciflow/periodic/2a6d37d 2025-12-04T09:20:36.0457512Z * [new tag] ciflow/periodic/317eeb8 -> ciflow/periodic/317eeb8 2025-12-04T09:20:36.0458933Z * [new tag] ciflow/periodic/3c32 -> ciflow/periodic/3c32 2025-12-04T09:20:36.0460384Z * [new tag] ciflow/periodic/3e98831 -> ciflow/periodic/3e98831 2025-12-04T09:20:36.0462380Z * [new tag] ciflow/periodic/7c648509a7470ace9fb2bae960dd4790f7e943e9 -> ciflow/periodic/7c648509a7470ace9fb2bae960dd4790f7e943e9 2025-12-04T09:20:36.0463890Z * [new tag] ciflow/periodic/94512-point -> ciflow/periodic/94512-point 2025-12-04T09:20:36.0465679Z * [new tag] ciflow/periodic/csl/test87519 -> ciflow/periodic/csl/test87519 2025-12-04T09:20:36.0467078Z * [new tag] ciflow/periodic/csltest88275 -> ciflow/periodic/csltest88275 2025-12-04T09:20:36.0468936Z * [new tag] ciflow/periodic/csltest88761 -> ciflow/periodic/csltest88761 2025-12-04T09:20:36.0470359Z * [new tag] ciflow/periodic/release_1.12 -> ciflow/periodic/release_1.12 2025-12-04T09:20:36.0472060Z * [new tag] ciflow/periodic/release_1.12.0 -> ciflow/periodic/release_1.12.0 2025-12-04T09:20:36.0473588Z * [new tag] ciflow/periodic/sha-ec5b83 -> ciflow/periodic/sha-ec5b83 2025-12-04T09:20:36.0475027Z * [new tag] ciflow/pull/167207 -> ciflow/pull/167207 2025-12-04T09:20:36.0476666Z * [new tag] ciflow/quantization-periodic/169207 -> ciflow/quantization-periodic/169207 2025-12-04T09:20:36.0478065Z * [new tag] ciflow/rocm-mi200/165545 -> ciflow/rocm-mi200/165545 2025-12-04T09:20:36.0479243Z * [new tag] ciflow/rocm-mi200/165997 -> ciflow/rocm-mi200/165997 2025-12-04T09:20:36.0480428Z * [new tag] ciflow/rocm-mi200/168096 -> ciflow/rocm-mi200/168096 2025-12-04T09:20:36.0481704Z * [new tag] ciflow/rocm-mi200/168275 -> ciflow/rocm-mi200/168275 2025-12-04T09:20:36.0482883Z * [new tag] ciflow/rocm-mi200/169063 -> ciflow/rocm-mi200/169063 2025-12-04T09:20:36.0484259Z * [new tag] ciflow/rocm-mi200/169356 -> ciflow/rocm-mi200/169356 2025-12-04T09:20:36.0485369Z * [new tag] ciflow/rocm-mi200/169425 -> ciflow/rocm-mi200/169425 2025-12-04T09:20:36.0486820Z * [new tag] ciflow/rocm-mi300/165545 -> ciflow/rocm-mi300/165545 2025-12-04T09:20:36.0488196Z * [new tag] ciflow/rocm-mi300/167157 -> ciflow/rocm-mi300/167157 2025-12-04T09:20:36.0489415Z * [new tag] ciflow/rocm-mi300/168096 -> ciflow/rocm-mi300/168096 2025-12-04T09:20:36.0490558Z * [new tag] ciflow/rocm-mi300/169063 -> ciflow/rocm-mi300/169063 2025-12-04T09:20:36.0491735Z * [new tag] ciflow/rocm-mi300/169425 -> ciflow/rocm-mi300/169425 2025-12-04T09:20:36.0493168Z * [new tag] ciflow/rocm-mi355/167157 -> ciflow/rocm-mi355/167157 2025-12-04T09:20:36.0494522Z * [new tag] ciflow/rocm-mi355/168275 -> ciflow/rocm-mi355/168275 2025-12-04T09:20:36.0495653Z * [new tag] ciflow/rocm-mi355/169425 -> ciflow/rocm-mi355/169425 2025-12-04T09:20:36.0497085Z * [new tag] ciflow/rocm-navi31/168275 -> ciflow/rocm-navi31/168275 2025-12-04T09:20:36.0498231Z * [new tag] ciflow/rocm-navi31/169425 -> ciflow/rocm-navi31/169425 2025-12-04T09:20:36.0499887Z * [new tag] ciflow/rocm/115316 -> ciflow/rocm/115316 2025-12-04T09:20:36.0501010Z * [new tag] ciflow/rocm/148492 -> ciflow/rocm/148492 2025-12-04T09:20:36.0502164Z * [new tag] ciflow/rocm/160685 -> ciflow/rocm/160685 2025-12-04T09:20:36.0503326Z * [new tag] ciflow/rocm/161607 -> ciflow/rocm/161607 2025-12-04T09:20:36.0504548Z * [new tag] ciflow/rocm/162052 -> ciflow/rocm/162052 2025-12-04T09:20:36.0505692Z * [new tag] ciflow/rocm/165997 -> ciflow/rocm/165997 2025-12-04T09:20:36.0506970Z * [new tag] ciflow/rocm/166165 -> ciflow/rocm/166165 2025-12-04T09:20:36.0507991Z * [new tag] ciflow/rocm/166517 -> ciflow/rocm/166517 2025-12-04T09:20:36.0509243Z * [new tag] ciflow/rocm/167207 -> ciflow/rocm/167207 2025-12-04T09:20:36.0510387Z * [new tag] ciflow/rocm/167536 -> ciflow/rocm/167536 2025-12-04T09:20:36.0511575Z * [new tag] ciflow/rocm/167781 -> ciflow/rocm/167781 2025-12-04T09:20:36.0513122Z * [new tag] ciflow/rocm/167989 -> ciflow/rocm/167989 2025-12-04T09:20:36.0514703Z * [new tag] ciflow/rocm/168073 -> ciflow/rocm/168073 2025-12-04T09:20:36.0516171Z * [new tag] ciflow/rocm/168195 -> ciflow/rocm/168195 2025-12-04T09:20:36.0517380Z * [new tag] ciflow/rocm/168939 -> ciflow/rocm/168939 2025-12-04T09:20:36.0518612Z * [new tag] ciflow/rocm/168971 -> ciflow/rocm/168971 2025-12-04T09:20:36.0519913Z * [new tag] ciflow/rocm/169024 -> ciflow/rocm/169024 2025-12-04T09:20:36.0521071Z * [new tag] ciflow/rocm/169200 -> ciflow/rocm/169200 2025-12-04T09:20:36.0522275Z * [new tag] ciflow/rocm/169216 -> ciflow/rocm/169216 2025-12-04T09:20:36.0523547Z * [new tag] ciflow/rocm/169312 -> ciflow/rocm/169312 2025-12-04T09:20:36.0524833Z * [new tag] ciflow/rocm/169380 -> ciflow/rocm/169380 2025-12-04T09:20:36.0526024Z * [new tag] ciflow/rocm/169427 -> ciflow/rocm/169427 2025-12-04T09:20:36.0527268Z * [new tag] ciflow/rocm/169455 -> ciflow/rocm/169455 2025-12-04T09:20:36.0528510Z * [new tag] ciflow/rocm/169470 -> ciflow/rocm/169470 2025-12-04T09:20:36.0529839Z * [new tag] ciflow/rocm/169471 -> ciflow/rocm/169471 2025-12-04T09:20:36.0531090Z * [new tag] ciflow/rocm/169472 -> ciflow/rocm/169472 2025-12-04T09:20:36.0532322Z * [new tag] ciflow/rocm/169514 -> ciflow/rocm/169514 2025-12-04T09:20:36.0533902Z * [new tag] ciflow/slow/01c7106 -> ciflow/slow/01c7106 2025-12-04T09:20:36.0535218Z * [new tag] ciflow/slow/0577043 -> ciflow/slow/0577043 2025-12-04T09:20:36.0536882Z * [new tag] ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym -> ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym 2025-12-04T09:20:36.0538083Z * [new tag] ciflow/slow/0e81104 -> ciflow/slow/0e81104 2025-12-04T09:20:36.0539347Z * [new tag] ciflow/slow/167207 -> ciflow/slow/167207 2025-12-04T09:20:36.0540557Z * [new tag] ciflow/slow/168050 -> ciflow/slow/168050 2025-12-04T09:20:36.0541893Z * [new tag] ciflow/slow/1732077 -> ciflow/slow/1732077 2025-12-04T09:20:36.0543332Z * [new tag] ciflow/slow/187eb7c -> ciflow/slow/187eb7c 2025-12-04T09:20:36.0544975Z * [new tag] ciflow/slow/1faef89 -> ciflow/slow/1faef89 2025-12-04T09:20:36.0546669Z * [new tag] ciflow/slow/3920ec1 -> ciflow/slow/3920ec1 2025-12-04T09:20:36.0548216Z * [new tag] ciflow/slow/3b7c6b2 -> ciflow/slow/3b7c6b2 2025-12-04T09:20:36.0549715Z * [new tag] ciflow/slow/59a3759 -> ciflow/slow/59a3759 2025-12-04T09:20:36.0551093Z * [new tag] ciflow/slow/70ef0bb -> ciflow/slow/70ef0bb 2025-12-04T09:20:36.0552465Z * [new tag] ciflow/slow/788ff06 -> ciflow/slow/788ff06 2025-12-04T09:20:36.0554164Z * [new tag] ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym -> ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym 2025-12-04T09:20:36.0555320Z * [new tag] ciflow/slow/9d85864 -> ciflow/slow/9d85864 2025-12-04T09:20:36.0556793Z * [new tag] ciflow/slow/9ffad5b -> ciflow/slow/9ffad5b 2025-12-04T09:20:36.0558111Z * [new tag] ciflow/slow/a206e8b -> ciflow/slow/a206e8b 2025-12-04T09:20:36.0559487Z * [new tag] ciflow/slow/a837609 -> ciflow/slow/a837609 2025-12-04T09:20:36.0560944Z * [new tag] ciflow/slow/af841f3 -> ciflow/slow/af841f3 2025-12-04T09:20:36.0562759Z * [new tag] ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym -> ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym 2025-12-04T09:20:36.0564060Z * [new tag] ciflow/torchbench/168175 -> ciflow/torchbench/168175 2025-12-04T09:20:36.0565490Z * [new tag] ciflow/trunk/148492 -> ciflow/trunk/148492 2025-12-04T09:20:36.0566649Z * [new tag] ciflow/trunk/157149 -> ciflow/trunk/157149 2025-12-04T09:20:36.0567860Z * [new tag] ciflow/trunk/157994 -> ciflow/trunk/157994 2025-12-04T09:20:36.0569029Z * [new tag] ciflow/trunk/159718 -> ciflow/trunk/159718 2025-12-04T09:20:36.0570189Z * [new tag] ciflow/trunk/160685 -> ciflow/trunk/160685 2025-12-04T09:20:36.0571414Z * [new tag] ciflow/trunk/160729 -> ciflow/trunk/160729 2025-12-04T09:20:36.0572817Z * [new tag] ciflow/trunk/162275 -> ciflow/trunk/162275 2025-12-04T09:20:36.0573969Z * [new tag] ciflow/trunk/162795 -> ciflow/trunk/162795 2025-12-04T09:20:36.0575139Z * [new tag] ciflow/trunk/163245 -> ciflow/trunk/163245 2025-12-04T09:20:36.0576320Z * [new tag] ciflow/trunk/163942 -> ciflow/trunk/163942 2025-12-04T09:20:36.0577497Z * [new tag] ciflow/trunk/165274 -> ciflow/trunk/165274 2025-12-04T09:20:36.0579336Z * [new tag] ciflow/trunk/165483 -> ciflow/trunk/165483 2025-12-04T09:20:36.0580990Z * [new tag] ciflow/trunk/165728 -> ciflow/trunk/165728 2025-12-04T09:20:36.0582486Z * [new tag] ciflow/trunk/165922 -> ciflow/trunk/165922 2025-12-04T09:20:36.0583724Z * [new tag] ciflow/trunk/166075 -> ciflow/trunk/166075 2025-12-04T09:20:36.0585061Z * [new tag] ciflow/trunk/166165 -> ciflow/trunk/166165 2025-12-04T09:20:36.0586240Z * [new tag] ciflow/trunk/166829 -> ciflow/trunk/166829 2025-12-04T09:20:36.0587637Z * [new tag] ciflow/trunk/166843 -> ciflow/trunk/166843 2025-12-04T09:20:36.0588885Z * [new tag] ciflow/trunk/166876 -> ciflow/trunk/166876 2025-12-04T09:20:36.0590148Z * [new tag] ciflow/trunk/167207 -> ciflow/trunk/167207 2025-12-04T09:20:36.0591391Z * [new tag] ciflow/trunk/167536 -> ciflow/trunk/167536 2025-12-04T09:20:36.0592616Z * [new tag] ciflow/trunk/167552 -> ciflow/trunk/167552 2025-12-04T09:20:36.0593869Z * [new tag] ciflow/trunk/167555 -> ciflow/trunk/167555 2025-12-04T09:20:36.0595164Z * [new tag] ciflow/trunk/167599 -> ciflow/trunk/167599 2025-12-04T09:20:36.0596421Z * [new tag] ciflow/trunk/167659 -> ciflow/trunk/167659 2025-12-04T09:20:36.0597822Z * [new tag] ciflow/trunk/167672 -> ciflow/trunk/167672 2025-12-04T09:20:36.0599105Z * [new tag] ciflow/trunk/167742 -> ciflow/trunk/167742 2025-12-04T09:20:36.0600339Z * [new tag] ciflow/trunk/167781 -> ciflow/trunk/167781 2025-12-04T09:20:36.0601712Z * [new tag] ciflow/trunk/167837 -> ciflow/trunk/167837 2025-12-04T09:20:36.0603017Z * [new tag] ciflow/trunk/167887 -> ciflow/trunk/167887 2025-12-04T09:20:36.0604317Z * [new tag] ciflow/trunk/167978 -> ciflow/trunk/167978 2025-12-04T09:20:36.0605668Z * [new tag] ciflow/trunk/168050 -> ciflow/trunk/168050 2025-12-04T09:20:36.0606908Z * [new tag] ciflow/trunk/168051 -> ciflow/trunk/168051 2025-12-04T09:20:36.0608116Z * [new tag] ciflow/trunk/168096 -> ciflow/trunk/168096 2025-12-04T09:20:36.0609395Z * [new tag] ciflow/trunk/168127 -> ciflow/trunk/168127 2025-12-04T09:20:36.0610611Z * [new tag] ciflow/trunk/168157 -> ciflow/trunk/168157 2025-12-04T09:20:36.0611852Z * [new tag] ciflow/trunk/168175 -> ciflow/trunk/168175 2025-12-04T09:20:36.0613109Z * [new tag] ciflow/trunk/168209 -> ciflow/trunk/168209 2025-12-04T09:20:36.0614502Z * [new tag] ciflow/trunk/168213 -> ciflow/trunk/168213 2025-12-04T09:20:36.0615882Z * [new tag] ciflow/trunk/168226 -> ciflow/trunk/168226 2025-12-04T09:20:36.0617141Z * [new tag] ciflow/trunk/168262 -> ciflow/trunk/168262 2025-12-04T09:20:36.0618384Z * [new tag] ciflow/trunk/168275 -> ciflow/trunk/168275 2025-12-04T09:20:36.0619908Z * [new tag] ciflow/trunk/168328 -> ciflow/trunk/168328 2025-12-04T09:20:36.0621174Z * [new tag] ciflow/trunk/168368 -> ciflow/trunk/168368 2025-12-04T09:20:36.0622389Z * [new tag] ciflow/trunk/168917 -> ciflow/trunk/168917 2025-12-04T09:20:36.0623655Z * [new tag] ciflow/trunk/168933 -> ciflow/trunk/168933 2025-12-04T09:20:36.0625105Z * [new tag] ciflow/trunk/168941 -> ciflow/trunk/168941 2025-12-04T09:20:36.0626317Z * [new tag] ciflow/trunk/168955 -> ciflow/trunk/168955 2025-12-04T09:20:36.0627557Z * [new tag] ciflow/trunk/168980 -> ciflow/trunk/168980 2025-12-04T09:20:36.0628978Z * [new tag] ciflow/trunk/169004 -> ciflow/trunk/169004 2025-12-04T09:20:36.0630281Z * [new tag] ciflow/trunk/169006 -> ciflow/trunk/169006 2025-12-04T09:20:36.0631459Z * [new tag] ciflow/trunk/169023 -> ciflow/trunk/169023 2025-12-04T09:20:36.0632687Z * [new tag] ciflow/trunk/169025 -> ciflow/trunk/169025 2025-12-04T09:20:36.0633931Z * [new tag] ciflow/trunk/169048 -> ciflow/trunk/169048 2025-12-04T09:20:36.0635204Z * [new tag] ciflow/trunk/169066 -> ciflow/trunk/169066 2025-12-04T09:20:36.0636430Z * [new tag] ciflow/trunk/169091 -> ciflow/trunk/169091 2025-12-04T09:20:36.0637663Z * [new tag] ciflow/trunk/169102 -> ciflow/trunk/169102 2025-12-04T09:20:36.0638895Z * [new tag] ciflow/trunk/169103 -> ciflow/trunk/169103 2025-12-04T09:20:36.0640275Z * [new tag] ciflow/trunk/169125 -> ciflow/trunk/169125 2025-12-04T09:20:36.0642089Z * [new tag] ciflow/trunk/169139 -> ciflow/trunk/169139 2025-12-04T09:20:36.0643452Z * [new tag] ciflow/trunk/169148 -> ciflow/trunk/169148 2025-12-04T09:20:36.0644715Z * [new tag] ciflow/trunk/169151 -> ciflow/trunk/169151 2025-12-04T09:20:36.0646039Z * [new tag] ciflow/trunk/169156 -> ciflow/trunk/169156 2025-12-04T09:20:36.0647394Z * [new tag] ciflow/trunk/169176 -> ciflow/trunk/169176 2025-12-04T09:20:36.0648648Z * [new tag] ciflow/trunk/169204 -> ciflow/trunk/169204 2025-12-04T09:20:36.0649930Z * [new tag] ciflow/trunk/169207 -> ciflow/trunk/169207 2025-12-04T09:20:36.0651189Z * [new tag] ciflow/trunk/169211 -> ciflow/trunk/169211 2025-12-04T09:20:36.0652655Z * [new tag] ciflow/trunk/169231 -> ciflow/trunk/169231 2025-12-04T09:20:36.0653997Z * [new tag] ciflow/trunk/169260 -> ciflow/trunk/169260 2025-12-04T09:20:36.0655397Z * [new tag] ciflow/trunk/169271 -> ciflow/trunk/169271 2025-12-04T09:20:36.0656683Z * [new tag] ciflow/trunk/169280 -> ciflow/trunk/169280 2025-12-04T09:20:36.0657912Z * [new tag] ciflow/trunk/169281 -> ciflow/trunk/169281 2025-12-04T09:20:36.0659278Z * [new tag] ciflow/trunk/169286 -> ciflow/trunk/169286 2025-12-04T09:20:36.0660723Z * [new tag] ciflow/trunk/169293 -> ciflow/trunk/169293 2025-12-04T09:20:36.0662253Z * [new tag] ciflow/trunk/169296 -> ciflow/trunk/169296 2025-12-04T09:20:36.0663580Z * [new tag] ciflow/trunk/169304 -> ciflow/trunk/169304 2025-12-04T09:20:36.0664786Z * [new tag] ciflow/trunk/169305 -> ciflow/trunk/169305 2025-12-04T09:20:36.0666026Z * [new tag] ciflow/trunk/169312 -> ciflow/trunk/169312 2025-12-04T09:20:36.0667540Z * [new tag] ciflow/trunk/169328 -> ciflow/trunk/169328 2025-12-04T09:20:36.0668789Z * [new tag] ciflow/trunk/169343 -> ciflow/trunk/169343 2025-12-04T09:20:36.0670040Z * [new tag] ciflow/trunk/169355 -> ciflow/trunk/169355 2025-12-04T09:20:36.0671453Z * [new tag] ciflow/trunk/169370 -> ciflow/trunk/169370 2025-12-04T09:20:36.0674533Z * [new tag] ciflow/trunk/169379 -> ciflow/trunk/169379 2025-12-04T09:20:36.0675761Z * [new tag] ciflow/trunk/169380 -> ciflow/trunk/169380 2025-12-04T09:20:36.0677028Z * [new tag] ciflow/trunk/169385 -> ciflow/trunk/169385 2025-12-04T09:20:36.0678246Z * [new tag] ciflow/trunk/169387 -> ciflow/trunk/169387 2025-12-04T09:20:36.0679623Z * [new tag] ciflow/trunk/169410 -> ciflow/trunk/169410 2025-12-04T09:20:36.0680881Z * [new tag] ciflow/trunk/169412 -> ciflow/trunk/169412 2025-12-04T09:20:36.0682202Z * [new tag] ciflow/trunk/169418 -> ciflow/trunk/169418 2025-12-04T09:20:36.0683457Z * [new tag] ciflow/trunk/169423 -> ciflow/trunk/169423 2025-12-04T09:20:36.0684700Z * [new tag] ciflow/trunk/169427 -> ciflow/trunk/169427 2025-12-04T09:20:36.0685972Z * [new tag] ciflow/trunk/169430 -> ciflow/trunk/169430 2025-12-04T09:20:36.0687219Z * [new tag] ciflow/trunk/169437 -> ciflow/trunk/169437 2025-12-04T09:20:36.0688460Z * [new tag] ciflow/trunk/169442 -> ciflow/trunk/169442 2025-12-04T09:20:36.0689706Z * [new tag] ciflow/trunk/169452 -> ciflow/trunk/169452 2025-12-04T09:20:36.0690964Z * [new tag] ciflow/trunk/169454 -> ciflow/trunk/169454 2025-12-04T09:20:36.0692185Z * [new tag] ciflow/trunk/169459 -> ciflow/trunk/169459 2025-12-04T09:20:36.0693551Z * [new tag] ciflow/trunk/169474 -> ciflow/trunk/169474 2025-12-04T09:20:36.0694794Z * [new tag] ciflow/trunk/169475 -> ciflow/trunk/169475 2025-12-04T09:20:36.0696026Z * [new tag] ciflow/trunk/169476 -> ciflow/trunk/169476 2025-12-04T09:20:36.0697452Z * [new tag] ciflow/trunk/169487 -> ciflow/trunk/169487 2025-12-04T09:20:36.0698701Z * [new tag] ciflow/trunk/169497 -> ciflow/trunk/169497 2025-12-04T09:20:36.0700064Z * [new tag] ciflow/trunk/169503 -> ciflow/trunk/169503 2025-12-04T09:20:36.0701274Z * [new tag] ciflow/trunk/169505 -> ciflow/trunk/169505 2025-12-04T09:20:36.0702435Z * [new tag] ciflow/trunk/169507 -> ciflow/trunk/169507 2025-12-04T09:20:36.0703678Z * [new tag] ciflow/trunk/169514 -> ciflow/trunk/169514 2025-12-04T09:20:36.0705047Z * [new tag] ciflow/trunk/169517 -> ciflow/trunk/169517 2025-12-04T09:20:36.0706240Z * [new tag] ciflow/trunk/169519 -> ciflow/trunk/169519 2025-12-04T09:20:36.0707433Z * [new tag] ciflow/trunk/169528 -> ciflow/trunk/169528 2025-12-04T09:20:36.0708666Z * [new tag] ciflow/trunk/169541 -> ciflow/trunk/169541 2025-12-04T09:20:36.0710073Z * [new tag] ciflow/trunk/169555 -> ciflow/trunk/169555 2025-12-04T09:20:36.0711828Z * [new tag] ciflow/unstable/123 -> ciflow/unstable/123 2025-12-04T09:20:36.0713302Z * [new tag] ciflow/vllm/165270 -> ciflow/vllm/165270 2025-12-04T09:20:36.0714555Z * [new tag] ciflow/vllm/165274 -> ciflow/vllm/165274 2025-12-04T09:20:36.0715708Z * [new tag] ciflow/vllm/166494 -> ciflow/vllm/166494 2025-12-04T09:20:36.0716890Z * [new tag] ciflow/vllm/169219 -> ciflow/vllm/169219 2025-12-04T09:20:36.0718056Z * [new tag] ciflow/vllm/169220 -> ciflow/vllm/169220 2025-12-04T09:20:36.0719485Z * [new tag] ciflow/xpu/157994 -> ciflow/xpu/157994 2025-12-04T09:20:36.0720630Z * [new tag] ciflow/xpu/159718 -> ciflow/xpu/159718 2025-12-04T09:20:36.0721811Z * [new tag] ciflow/xpu/161940 -> ciflow/xpu/161940 2025-12-04T09:20:36.0723078Z * [new tag] ciflow/xpu/163251 -> ciflow/xpu/163251 2025-12-04T09:20:36.0724301Z * [new tag] ciflow/xpu/166829 -> ciflow/xpu/166829 2025-12-04T09:20:36.0725402Z * [new tag] ciflow/xpu/166843 -> ciflow/xpu/166843 2025-12-04T09:20:36.0726554Z * [new tag] ciflow/xpu/167972 -> ciflow/xpu/167972 2025-12-04T09:20:36.0727699Z * [new tag] ciflow/xpu/167981 -> ciflow/xpu/167981 2025-12-04T09:20:36.0728892Z * [new tag] ciflow/xpu/168213 -> ciflow/xpu/168213 2025-12-04T09:20:36.0730038Z * [new tag] ciflow/xpu/168262 -> ciflow/xpu/168262 2025-12-04T09:20:36.0731180Z * [new tag] ciflow/xpu/168328 -> ciflow/xpu/168328 2025-12-04T09:20:36.0732684Z * [new tag] ciflow/xpu/168950 -> ciflow/xpu/168950 2025-12-04T09:20:36.0734414Z * [new tag] ciflow/xpu/169039 -> ciflow/xpu/169039 2025-12-04T09:20:36.0735825Z * [new tag] ciflow/xpu/169200 -> ciflow/xpu/169200 2025-12-04T09:20:36.0737084Z * [new tag] ciflow/xpu/169203 -> ciflow/xpu/169203 2025-12-04T09:20:36.0738302Z * [new tag] ciflow/xpu/169230 -> ciflow/xpu/169230 2025-12-04T09:20:36.0739661Z * [new tag] ciflow/xpu/169231 -> ciflow/xpu/169231 2025-12-04T09:20:36.0740991Z * [new tag] ciflow/xpu/169241 -> ciflow/xpu/169241 2025-12-04T09:20:36.0742268Z * [new tag] ciflow/xpu/169280 -> ciflow/xpu/169280 2025-12-04T09:20:36.0743622Z * [new tag] ciflow/xpu/169296 -> ciflow/xpu/169296 2025-12-04T09:20:36.0745040Z * [new tag] ciflow/xpu/169353 -> ciflow/xpu/169353 2025-12-04T09:20:36.0746223Z * [new tag] ciflow/xpu/169410 -> ciflow/xpu/169410 2025-12-04T09:20:36.0747449Z * [new tag] ciflow/xpu/169442 -> ciflow/xpu/169442 2025-12-04T09:20:36.0748720Z * [new tag] ciflow/xpu/169555 -> ciflow/xpu/169555 2025-12-04T09:20:36.0750034Z * [new tag] cslpull75 -> cslpull75 2025-12-04T09:20:36.0751254Z * [new tag] cslpull76 -> cslpull76 2025-12-04T09:20:36.0752394Z * [new tag] cslpull77 -> cslpull77 2025-12-04T09:20:36.0753725Z * [new tag] cslpull78 -> cslpull78 2025-12-04T09:20:36.0755198Z * [new tag] cslpull79 -> cslpull79 2025-12-04T09:20:36.0756706Z * [new tag] cslpull80 -> cslpull80 2025-12-04T09:20:36.0758013Z * [new tag] cslpull81 -> cslpull81 2025-12-04T09:20:36.0759376Z * [new tag] cslpull82 -> cslpull82 2025-12-04T09:20:36.0760854Z * [new tag] cslpull83 -> cslpull83 2025-12-04T09:20:36.0762153Z * [new tag] cslpull84 -> cslpull84 2025-12-04T09:20:36.0763480Z * [new tag] cslpull85 -> cslpull85 2025-12-04T09:20:36.0764836Z * [new tag] cslpull86 -> cslpull86 2025-12-04T09:20:36.0766164Z * [new tag] cslpull87 -> cslpull87 2025-12-04T09:20:36.0767425Z * [new tag] cslpull88 -> cslpull88 2025-12-04T09:20:36.0769202Z * [new tag] cslpull89 -> cslpull89 2025-12-04T09:20:36.0770373Z * [new tag] cslpull90 -> cslpull90 2025-12-04T09:20:36.0772218Z * [new tag] cslpull91 -> cslpull91 2025-12-04T09:20:36.0773394Z * [new tag] cslpull92 -> cslpull92 2025-12-04T09:20:36.0774757Z * [new tag] flight_5 -> flight_5 2025-12-04T09:20:36.0776237Z * [new tag] flight_5.1 -> flight_5.1 2025-12-04T09:20:36.0777590Z * [new tag] flight_5.2 -> flight_5.2 2025-12-04T09:20:36.0778922Z * [new tag] flight_5.3 -> flight_5.3 2025-12-04T09:20:36.0780420Z * [new tag] forpull1 -> forpull1 2025-12-04T09:20:36.0781958Z * [new tag] malfet/tag-2ef5611 -> malfet/tag-2ef5611 2025-12-04T09:20:36.0783302Z * [new tag] malfet/tag-317b1a0 -> malfet/tag-317b1a0 2025-12-04T09:20:36.0784607Z * [new tag] malfet/tag-ec6f767 -> malfet/tag-ec6f767 2025-12-04T09:20:36.0786081Z * [new tag] nightly-binary -> nightly-binary 2025-12-04T09:20:36.0787404Z * [new tag] sqzhang_flight4_plus -> sqzhang_flight4_plus 2025-12-04T09:20:36.0788857Z * [new tag] sqzhang_flight_3 -> sqzhang_flight_3 2025-12-04T09:20:36.0790535Z * [new tag] trunk/02d8bd6974cf84b721680d773dbdb1b6f40ce272 -> trunk/02d8bd6974cf84b721680d773dbdb1b6f40ce272 2025-12-04T09:20:36.0791884Z * [new tag] trunk/066997fb38ade71e00d78e9d572e380b5f02bd3e -> trunk/066997fb38ade71e00d78e9d572e380b5f02bd3e 2025-12-04T09:20:36.0793506Z * [new tag] trunk/076e7b19fa1d481ad778d06d2b49ba57d3ce8c88 -> trunk/076e7b19fa1d481ad778d06d2b49ba57d3ce8c88 2025-12-04T09:20:36.0795046Z * [new tag] trunk/07dcc0b83db3211653a38565a24e15acdba75654 -> trunk/07dcc0b83db3211653a38565a24e15acdba75654 2025-12-04T09:20:36.0796427Z * [new tag] trunk/082e96b68dfcd16cab7cfafc4d3d055767dab3eb -> trunk/082e96b68dfcd16cab7cfafc4d3d055767dab3eb 2025-12-04T09:20:36.0797746Z * [new tag] trunk/088048f2fea28ff7d450f65c72419ca45780d30b -> trunk/088048f2fea28ff7d450f65c72419ca45780d30b 2025-12-04T09:20:36.0799152Z * [new tag] trunk/09076941a95c76f4d9ad189d064dfd8baa39e672 -> trunk/09076941a95c76f4d9ad189d064dfd8baa39e672 2025-12-04T09:20:36.0800475Z * [new tag] trunk/0b80a4c62b94402844bf221791c096b0035c6d75 -> trunk/0b80a4c62b94402844bf221791c096b0035c6d75 2025-12-04T09:20:36.0802033Z * [new tag] trunk/0bbbdf1750567a980634ad907a325357ba8ba8f2 -> trunk/0bbbdf1750567a980634ad907a325357ba8ba8f2 2025-12-04T09:20:36.0803555Z * [new tag] trunk/0c281dd78773b2bc17c58ead0e4cd4ac46e775c5 -> trunk/0c281dd78773b2bc17c58ead0e4cd4ac46e775c5 2025-12-04T09:20:36.0804806Z * [new tag] trunk/135f3753c418a6879b1954904184937b67e61688 -> trunk/135f3753c418a6879b1954904184937b67e61688 2025-12-04T09:20:36.0806240Z * [new tag] trunk/15da21026cb13cd20257dc9e96830db108743c10 -> trunk/15da21026cb13cd20257dc9e96830db108743c10 2025-12-04T09:20:36.0808301Z * [new tag] trunk/166efdad2ac827f30fb02504c6017520257f88ec -> trunk/166efdad2ac827f30fb02504c6017520257f88ec 2025-12-04T09:20:36.0809083Z * [new tag] trunk/174272c15fae553d8488140af931f7d8050a313f -> trunk/174272c15fae553d8488140af931f7d8050a313f 2025-12-04T09:20:36.0810216Z * [new tag] trunk/18f3ca08f13b8de61307f5e8cd7d4cccb67e9d11 -> trunk/18f3ca08f13b8de61307f5e8cd7d4cccb67e9d11 2025-12-04T09:20:36.0811749Z * [new tag] trunk/1902eddfe655a15ebcf2c72bd81ade110fdeef63 -> trunk/1902eddfe655a15ebcf2c72bd81ade110fdeef63 2025-12-04T09:20:36.0813033Z * [new tag] trunk/195f92e98d3d66738577f11f22c4b5c8a1c76dd5 -> trunk/195f92e98d3d66738577f11f22c4b5c8a1c76dd5 2025-12-04T09:20:36.0814405Z * [new tag] trunk/1aa13e17de39e3c768ea7aebaad166ce72a06676 -> trunk/1aa13e17de39e3c768ea7aebaad166ce72a06676 2025-12-04T09:20:36.0815753Z * [new tag] trunk/1afe2832f58e24e54a5bfda5a5afa9b96fdea40e -> trunk/1afe2832f58e24e54a5bfda5a5afa9b96fdea40e 2025-12-04T09:20:36.0817155Z * [new tag] trunk/1c87554d74140eaee964ca8b1832cede67f5f520 -> trunk/1c87554d74140eaee964ca8b1832cede67f5f520 2025-12-04T09:20:36.0818479Z * [new tag] trunk/1ccb743b7b5be955f49736c162c4f5004b8a0dd8 -> trunk/1ccb743b7b5be955f49736c162c4f5004b8a0dd8 2025-12-04T09:20:36.0820003Z * [new tag] trunk/1cee47d6ce0a02227185b566593f002dd639ca0c -> trunk/1cee47d6ce0a02227185b566593f002dd639ca0c 2025-12-04T09:20:36.0821247Z * [new tag] trunk/1d21b4df2babe322e5d085ceb6de884eb260a62d -> trunk/1d21b4df2babe322e5d085ceb6de884eb260a62d 2025-12-04T09:20:36.0822641Z * [new tag] trunk/1e34fb2550e4aa650314f7a6d9f6daf4da7478a8 -> trunk/1e34fb2550e4aa650314f7a6d9f6daf4da7478a8 2025-12-04T09:20:36.0824047Z * [new tag] trunk/1e526fb5b1d93bfc70691c5c3955fdffc1b7b7de -> trunk/1e526fb5b1d93bfc70691c5c3955fdffc1b7b7de 2025-12-04T09:20:36.0825372Z * [new tag] trunk/1ee32a8b1f554a312d79bad01ded24f38cd95543 -> trunk/1ee32a8b1f554a312d79bad01ded24f38cd95543 2025-12-04T09:20:36.0826738Z * [new tag] trunk/201e2c4117eb9744594dad6a5c18213d7b4705d7 -> trunk/201e2c4117eb9744594dad6a5c18213d7b4705d7 2025-12-04T09:20:36.0828148Z * [new tag] trunk/2353a0f60eb4b4cb6675907a7fa9fbedc1c02e7f -> trunk/2353a0f60eb4b4cb6675907a7fa9fbedc1c02e7f 2025-12-04T09:20:36.0829697Z * [new tag] trunk/285779b1621cf9f073a062b0889a642d200308d9 -> trunk/285779b1621cf9f073a062b0889a642d200308d9 2025-12-04T09:20:36.0830930Z * [new tag] trunk/2887faaec6295d081580d09fce161201826c6d87 -> trunk/2887faaec6295d081580d09fce161201826c6d87 2025-12-04T09:20:36.0832339Z * [new tag] trunk/296e67c92635443c67b11c0ae1bd045f03ebb7bc -> trunk/296e67c92635443c67b11c0ae1bd045f03ebb7bc 2025-12-04T09:20:36.0833775Z * [new tag] trunk/29856679769b3dede478767e2fe6cfb51197cb25 -> trunk/29856679769b3dede478767e2fe6cfb51197cb25 2025-12-04T09:20:36.0835193Z * [new tag] trunk/29e5455a4740c326ab187c7aa7b5ef98034ea563 -> trunk/29e5455a4740c326ab187c7aa7b5ef98034ea563 2025-12-04T09:20:36.0836605Z * [new tag] trunk/2ac3ef882afb23136adc188975f0a8802fc68adf -> trunk/2ac3ef882afb23136adc188975f0a8802fc68adf 2025-12-04T09:20:36.0837840Z * [new tag] trunk/2bec68e73b64715354af076ad309335f943e36cd -> trunk/2bec68e73b64715354af076ad309335f943e36cd 2025-12-04T09:20:36.0839141Z * [new tag] trunk/2c87367e6f88662cd5cedbd1537748b7948c38e1 -> trunk/2c87367e6f88662cd5cedbd1537748b7948c38e1 2025-12-04T09:20:36.0840571Z * [new tag] trunk/2d1f78fe3ec13820f136a2e0336da12a25f41708 -> trunk/2d1f78fe3ec13820f136a2e0336da12a25f41708 2025-12-04T09:20:36.0841993Z * [new tag] trunk/2df6058f116a65722a0e03073402feb242572d35 -> trunk/2df6058f116a65722a0e03073402feb242572d35 2025-12-04T09:20:36.0843255Z * [new tag] trunk/2e0c2e170fe658c440775c8e5c44228aafcc47ec -> trunk/2e0c2e170fe658c440775c8e5c44228aafcc47ec 2025-12-04T09:20:36.0844722Z * [new tag] trunk/2f9b7dad7b5419b063bd0f2e204de192720ebb94 -> trunk/2f9b7dad7b5419b063bd0f2e204de192720ebb94 2025-12-04T09:20:36.0846075Z * [new tag] trunk/305168768a95d69c444df5cd334bb774edfe06f1 -> trunk/305168768a95d69c444df5cd334bb774edfe06f1 2025-12-04T09:20:36.0847459Z * [new tag] trunk/31fc12773026e8e00f054dd79ad9b2491e693b48 -> trunk/31fc12773026e8e00f054dd79ad9b2491e693b48 2025-12-04T09:20:36.0848867Z * [new tag] trunk/320de0c6b0a3e7c6d2693ea5c28d5d0156ba7991 -> trunk/320de0c6b0a3e7c6d2693ea5c28d5d0156ba7991 2025-12-04T09:20:36.0850367Z * [new tag] trunk/3418bd29475dff06695045fcdf93e7d0dac67da8 -> trunk/3418bd29475dff06695045fcdf93e7d0dac67da8 2025-12-04T09:20:36.0851729Z * [new tag] trunk/34a98608afa0cb5b48f0d6d30432fdd0a2614ddf -> trunk/34a98608afa0cb5b48f0d6d30432fdd0a2614ddf 2025-12-04T09:20:36.0853089Z * [new tag] trunk/35b7a9a26c5923d98aebaa41a031dae21788a9ee -> trunk/35b7a9a26c5923d98aebaa41a031dae21788a9ee 2025-12-04T09:20:36.0854513Z * [new tag] trunk/39d07dbf03a911bdd45d1af78d8638dc92074938 -> trunk/39d07dbf03a911bdd45d1af78d8638dc92074938 2025-12-04T09:20:36.0855728Z * [new tag] trunk/3cd98b4205ada151042cc7ff097a82d4a4b18725 -> trunk/3cd98b4205ada151042cc7ff097a82d4a4b18725 2025-12-04T09:20:36.0857088Z * [new tag] trunk/3d35fd20a78ff4d016fa80f4e5fad37191d7bcae -> trunk/3d35fd20a78ff4d016fa80f4e5fad37191d7bcae 2025-12-04T09:20:36.0858495Z * [new tag] trunk/409a5fee945c46a3edaf5df162812f201bfd7b2f -> trunk/409a5fee945c46a3edaf5df162812f201bfd7b2f 2025-12-04T09:20:36.0859879Z * [new tag] trunk/42e9005cda22da3f1c559c3649218cebd671027c -> trunk/42e9005cda22da3f1c559c3649218cebd671027c 2025-12-04T09:20:36.0861275Z * [new tag] trunk/43b94713bbf340d3c124fde02d0f73add4021247 -> trunk/43b94713bbf340d3c124fde02d0f73add4021247 2025-12-04T09:20:36.0862684Z * [new tag] trunk/44ac69388a4a5eb463dbd2a13f00d1e3b924566c -> trunk/44ac69388a4a5eb463dbd2a13f00d1e3b924566c 2025-12-04T09:20:36.0864080Z * [new tag] trunk/45d14e2497292be06ad36eaa1aaaf7c630a2586a -> trunk/45d14e2497292be06ad36eaa1aaaf7c630a2586a 2025-12-04T09:20:36.0865331Z * [new tag] trunk/45d310ad84854dff730c0b12e577d7998d978686 -> trunk/45d310ad84854dff730c0b12e577d7998d978686 2025-12-04T09:20:36.0866900Z * [new tag] trunk/47b28ddf7bd74b50fa93b307a7d3b183a6d77f54 -> trunk/47b28ddf7bd74b50fa93b307a7d3b183a6d77f54 2025-12-04T09:20:36.0868157Z * [new tag] trunk/481e5ab336275bd3acd5fa8a611b05b4469012af -> trunk/481e5ab336275bd3acd5fa8a611b05b4469012af 2025-12-04T09:20:36.0870027Z * [new tag] trunk/491731647f6b8a9345dcfb3bc9416aea254a7d96 -> trunk/491731647f6b8a9345dcfb3bc9416aea254a7d96 2025-12-04T09:20:36.0871597Z * [new tag] trunk/49a04d26088acc17d948ddd66920f3e16371e873 -> trunk/49a04d26088acc17d948ddd66920f3e16371e873 2025-12-04T09:20:36.0872973Z * [new tag] trunk/4bebc827c47d2f1f0fa1a417a5201a97aef3d985 -> trunk/4bebc827c47d2f1f0fa1a417a5201a97aef3d985 2025-12-04T09:20:36.0874283Z * [new tag] trunk/4c246677784c6a14bc2dbb9ff8773ef0a3a3222f -> trunk/4c246677784c6a14bc2dbb9ff8773ef0a3a3222f 2025-12-04T09:20:36.0875776Z * [new tag] trunk/4cfb47ff548b6d996641058cf04a70e311a4c3aa -> trunk/4cfb47ff548b6d996641058cf04a70e311a4c3aa 2025-12-04T09:20:36.0877253Z * [new tag] trunk/4e0061c1aa52f606dda8cfab0bd7591e588faf2c -> trunk/4e0061c1aa52f606dda8cfab0bd7591e588faf2c 2025-12-04T09:20:36.0878936Z * [new tag] trunk/4fefb8e7e942386ffac764a41b232241f82bea3a -> trunk/4fefb8e7e942386ffac764a41b232241f82bea3a 2025-12-04T09:20:36.0880290Z * [new tag] trunk/503b2640023521f5a35cd9a52fc8033d73a95d0d -> trunk/503b2640023521f5a35cd9a52fc8033d73a95d0d 2025-12-04T09:20:36.0881682Z * [new tag] trunk/518c2b1b3dab9a2ef2849e04b3bc2f20c1c41db9 -> trunk/518c2b1b3dab9a2ef2849e04b3bc2f20c1c41db9 2025-12-04T09:20:36.0883043Z * [new tag] trunk/5191b2fa68ba19960912bfd7fd721c79d76bb1f3 -> trunk/5191b2fa68ba19960912bfd7fd721c79d76bb1f3 2025-12-04T09:20:36.0884544Z * [new tag] trunk/52ac0f0dc4acacd219f1317fbc28ec631c01e07a -> trunk/52ac0f0dc4acacd219f1317fbc28ec631c01e07a 2025-12-04T09:20:36.0885945Z * [new tag] trunk/539ba711b029de9f191070f4f0d12f18f5b7f292 -> trunk/539ba711b029de9f191070f4f0d12f18f5b7f292 2025-12-04T09:20:36.0887306Z * [new tag] trunk/556375b55deebebbc56cb7aef81f4d52f031ba28 -> trunk/556375b55deebebbc56cb7aef81f4d52f031ba28 2025-12-04T09:20:36.0888757Z * [new tag] trunk/55c4ab554845481d0a69a3811937575fe8bb1a66 -> trunk/55c4ab554845481d0a69a3811937575fe8bb1a66 2025-12-04T09:20:36.0890121Z * [new tag] trunk/5634469fda9e5d98869c82c7d03bb08914245f96 -> trunk/5634469fda9e5d98869c82c7d03bb08914245f96 2025-12-04T09:20:36.0891396Z * [new tag] trunk/5778f6ff894686a975a9a23645178ae4c87ad5dc -> trunk/5778f6ff894686a975a9a23645178ae4c87ad5dc 2025-12-04T09:20:36.0892794Z * [new tag] trunk/587d63a3e07de5dc91065f9ef70bcacda9989068 -> trunk/587d63a3e07de5dc91065f9ef70bcacda9989068 2025-12-04T09:20:36.0894164Z * [new tag] trunk/597930f6b568852356ca9795dac76f9e4653adbd -> trunk/597930f6b568852356ca9795dac76f9e4653adbd 2025-12-04T09:20:36.0895495Z * [new tag] trunk/597df3a4e2a67b9fdbe1a89b2f4d74f822274db6 -> trunk/597df3a4e2a67b9fdbe1a89b2f4d74f822274db6 2025-12-04T09:20:36.0896987Z * [new tag] trunk/59abd50e931f4efb21b053f7a2911f5d8a49d883 -> trunk/59abd50e931f4efb21b053f7a2911f5d8a49d883 2025-12-04T09:20:36.0898325Z * [new tag] trunk/5a607febc04c3a2b5824c75f3f60307867439a2c -> trunk/5a607febc04c3a2b5824c75f3f60307867439a2c 2025-12-04T09:20:36.0899849Z * [new tag] trunk/5bf1cdf4755c54ef462b44cb8041b0a57311556b -> trunk/5bf1cdf4755c54ef462b44cb8041b0a57311556b 2025-12-04T09:20:36.0901106Z * [new tag] trunk/5f0030ba63d334d7e8c93a09e41403b89e4c573c -> trunk/5f0030ba63d334d7e8c93a09e41403b89e4c573c 2025-12-04T09:20:36.0902550Z * [new tag] trunk/5f21d27e71268464d362a96c9ac09ea475f7f202 -> trunk/5f21d27e71268464d362a96c9ac09ea475f7f202 2025-12-04T09:20:36.0903944Z * [new tag] trunk/5fafc13038c9988d9ac21fa793fbd5890604b447 -> trunk/5fafc13038c9988d9ac21fa793fbd5890604b447 2025-12-04T09:20:36.0905364Z * [new tag] trunk/61be54a31dc09b59d99b62176fb935aee0b924ef -> trunk/61be54a31dc09b59d99b62176fb935aee0b924ef 2025-12-04T09:20:36.0906735Z * [new tag] trunk/62d3ccd71484ed6a760d909b41487101bbc65719 -> trunk/62d3ccd71484ed6a760d909b41487101bbc65719 2025-12-04T09:20:36.0908097Z * [new tag] trunk/641cdb68ae27668eb441d0e49c87a0602c120c2b -> trunk/641cdb68ae27668eb441d0e49c87a0602c120c2b 2025-12-04T09:20:36.0909458Z * [new tag] trunk/65c4620d6bb0c6029f69762c22b91dda2294da9a -> trunk/65c4620d6bb0c6029f69762c22b91dda2294da9a 2025-12-04T09:20:36.0910837Z * [new tag] trunk/66004b993744b4106bf8afaba71f3c228a804206 -> trunk/66004b993744b4106bf8afaba71f3c228a804206 2025-12-04T09:20:36.0912254Z * [new tag] trunk/6658a04c7ca67acb64512341342e7b3ee13ee386 -> trunk/6658a04c7ca67acb64512341342e7b3ee13ee386 2025-12-04T09:20:36.0913508Z * [new tag] trunk/6864e309092a71f8ab0ca6a4dc7f8a4073fd31c4 -> trunk/6864e309092a71f8ab0ca6a4dc7f8a4073fd31c4 2025-12-04T09:20:36.0915063Z * [new tag] trunk/6c261c6cb07892c90ca19ed51c9705b1659a3f7d -> trunk/6c261c6cb07892c90ca19ed51c9705b1659a3f7d 2025-12-04T09:20:36.0916330Z * [new tag] trunk/6c8b6a043f1628188b6396b3a2a6e000ca68362b -> trunk/6c8b6a043f1628188b6396b3a2a6e000ca68362b 2025-12-04T09:20:36.0917700Z * [new tag] trunk/6ceb4a32f92ae67ce5d7d97931d17401ebf5ffa5 -> trunk/6ceb4a32f92ae67ce5d7d97931d17401ebf5ffa5 2025-12-04T09:20:36.0919087Z * [new tag] trunk/6e404e9b7d6f5fb0de86aa73888c3038248c17f8 -> trunk/6e404e9b7d6f5fb0de86aa73888c3038248c17f8 2025-12-04T09:20:36.0920537Z * [new tag] trunk/6ec30b490aee1db6bcdc7340abddef25784f08ec -> trunk/6ec30b490aee1db6bcdc7340abddef25784f08ec 2025-12-04T09:20:36.0921953Z * [new tag] trunk/6f2783a6c08e1db34275ff25176ffe9aebc30a71 -> trunk/6f2783a6c08e1db34275ff25176ffe9aebc30a71 2025-12-04T09:20:36.0923302Z * [new tag] trunk/6f53fefeb90ad3281119b5cfc4aa9ffd8a066e3d -> trunk/6f53fefeb90ad3281119b5cfc4aa9ffd8a066e3d 2025-12-04T09:20:36.0924694Z * [new tag] trunk/6f7dcf51e46d0c880db1a2f5c70de57adb576f4a -> trunk/6f7dcf51e46d0c880db1a2f5c70de57adb576f4a 2025-12-04T09:20:36.0926152Z * [new tag] trunk/6ff831180d2fa436c7f1c1af3adac641fce9d60e -> trunk/6ff831180d2fa436c7f1c1af3adac641fce9d60e 2025-12-04T09:20:36.0927521Z * [new tag] trunk/70076464a63ab218a7ceefb0e76ccd7131deb8f8 -> trunk/70076464a63ab218a7ceefb0e76ccd7131deb8f8 2025-12-04T09:20:36.0928892Z * [new tag] trunk/70d797a5fc109b20a517646fcaa819477cd0d485 -> trunk/70d797a5fc109b20a517646fcaa819477cd0d485 2025-12-04T09:20:36.0930226Z * [new tag] trunk/7348cb355ff0a6f79cd4871215aea72185748734 -> trunk/7348cb355ff0a6f79cd4871215aea72185748734 2025-12-04T09:20:36.0931664Z * [new tag] trunk/74fe26a1ebe32931783569f2e762e3c2c974901f -> trunk/74fe26a1ebe32931783569f2e762e3c2c974901f 2025-12-04T09:20:36.0933102Z * [new tag] trunk/76aeb8c7e0f795b3fddca134cbea9a69da3ee696 -> trunk/76aeb8c7e0f795b3fddca134cbea9a69da3ee696 2025-12-04T09:20:36.0934344Z * [new tag] trunk/7716da9fb23f27a65b41f9f016a2afadf281c18f -> trunk/7716da9fb23f27a65b41f9f016a2afadf281c18f 2025-12-04T09:20:36.0935735Z * [new tag] trunk/7741edd4ed665f3988052e260863efb508d61a03 -> trunk/7741edd4ed665f3988052e260863efb508d61a03 2025-12-04T09:20:36.0937151Z * [new tag] trunk/78adb3b3df41b45d2368b67226d2f864b78939a6 -> trunk/78adb3b3df41b45d2368b67226d2f864b78939a6 2025-12-04T09:20:36.0938651Z * [new tag] trunk/79d7b178225e5ed24d4e1db74e5abbff848f5fb7 -> trunk/79d7b178225e5ed24d4e1db74e5abbff848f5fb7 2025-12-04T09:20:36.0939884Z * [new tag] trunk/7a1e316115fc6996b3f2336822ba5d5f6179f0c3 -> trunk/7a1e316115fc6996b3f2336822ba5d5f6179f0c3 2025-12-04T09:20:36.0941261Z * [new tag] trunk/7a41b66367c38d0af3e8a90f7be48d6b281e7bca -> trunk/7a41b66367c38d0af3e8a90f7be48d6b281e7bca 2025-12-04T09:20:36.0942717Z * [new tag] trunk/7b7af390ea8541c611d1ce2018a6934188fc197b -> trunk/7b7af390ea8541c611d1ce2018a6934188fc197b 2025-12-04T09:20:36.0944151Z * [new tag] trunk/7ba4680f3755a560af81aa0f688791e367aa3609 -> trunk/7ba4680f3755a560af81aa0f688791e367aa3609 2025-12-04T09:20:36.0945511Z * [new tag] trunk/7bc2a66ded06a0b2549aa51d807edc5dc3e73d1b -> trunk/7bc2a66ded06a0b2549aa51d807edc5dc3e73d1b 2025-12-04T09:20:36.0946754Z * [new tag] trunk/7c648509a7470ace9fb2bae960dd4790f7e943e9 -> trunk/7c648509a7470ace9fb2bae960dd4790f7e943e9 2025-12-04T09:20:36.0948070Z * [new tag] trunk/7cbc2d034cecd21ab5c9707d0a9c525c17143fb8 -> trunk/7cbc2d034cecd21ab5c9707d0a9c525c17143fb8 2025-12-04T09:20:36.0949454Z * [new tag] trunk/7d1bbaf4ba301ea3fba6f3c7bc02d58f6417aaed -> trunk/7d1bbaf4ba301ea3fba6f3c7bc02d58f6417aaed 2025-12-04T09:20:36.0950917Z * [new tag] trunk/7d2a33e4ebf60b217a3cd77feae19231eb996fc8 -> trunk/7d2a33e4ebf60b217a3cd77feae19231eb996fc8 2025-12-04T09:20:36.0952215Z * [new tag] trunk/7eb625920054b1126a7d2d99818aaa188c6ba95e -> trunk/7eb625920054b1126a7d2d99818aaa188c6ba95e 2025-12-04T09:20:36.0953244Z * [new tag] trunk/7f55ba19c456a3d6cc443dd9edb6bb7cca677ead -> trunk/7f55ba19c456a3d6cc443dd9edb6bb7cca677ead 2025-12-04T09:20:36.0954730Z * [new tag] trunk/81af382128efa094d8702e18f2c133760904c718 -> trunk/81af382128efa094d8702e18f2c133760904c718 2025-12-04T09:20:36.0956309Z * [new tag] trunk/84149583d483e9c973c9a0feda70e4f3964947b0 -> trunk/84149583d483e9c973c9a0feda70e4f3964947b0 2025-12-04T09:20:36.0957925Z * [new tag] trunk/85a315917efe82c24306be805c584ec044951c75 -> trunk/85a315917efe82c24306be805c584ec044951c75 2025-12-04T09:20:36.0959657Z * [new tag] trunk/87329491c82a5f8c1cc4ec11d8f55a5de2551ece -> trunk/87329491c82a5f8c1cc4ec11d8f55a5de2551ece 2025-12-04T09:20:36.0960926Z * [new tag] trunk/892640e25aeefa8007c5af837214b4502b6b62a6 -> trunk/892640e25aeefa8007c5af837214b4502b6b62a6 2025-12-04T09:20:36.0962553Z * [new tag] trunk/89e3bbcb5b5321dc8b9520b4d5a8ee60cea1d0b4 -> trunk/89e3bbcb5b5321dc8b9520b4d5a8ee60cea1d0b4 2025-12-04T09:20:36.0963980Z * [new tag] trunk/8c73bbbb02159223c0c97d268a0a74cb78158a1c -> trunk/8c73bbbb02159223c0c97d268a0a74cb78158a1c 2025-12-04T09:20:36.0965437Z * [new tag] trunk/8d56e98c8db988a22cb2dfaeefb30bc7d2a3cc43 -> trunk/8d56e98c8db988a22cb2dfaeefb30bc7d2a3cc43 2025-12-04T09:20:36.0966909Z * [new tag] trunk/8d9dd9603e5ee26c01007f0cd4f018e584840922 -> trunk/8d9dd9603e5ee26c01007f0cd4f018e584840922 2025-12-04T09:20:36.0968313Z * [new tag] trunk/8ef0c0b02b062d75e7c9be2594914a3e784d23ca -> trunk/8ef0c0b02b062d75e7c9be2594914a3e784d23ca 2025-12-04T09:20:36.0969728Z * [new tag] trunk/90b27e7e8352cde97d32ddad24740ef819633f38 -> trunk/90b27e7e8352cde97d32ddad24740ef819633f38 2025-12-04T09:20:36.0971090Z * [new tag] trunk/90f0139e64b2951815d524b6a373bed20c4fbf90 -> trunk/90f0139e64b2951815d524b6a373bed20c4fbf90 2025-12-04T09:20:36.0972511Z * [new tag] trunk/93d0d6838c56af59b0dba794e6aa08f0c1c7799c -> trunk/93d0d6838c56af59b0dba794e6aa08f0c1c7799c 2025-12-04T09:20:36.0973973Z * [new tag] trunk/94ca8d5f1e81fea3ae488650a0fb6795049a9f87 -> trunk/94ca8d5f1e81fea3ae488650a0fb6795049a9f87 2025-12-04T09:20:36.0975298Z * [new tag] trunk/9844fbeadd5cebdf1281d6fbf79164139c352693 -> trunk/9844fbeadd5cebdf1281d6fbf79164139c352693 2025-12-04T09:20:36.0976690Z * [new tag] trunk/99024dec888ec1e50b546822a32b6fb2f35e5eaa -> trunk/99024dec888ec1e50b546822a32b6fb2f35e5eaa 2025-12-04T09:20:36.0978064Z * [new tag] trunk/9a296e640fc88aa44d275b48cd9cc30c573b169d -> trunk/9a296e640fc88aa44d275b48cd9cc30c573b169d 2025-12-04T09:20:36.0979512Z * [new tag] trunk/9b3e34d8589b29f7b4e7fab6f78711b7ca6e4639 -> trunk/9b3e34d8589b29f7b4e7fab6f78711b7ca6e4639 2025-12-04T09:20:36.0980969Z * [new tag] trunk/9cd055e547e9b67a5f9827f8999c38d7eda1bcb8 -> trunk/9cd055e547e9b67a5f9827f8999c38d7eda1bcb8 2025-12-04T09:20:36.0982432Z * [new tag] trunk/9f0df5686cb4ada94f94620acba2e3c3f363b11d -> trunk/9f0df5686cb4ada94f94620acba2e3c3f363b11d 2025-12-04T09:20:36.0983786Z * [new tag] trunk/9f7fceb887d0cfa0326a59b887821c63ff11340a -> trunk/9f7fceb887d0cfa0326a59b887821c63ff11340a 2025-12-04T09:20:36.0985213Z * [new tag] trunk/9f8ef8855d3078d70f7b782540ff2aaf158d6742 -> trunk/9f8ef8855d3078d70f7b782540ff2aaf158d6742 2025-12-04T09:20:36.0986636Z * [new tag] trunk/9fb52efc797b47a1f425a03aa5e47b866d8b1098 -> trunk/9fb52efc797b47a1f425a03aa5e47b866d8b1098 2025-12-04T09:20:36.0988000Z * [new tag] trunk/9ff4a2ebc5762d46c73e46b1b523d7ff349fedfa -> trunk/9ff4a2ebc5762d46c73e46b1b523d7ff349fedfa 2025-12-04T09:20:36.0989639Z * [new tag] trunk/a0f3937b94422354538ebbd47202d5b0e8a3fd0d -> trunk/a0f3937b94422354538ebbd47202d5b0e8a3fd0d 2025-12-04T09:20:36.0991041Z * [new tag] trunk/a15066c28b3145e6edbfc88359d0411d14cfc70c -> trunk/a15066c28b3145e6edbfc88359d0411d14cfc70c 2025-12-04T09:20:36.0992471Z * [new tag] trunk/a20f775e82564d2a9979221ed7f3b8d7cf54ce90 -> trunk/a20f775e82564d2a9979221ed7f3b8d7cf54ce90 2025-12-04T09:20:36.0993844Z * [new tag] trunk/a2973fb00ec002dd4b6bbf07385f066efb259b8c -> trunk/a2973fb00ec002dd4b6bbf07385f066efb259b8c 2025-12-04T09:20:36.0995083Z * [new tag] trunk/a7dc6dab9ad911259d4801c502907e531594db45 -> trunk/a7dc6dab9ad911259d4801c502907e531594db45 2025-12-04T09:20:36.0996551Z * [new tag] trunk/a951a9cee65c01660bbc6e6fded90ecb10fa6109 -> trunk/a951a9cee65c01660bbc6e6fded90ecb10fa6109 2025-12-04T09:20:36.0997911Z * [new tag] trunk/abfa1a6d65c7c159e35c72c25979b9da4971689e -> trunk/abfa1a6d65c7c159e35c72c25979b9da4971689e 2025-12-04T09:20:36.0999282Z * [new tag] trunk/ae3a2395bf66151078e2d201716f7d63ce1c6f3e -> trunk/ae3a2395bf66151078e2d201716f7d63ce1c6f3e 2025-12-04T09:20:36.1000582Z * [new tag] trunk/afdff7f0325080dedac44d080cb5a3b0e65e6c5e -> trunk/afdff7f0325080dedac44d080cb5a3b0e65e6c5e 2025-12-04T09:20:36.1001831Z * [new tag] trunk/b1aed4e7a72c03a38f44543aaea0dae2e9b76d48 -> trunk/b1aed4e7a72c03a38f44543aaea0dae2e9b76d48 2025-12-04T09:20:36.1003223Z * [new tag] trunk/b1decff555cd50e2123c8c6e25cc0d447c411f62 -> trunk/b1decff555cd50e2123c8c6e25cc0d447c411f62 2025-12-04T09:20:36.1004645Z * [new tag] trunk/b2b6b034c9fd08672c40e63ef243556ad4c49bd2 -> trunk/b2b6b034c9fd08672c40e63ef243556ad4c49bd2 2025-12-04T09:20:36.1006050Z * [new tag] trunk/b39813b4a04931682b0491adba2138d01d716d99 -> trunk/b39813b4a04931682b0491adba2138d01d716d99 2025-12-04T09:20:36.1007462Z * [new tag] trunk/b3a7edb2311367974cc7cd764cfb11a5d6758b24 -> trunk/b3a7edb2311367974cc7cd764cfb11a5d6758b24 2025-12-04T09:20:36.1008970Z * [new tag] trunk/b4cc1329c86acaef6d42c1fac7169b8d870ab0d7 -> trunk/b4cc1329c86acaef6d42c1fac7169b8d870ab0d7 2025-12-04T09:20:36.1010433Z * [new tag] trunk/b555c39217f765759954a4f9f9bd1e9b87bed11a -> trunk/b555c39217f765759954a4f9f9bd1e9b87bed11a 2025-12-04T09:20:36.1011866Z * [new tag] trunk/b6b6c80379388b7f9932c3e6a0f9907bf430e417 -> trunk/b6b6c80379388b7f9932c3e6a0f9907bf430e417 2025-12-04T09:20:36.1013336Z * [new tag] trunk/b6b6d912df0b6f4082f8e50b18bd1de1dd7325f4 -> trunk/b6b6d912df0b6f4082f8e50b18bd1de1dd7325f4 2025-12-04T09:20:36.1014723Z * [new tag] trunk/b7d60685f8cbc939b68a20871e90db67e729329b -> trunk/b7d60685f8cbc939b68a20871e90db67e729329b 2025-12-04T09:20:36.1016192Z * [new tag] trunk/b7f6b9a4fc6259f7af068f31868b3119bb1bac3e -> trunk/b7f6b9a4fc6259f7af068f31868b3119bb1bac3e 2025-12-04T09:20:36.1017712Z * [new tag] trunk/b8c4ba3593761e7b2a3ebd86f040fb07b47c02cf -> trunk/b8c4ba3593761e7b2a3ebd86f040fb07b47c02cf 2025-12-04T09:20:36.1019090Z * [new tag] trunk/b9c8f3a4884befb965ff42620ce44a71b04887f5 -> trunk/b9c8f3a4884befb965ff42620ce44a71b04887f5 2025-12-04T09:20:36.1020568Z * [new tag] trunk/ba1412546f3082c0958c077acc2025e4dbc33f1f -> trunk/ba1412546f3082c0958c077acc2025e4dbc33f1f 2025-12-04T09:20:36.1022001Z * [new tag] trunk/bac403c0b38c63bdbcc0c31f1c2b0bc0260f610f -> trunk/bac403c0b38c63bdbcc0c31f1c2b0bc0260f610f 2025-12-04T09:20:36.1023477Z * [new tag] trunk/bb3034198b459401fabeab254e1b99f0115046e2 -> trunk/bb3034198b459401fabeab254e1b99f0115046e2 2025-12-04T09:20:36.1024858Z * [new tag] trunk/bc39b2b3bc7a6e19a42e62bd576974035086fe55 -> trunk/bc39b2b3bc7a6e19a42e62bd576974035086fe55 2025-12-04T09:20:36.1026531Z * [new tag] trunk/bc43d5b297f207a11d83d77ddf0152bdaabe15a8 -> trunk/bc43d5b297f207a11d83d77ddf0152bdaabe15a8 2025-12-04T09:20:36.1027888Z * [new tag] trunk/bc6a4863c7246a6493d16d4ea6eee71ec07c6a09 -> trunk/bc6a4863c7246a6493d16d4ea6eee71ec07c6a09 2025-12-04T09:20:36.1029286Z * [new tag] trunk/bea4912944defdbcb8b061800caab6cbbbd01df5 -> trunk/bea4912944defdbcb8b061800caab6cbbbd01df5 2025-12-04T09:20:36.1030930Z * [new tag] trunk/c04e2c656f48d82d1521b867bbbf03967b9b7564 -> trunk/c04e2c656f48d82d1521b867bbbf03967b9b7564 2025-12-04T09:20:36.1032325Z * [new tag] trunk/c0660bcee27e7d7731634e274576a7081882bede -> trunk/c0660bcee27e7d7731634e274576a7081882bede 2025-12-04T09:20:36.1033838Z * [new tag] trunk/c178ed43d3d99cbefe84fbfb21d6f282b20d62ac -> trunk/c178ed43d3d99cbefe84fbfb21d6f282b20d62ac 2025-12-04T09:20:36.1035272Z * [new tag] trunk/c55b1e8f61d041ee436d697449eb028931d574fb -> trunk/c55b1e8f61d041ee436d697449eb028931d574fb 2025-12-04T09:20:36.1036554Z * [new tag] trunk/c6ae7579fe12fe75f1a8f7043a494c90567273f1 -> trunk/c6ae7579fe12fe75f1a8f7043a494c90567273f1 2025-12-04T09:20:36.1038084Z * [new tag] trunk/c8210e7d94bad5ae21ac389fa4ba8a463c76c4d0 -> trunk/c8210e7d94bad5ae21ac389fa4ba8a463c76c4d0 2025-12-04T09:20:36.1039674Z * [new tag] trunk/cc0853af42122f8185321f542616f4474e717f09 -> trunk/cc0853af42122f8185321f542616f4474e717f09 2025-12-04T09:20:36.1041012Z * [new tag] trunk/cddec6562eabfa390d014fa3741a5659cf9c94c9 -> trunk/cddec6562eabfa390d014fa3741a5659cf9c94c9 2025-12-04T09:20:36.1042406Z * [new tag] trunk/ce5e7e3bf1f4b69a4f4f93d288ba75b906df492a -> trunk/ce5e7e3bf1f4b69a4f4f93d288ba75b906df492a 2025-12-04T09:20:36.1043819Z * [new tag] trunk/d038b0130ec7c20ebcac219301292fd8e98a1ace -> trunk/d038b0130ec7c20ebcac219301292fd8e98a1ace 2025-12-04T09:20:36.1045147Z * [new tag] trunk/d16447dacaf2420ea175f0c275c75da951f57d39 -> trunk/d16447dacaf2420ea175f0c275c75da951f57d39 2025-12-04T09:20:36.1046633Z * [new tag] trunk/d19f1e8cab6810bb2e99141f9976665954c67a50 -> trunk/d19f1e8cab6810bb2e99141f9976665954c67a50 2025-12-04T09:20:36.1048006Z * [new tag] trunk/d1c9f03b2a5af4104721712f8cdffe9b4f340c01 -> trunk/d1c9f03b2a5af4104721712f8cdffe9b4f340c01 2025-12-04T09:20:36.1049436Z * [new tag] trunk/d40f4950f2b7f7aa380a22fe0f6166e71680fbcf -> trunk/d40f4950f2b7f7aa380a22fe0f6166e71680fbcf 2025-12-04T09:20:36.1051305Z * [new tag] trunk/d5038950bacfe36bbf24a47a455fe76901deb8e8 -> trunk/d5038950bacfe36bbf24a47a455fe76901deb8e8 2025-12-04T09:20:36.1052675Z * [new tag] trunk/d54ff42903c2ae0533931ff11d23b35f875bdb3d -> trunk/d54ff42903c2ae0533931ff11d23b35f875bdb3d 2025-12-04T09:20:36.1054125Z * [new tag] trunk/d76697633a2d2b9cced1ae21161849b33bfe7e47 -> trunk/d76697633a2d2b9cced1ae21161849b33bfe7e47 2025-12-04T09:20:36.1055594Z * [new tag] trunk/d78f52b199c547106d4cd9d2856dd0805c118bf1 -> trunk/d78f52b199c547106d4cd9d2856dd0805c118bf1 2025-12-04T09:20:36.1057001Z * [new tag] trunk/d8fd5c6eed28e5004150691d048a3f6785e19a8e -> trunk/d8fd5c6eed28e5004150691d048a3f6785e19a8e 2025-12-04T09:20:36.1058411Z * [new tag] trunk/d900f5e86745dec76713f4b0ef07005ef36b2f5a -> trunk/d900f5e86745dec76713f4b0ef07005ef36b2f5a 2025-12-04T09:20:36.1059924Z * [new tag] trunk/d973dc6b87d763859fe1c5bd1287e3b6b1c49d1b -> trunk/d973dc6b87d763859fe1c5bd1287e3b6b1c49d1b 2025-12-04T09:20:36.1061358Z * [new tag] trunk/d998c03304cb6ede76e1ed535b4ddeb6c2bf40ec -> trunk/d998c03304cb6ede76e1ed535b4ddeb6c2bf40ec 2025-12-04T09:20:36.1062728Z * [new tag] trunk/d9cb8a70833101dbbe16b99520cfbdd70d0a87bf -> trunk/d9cb8a70833101dbbe16b99520cfbdd70d0a87bf 2025-12-04T09:20:36.1064195Z * [new tag] trunk/d9d5e91b43f70eb8637af55db6856d49be391ffd -> trunk/d9d5e91b43f70eb8637af55db6856d49be391ffd 2025-12-04T09:20:36.1065567Z * [new tag] trunk/dd18a75336a4fbd7497955cc5665904724fce889 -> trunk/dd18a75336a4fbd7497955cc5665904724fce889 2025-12-04T09:20:36.1066950Z * [new tag] trunk/ded9bcd61a059bf723e6e84689552962b480ea77 -> trunk/ded9bcd61a059bf723e6e84689552962b480ea77 2025-12-04T09:20:36.1068692Z * [new tag] trunk/dfbd3714d15c37a7b83b322a6b60f997fc00f50c -> trunk/dfbd3714d15c37a7b83b322a6b60f997fc00f50c 2025-12-04T09:20:36.1070136Z * [new tag] trunk/e115f9f4e4b039f8e9a642aaa2bd8254a920541b -> trunk/e115f9f4e4b039f8e9a642aaa2bd8254a920541b 2025-12-04T09:20:36.1071502Z * [new tag] trunk/e3f24fd73ad74c6e7176687986436956c7c18235 -> trunk/e3f24fd73ad74c6e7176687986436956c7c18235 2025-12-04T09:20:36.1073993Z * [new tag] trunk/e7d24d3ff93d1503ba63860b7057438ad93f918e -> trunk/e7d24d3ff93d1503ba63860b7057438ad93f918e 2025-12-04T09:20:36.1075478Z * [new tag] trunk/ea7035f462a0d2830865ee86c832bd101e1427fc -> trunk/ea7035f462a0d2830865ee86c832bd101e1427fc 2025-12-04T09:20:36.1076844Z * [new tag] trunk/eabb7ad2128580ef674446027b95bcf4e21e8df3 -> trunk/eabb7ad2128580ef674446027b95bcf4e21e8df3 2025-12-04T09:20:36.1078326Z * [new tag] trunk/eb5c63652a33da42e7018c23df5f20a3eb4c6ccf -> trunk/eb5c63652a33da42e7018c23df5f20a3eb4c6ccf 2025-12-04T09:20:36.1079747Z * [new tag] trunk/ec2c71f5c85021b8938cdafadce24c15a36fd93e -> trunk/ec2c71f5c85021b8938cdafadce24c15a36fd93e 2025-12-04T09:20:36.1081140Z * [new tag] trunk/ecbcc3f6bf327856b435b259ac63cc2f328c4b4e -> trunk/ecbcc3f6bf327856b435b259ac63cc2f328c4b4e 2025-12-04T09:20:36.1082957Z * [new tag] trunk/ee87bbe876c42575e961b32a0827d76bc9782ca2 -> trunk/ee87bbe876c42575e961b32a0827d76bc9782ca2 2025-12-04T09:20:36.1084360Z * [new tag] trunk/ef019d1d431c4c5a95b594cb90d40a50cd00f5e4 -> trunk/ef019d1d431c4c5a95b594cb90d40a50cd00f5e4 2025-12-04T09:20:36.1085772Z * [new tag] trunk/ef8ecc13830a86c4b231f1aad9aba7851db61b53 -> trunk/ef8ecc13830a86c4b231f1aad9aba7851db61b53 2025-12-04T09:20:36.1087164Z * [new tag] trunk/f1076f5510920044912247b1abb8760cb820f598 -> trunk/f1076f5510920044912247b1abb8760cb820f598 2025-12-04T09:20:36.1088568Z * [new tag] trunk/f2d6a75a00a1d648ca9a0abc6a33e14c3dea6c40 -> trunk/f2d6a75a00a1d648ca9a0abc6a33e14c3dea6c40 2025-12-04T09:20:36.1090004Z * [new tag] trunk/f47dd0ddef1359e5b43e4b962412f67b30ecde56 -> trunk/f47dd0ddef1359e5b43e4b962412f67b30ecde56 2025-12-04T09:20:36.1091414Z * [new tag] trunk/f49d32dfa4730dcfb1b60eeeb369b5889da983c8 -> trunk/f49d32dfa4730dcfb1b60eeeb369b5889da983c8 2025-12-04T09:20:36.1092685Z * [new tag] trunk/f4dedf78fc30fd4b93975787ca6074ee89db9467 -> trunk/f4dedf78fc30fd4b93975787ca6074ee89db9467 2025-12-04T09:20:36.1094103Z * [new tag] trunk/f7c0d03819ebed05c4038f095d66d1b8c54aca17 -> trunk/f7c0d03819ebed05c4038f095d66d1b8c54aca17 2025-12-04T09:20:36.1095482Z * [new tag] trunk/f7e1bd80a063e17453c361837ba6ea2570920a73 -> trunk/f7e1bd80a063e17453c361837ba6ea2570920a73 2025-12-04T09:20:36.1096801Z * [new tag] trunk/f9bd6c53624c7c0ea3772de78498326e84c2f0e7 -> trunk/f9bd6c53624c7c0ea3772de78498326e84c2f0e7 2025-12-04T09:20:36.1098242Z * [new tag] trunk/fb5be221a46b51bfc9509013b0d85bc5a9d4f15b -> trunk/fb5be221a46b51bfc9509013b0d85bc5a9d4f15b 2025-12-04T09:20:36.1099777Z * [new tag] trunk/fdf863d5e1de3b2688c9511e96876e34581dbfd7 -> trunk/fdf863d5e1de3b2688c9511e96876e34581dbfd7 2025-12-04T09:20:36.1101592Z * [new tag] trunk/fe0e65adfc0e7ca6e5f57e6ea8b16bd5cc967307 -> trunk/fe0e65adfc0e7ca6e5f57e6ea8b16bd5cc967307 2025-12-04T09:20:36.1103212Z * [new tag] trunk/fec710bf89173f5355468a7ce1afe9157c3d9009 -> trunk/fec710bf89173f5355468a7ce1afe9157c3d9009 2025-12-04T09:20:36.1104767Z * [new tag] trunk/ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 -> trunk/ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:20:36.1105783Z * [new tag] v0.1.1 -> v0.1.1 2025-12-04T09:20:36.1107144Z * [new tag] v0.1.10 -> v0.1.10 2025-12-04T09:20:36.1108417Z * [new tag] v0.1.11 -> v0.1.11 2025-12-04T09:20:36.1109782Z * [new tag] v0.1.12 -> v0.1.12 2025-12-04T09:20:36.1111040Z * [new tag] v0.1.2 -> v0.1.2 2025-12-04T09:20:36.1112382Z * [new tag] v0.1.3 -> v0.1.3 2025-12-04T09:20:36.1113735Z * [new tag] v0.1.4 -> v0.1.4 2025-12-04T09:20:36.1115040Z * [new tag] v0.1.5 -> v0.1.5 2025-12-04T09:20:36.1116369Z * [new tag] v0.1.6 -> v0.1.6 2025-12-04T09:20:36.1117634Z * [new tag] v0.1.7 -> v0.1.7 2025-12-04T09:20:36.1118944Z * [new tag] v0.1.8 -> v0.1.8 2025-12-04T09:20:36.1120262Z * [new tag] v0.1.9 -> v0.1.9 2025-12-04T09:20:36.1121655Z * [new tag] v0.2.0 -> v0.2.0 2025-12-04T09:20:36.1122907Z * [new tag] v0.3.0 -> v0.3.0 2025-12-04T09:20:36.1124306Z * [new tag] v0.3.1 -> v0.3.1 2025-12-04T09:20:36.1125608Z * [new tag] v0.4.0 -> v0.4.0 2025-12-04T09:20:36.1126907Z * [new tag] v0.4.1 -> v0.4.1 2025-12-04T09:20:36.1128258Z * [new tag] v1.0.0 -> v1.0.0 2025-12-04T09:20:36.1129616Z * [new tag] v1.0.0a0 -> v1.0.0a0 2025-12-04T09:20:36.1130960Z * [new tag] v1.0.1 -> v1.0.1 2025-12-04T09:20:36.1132378Z * [new tag] v1.0rc0 -> v1.0rc0 2025-12-04T09:20:36.1133535Z * [new tag] v1.0rc1 -> v1.0rc1 2025-12-04T09:20:36.1134869Z * [new tag] v1.1.0 -> v1.1.0 2025-12-04T09:20:36.1136326Z * [new tag] v1.1.0a0 -> v1.1.0a0 2025-12-04T09:20:36.1137843Z * [new tag] v1.10.0 -> v1.10.0 2025-12-04T09:20:36.1139293Z * [new tag] v1.10.0-rc1 -> v1.10.0-rc1 2025-12-04T09:20:36.1140706Z * [new tag] v1.10.0-rc2 -> v1.10.0-rc2 2025-12-04T09:20:36.1141829Z * [new tag] v1.10.0-rc3 -> v1.10.0-rc3 2025-12-04T09:20:36.1143216Z * [new tag] v1.10.1 -> v1.10.1 2025-12-04T09:20:36.1144358Z * [new tag] v1.10.1-rc1 -> v1.10.1-rc1 2025-12-04T09:20:36.1145597Z * [new tag] v1.10.2 -> v1.10.2 2025-12-04T09:20:36.1146732Z * [new tag] v1.10.2-rc1 -> v1.10.2-rc1 2025-12-04T09:20:36.1148080Z * [new tag] v1.11.0 -> v1.11.0 2025-12-04T09:20:36.1149450Z * [new tag] v1.11.0-rc1 -> v1.11.0-rc1 2025-12-04T09:20:36.1151029Z * [new tag] v1.11.0-rc2 -> v1.11.0-rc2 2025-12-04T09:20:36.1152325Z * [new tag] v1.11.0-rc3 -> v1.11.0-rc3 2025-12-04T09:20:36.1153735Z * [new tag] v1.11.0-rc4 -> v1.11.0-rc4 2025-12-04T09:20:36.1155099Z * [new tag] v1.11.0-rc5 -> v1.11.0-rc5 2025-12-04T09:20:36.1156264Z * [new tag] v1.11.0-rc6 -> v1.11.0-rc6 2025-12-04T09:20:36.1157382Z * [new tag] v1.11.0-rc7 -> v1.11.0-rc7 2025-12-04T09:20:36.1158907Z * [new tag] v1.12.0 -> v1.12.0 2025-12-04T09:20:36.1160132Z * [new tag] v1.12.0-rc1 -> v1.12.0-rc1 2025-12-04T09:20:36.1161461Z * [new tag] v1.12.0-rc2 -> v1.12.0-rc2 2025-12-04T09:20:36.1162886Z * [new tag] v1.12.0-rc3 -> v1.12.0-rc3 2025-12-04T09:20:36.1164269Z * [new tag] v1.12.0-rc4 -> v1.12.0-rc4 2025-12-04T09:20:36.1165562Z * [new tag] v1.12.0-rc5 -> v1.12.0-rc5 2025-12-04T09:20:36.1167040Z * [new tag] v1.12.0-rc6 -> v1.12.0-rc6 2025-12-04T09:20:36.1168190Z * [new tag] v1.12.0-rc7 -> v1.12.0-rc7 2025-12-04T09:20:36.1169392Z * [new tag] v1.12.0-rc8 -> v1.12.0-rc8 2025-12-04T09:20:36.1170498Z * [new tag] v1.12.1 -> v1.12.1 2025-12-04T09:20:36.1172193Z * [new tag] v1.12.1-rc1 -> v1.12.1-rc1 2025-12-04T09:20:36.1173541Z * [new tag] v1.12.1-rc2 -> v1.12.1-rc2 2025-12-04T09:20:36.1175326Z * [new tag] v1.12.1-rc3 -> v1.12.1-rc3 2025-12-04T09:20:36.1176700Z * [new tag] v1.12.1-rc4 -> v1.12.1-rc4 2025-12-04T09:20:36.1177834Z * [new tag] v1.12.1-rc5 -> v1.12.1-rc5 2025-12-04T09:20:36.1179416Z * [new tag] v1.13.0 -> v1.13.0 2025-12-04T09:20:36.1180798Z * [new tag] v1.13.0-rc1 -> v1.13.0-rc1 2025-12-04T09:20:36.1182152Z * [new tag] v1.13.0-rc2 -> v1.13.0-rc2 2025-12-04T09:20:36.1183502Z * [new tag] v1.13.0-rc3 -> v1.13.0-rc3 2025-12-04T09:20:36.1184983Z * [new tag] v1.13.0-rc4 -> v1.13.0-rc4 2025-12-04T09:20:36.1186127Z * [new tag] v1.13.0-rc5 -> v1.13.0-rc5 2025-12-04T09:20:36.1187252Z * [new tag] v1.13.0-rc6 -> v1.13.0-rc6 2025-12-04T09:20:36.1188731Z * [new tag] v1.13.1 -> v1.13.1 2025-12-04T09:20:36.1189880Z * [new tag] v1.13.1-rc1 -> v1.13.1-rc1 2025-12-04T09:20:36.1191199Z * [new tag] v1.2.0 -> v1.2.0 2025-12-04T09:20:36.1192487Z * [new tag] v1.2.0a0 -> v1.2.0a0 2025-12-04T09:20:36.1193827Z * [new tag] v1.3.0 -> v1.3.0 2025-12-04T09:20:36.1195185Z * [new tag] v1.3.0a0 -> v1.3.0a0 2025-12-04T09:20:36.1196327Z * [new tag] v1.3.1 -> v1.3.1 2025-12-04T09:20:36.1197654Z * [new tag] v1.4.0 -> v1.4.0 2025-12-04T09:20:36.1199051Z * [new tag] v1.4.0a0 -> v1.4.0a0 2025-12-04T09:20:36.1200204Z * [new tag] v1.4.1 -> v1.4.1 2025-12-04T09:20:36.1201573Z * [new tag] v1.5.0 -> v1.5.0 2025-12-04T09:20:36.1202929Z * [new tag] v1.5.0-rc1 -> v1.5.0-rc1 2025-12-04T09:20:36.1204328Z * [new tag] v1.5.0-rc2 -> v1.5.0-rc2 2025-12-04T09:20:36.1205773Z * [new tag] v1.5.0-rc3 -> v1.5.0-rc3 2025-12-04T09:20:36.1207052Z * [new tag] v1.5.0-rc4 -> v1.5.0-rc4 2025-12-04T09:20:36.1208205Z * [new tag] v1.5.0-rc5 -> v1.5.0-rc5 2025-12-04T09:20:36.1209582Z * [new tag] v1.5.1 -> v1.5.1 2025-12-04T09:20:36.1210753Z * [new tag] v1.5.1-rc1 -> v1.5.1-rc1 2025-12-04T09:20:36.1211895Z * [new tag] v1.6.0 -> v1.6.0 2025-12-04T09:20:36.1213386Z * [new tag] v1.6.0-rc1 -> v1.6.0-rc1 2025-12-04T09:20:36.1214879Z * [new tag] v1.6.0-rc2 -> v1.6.0-rc2 2025-12-04T09:20:36.1216172Z * [new tag] v1.6.0-rc3 -> v1.6.0-rc3 2025-12-04T09:20:36.1217497Z * [new tag] v1.6.0-rc4 -> v1.6.0-rc4 2025-12-04T09:20:36.1218823Z * [new tag] v1.6.0-rc5 -> v1.6.0-rc5 2025-12-04T09:20:36.1220306Z * [new tag] v1.6.0-rc6 -> v1.6.0-rc6 2025-12-04T09:20:36.1221405Z * [new tag] v1.6.0-rc7 -> v1.6.0-rc7 2025-12-04T09:20:36.1222775Z * [new tag] v1.7.0 -> v1.7.0 2025-12-04T09:20:36.1224220Z * [new tag] v1.7.0-rc1 -> v1.7.0-rc1 2025-12-04T09:20:36.1225658Z * [new tag] v1.7.0-rc2 -> v1.7.0-rc2 2025-12-04T09:20:36.1226954Z * [new tag] v1.7.0-rc3 -> v1.7.0-rc3 2025-12-04T09:20:36.1228108Z * [new tag] v1.7.0-rc4 -> v1.7.0-rc4 2025-12-04T09:20:36.1229488Z * [new tag] v1.7.1 -> v1.7.1 2025-12-04T09:20:36.1231026Z * [new tag] v1.7.1-rc1 -> v1.7.1-rc1 2025-12-04T09:20:36.1232431Z * [new tag] v1.7.1-rc2 -> v1.7.1-rc2 2025-12-04T09:20:36.1233603Z * [new tag] v1.7.1-rc3 -> v1.7.1-rc3 2025-12-04T09:20:36.1234990Z * [new tag] v1.8.0 -> v1.8.0 2025-12-04T09:20:36.1236151Z * [new tag] v1.8.0-rc1 -> v1.8.0-rc1 2025-12-04T09:20:36.1237515Z * [new tag] v1.8.0-rc2 -> v1.8.0-rc2 2025-12-04T09:20:36.1238856Z * [new tag] v1.8.0-rc3 -> v1.8.0-rc3 2025-12-04T09:20:36.1240194Z * [new tag] v1.8.0-rc4 -> v1.8.0-rc4 2025-12-04T09:20:36.1241308Z * [new tag] v1.8.0-rc5 -> v1.8.0-rc5 2025-12-04T09:20:36.1242384Z * [new tag] v1.8.1 -> v1.8.1 2025-12-04T09:20:36.1243749Z * [new tag] v1.8.1-rc1 -> v1.8.1-rc1 2025-12-04T09:20:36.1244914Z * [new tag] v1.8.1-rc2 -> v1.8.1-rc2 2025-12-04T09:20:36.1246076Z * [new tag] v1.8.1-rc3 -> v1.8.1-rc3 2025-12-04T09:20:36.1247877Z * [new tag] v1.8.2 -> v1.8.2 2025-12-04T09:20:36.1249026Z * [new tag] v1.8.2-rc1 -> v1.8.2-rc1 2025-12-04T09:20:36.1250348Z * [new tag] v1.9.0 -> v1.9.0 2025-12-04T09:20:36.1251744Z * [new tag] v1.9.0-rc1 -> v1.9.0-rc1 2025-12-04T09:20:36.1253127Z * [new tag] v1.9.0-rc2 -> v1.9.0-rc2 2025-12-04T09:20:36.1254481Z * [new tag] v1.9.0-rc3 -> v1.9.0-rc3 2025-12-04T09:20:36.1255668Z * [new tag] v1.9.0-rc4 -> v1.9.0-rc4 2025-12-04T09:20:36.1257031Z * [new tag] v1.9.1 -> v1.9.1 2025-12-04T09:20:36.1258538Z * [new tag] v1.9.1-rc1 -> v1.9.1-rc1 2025-12-04T09:20:36.1260223Z * [new tag] v1.9.1-rc2 -> v1.9.1-rc2 2025-12-04T09:20:36.1261623Z * [new tag] v2.0.0 -> v2.0.0 2025-12-04T09:20:36.1263011Z * [new tag] v2.0.0-rc1 -> v2.0.0-rc1 2025-12-04T09:20:36.1264394Z * [new tag] v2.0.0-rc2 -> v2.0.0-rc2 2025-12-04T09:20:36.1265794Z * [new tag] v2.0.0-rc3 -> v2.0.0-rc3 2025-12-04T09:20:36.1267084Z * [new tag] v2.0.0-rc4 -> v2.0.0-rc4 2025-12-04T09:20:36.1268504Z * [new tag] v2.0.0-rc5 -> v2.0.0-rc5 2025-12-04T09:20:36.1269749Z * [new tag] v2.0.0-rc6 -> v2.0.0-rc6 2025-12-04T09:20:36.1271229Z * [new tag] v2.0.1 -> v2.0.1 2025-12-04T09:20:36.1272735Z * [new tag] v2.0.1-rc1 -> v2.0.1-rc1 2025-12-04T09:20:36.1273939Z * [new tag] v2.0.1-rc2 -> v2.0.1-rc2 2025-12-04T09:20:36.1275239Z * [new tag] v2.0.1-rc3 -> v2.0.1-rc3 2025-12-04T09:20:36.1276485Z * [new tag] v2.0.1-rc4 -> v2.0.1-rc4 2025-12-04T09:20:36.1278223Z * [new tag] v2.1.0 -> v2.1.0 2025-12-04T09:20:36.1279574Z * [new tag] v2.1.0-rc1 -> v2.1.0-rc1 2025-12-04T09:20:36.1280970Z * [new tag] v2.1.0-rc2 -> v2.1.0-rc2 2025-12-04T09:20:36.1282409Z * [new tag] v2.1.0-rc3 -> v2.1.0-rc3 2025-12-04T09:20:36.1283803Z * [new tag] v2.1.0-rc4 -> v2.1.0-rc4 2025-12-04T09:20:36.1285175Z * [new tag] v2.1.0-rc5 -> v2.1.0-rc5 2025-12-04T09:20:36.1286339Z * [new tag] v2.1.0-rc6 -> v2.1.0-rc6 2025-12-04T09:20:36.1287845Z * [new tag] v2.1.1 -> v2.1.1 2025-12-04T09:20:36.1289281Z * [new tag] v2.1.1-rc1 -> v2.1.1-rc1 2025-12-04T09:20:36.1290651Z * [new tag] v2.1.1-rc2 -> v2.1.1-rc2 2025-12-04T09:20:36.1292147Z * [new tag] v2.1.1-rc3 -> v2.1.1-rc3 2025-12-04T09:20:36.1293553Z * [new tag] v2.1.1-rc4 -> v2.1.1-rc4 2025-12-04T09:20:36.1294836Z * [new tag] v2.1.1-rc5 -> v2.1.1-rc5 2025-12-04T09:20:36.1295994Z * [new tag] v2.1.1-rc6 -> v2.1.1-rc6 2025-12-04T09:20:36.1297317Z * [new tag] v2.1.2 -> v2.1.2 2025-12-04T09:20:36.1298712Z * [new tag] v2.1.2-rc1 -> v2.1.2-rc1 2025-12-04T09:20:36.1300289Z * [new tag] v2.1.2-rc2 -> v2.1.2-rc2 2025-12-04T09:20:36.1301447Z * [new tag] v2.1.2-rc3 -> v2.1.2-rc3 2025-12-04T09:20:36.1302778Z * [new tag] v2.2.0 -> v2.2.0 2025-12-04T09:20:36.1304103Z * [new tag] v2.2.0-rc1 -> v2.2.0-rc1 2025-12-04T09:20:36.1305401Z * [new tag] v2.2.0-rc2 -> v2.2.0-rc2 2025-12-04T09:20:36.1306743Z * [new tag] v2.2.0-rc3 -> v2.2.0-rc3 2025-12-04T09:20:36.1308140Z * [new tag] v2.2.0-rc4 -> v2.2.0-rc4 2025-12-04T09:20:36.1309443Z * [new tag] v2.2.0-rc5 -> v2.2.0-rc5 2025-12-04T09:20:36.1310829Z * [new tag] v2.2.0-rc6 -> v2.2.0-rc6 2025-12-04T09:20:36.1311993Z * [new tag] v2.2.0-rc7 -> v2.2.0-rc7 2025-12-04T09:20:36.1313237Z * [new tag] v2.2.0-rc8 -> v2.2.0-rc8 2025-12-04T09:20:36.1314670Z * [new tag] v2.2.1 -> v2.2.1 2025-12-04T09:20:36.1316085Z * [new tag] v2.2.1-rc1 -> v2.2.1-rc1 2025-12-04T09:20:36.1317328Z * [new tag] v2.2.1-rc2 -> v2.2.1-rc2 2025-12-04T09:20:36.1318459Z * [new tag] v2.2.1-rc3 -> v2.2.1-rc3 2025-12-04T09:20:36.1319601Z * [new tag] v2.2.2 -> v2.2.2 2025-12-04T09:20:36.1321063Z * [new tag] v2.2.2-rc1 -> v2.2.2-rc1 2025-12-04T09:20:36.1322206Z * [new tag] v2.2.2-rc2 -> v2.2.2-rc2 2025-12-04T09:20:36.1323340Z * [new tag] v2.2.2-rc3 -> v2.2.2-rc3 2025-12-04T09:20:36.1324922Z * [new tag] v2.3.0 -> v2.3.0 2025-12-04T09:20:36.1326126Z * [new tag] v2.3.0-rc1 -> v2.3.0-rc1 2025-12-04T09:20:36.1327487Z * [new tag] v2.3.0-rc10 -> v2.3.0-rc10 2025-12-04T09:20:36.1328902Z * [new tag] v2.3.0-rc11 -> v2.3.0-rc11 2025-12-04T09:20:36.1330060Z * [new tag] v2.3.0-rc12 -> v2.3.0-rc12 2025-12-04T09:20:36.1331477Z * [new tag] v2.3.0-rc2 -> v2.3.0-rc2 2025-12-04T09:20:36.1332899Z * [new tag] v2.3.0-rc3 -> v2.3.0-rc3 2025-12-04T09:20:36.1334246Z * [new tag] v2.3.0-rc4 -> v2.3.0-rc4 2025-12-04T09:20:36.1335545Z * [new tag] v2.3.0-rc5 -> v2.3.0-rc5 2025-12-04T09:20:36.1336784Z * [new tag] v2.3.0-rc6 -> v2.3.0-rc6 2025-12-04T09:20:36.1338160Z * [new tag] v2.3.0-rc7 -> v2.3.0-rc7 2025-12-04T09:20:36.1339603Z * [new tag] v2.3.0-rc8 -> v2.3.0-rc8 2025-12-04T09:20:36.1340772Z * [new tag] v2.3.0-rc9 -> v2.3.0-rc9 2025-12-04T09:20:36.1341940Z * [new tag] v2.3.1 -> v2.3.1 2025-12-04T09:20:36.1343404Z * [new tag] v2.3.1-rc1 -> v2.3.1-rc1 2025-12-04T09:20:36.1344788Z * [new tag] v2.3.1-rc2 -> v2.3.1-rc2 2025-12-04T09:20:36.1346584Z * [new tag] v2.3.1-rc3 -> v2.3.1-rc3 2025-12-04T09:20:36.1347969Z * [new tag] v2.4.0 -> v2.4.0 2025-12-04T09:20:36.1349329Z * [new tag] v2.4.0-rc1 -> v2.4.0-rc1 2025-12-04T09:20:36.1350647Z * [new tag] v2.4.0-rc2 -> v2.4.0-rc2 2025-12-04T09:20:36.1351985Z * [new tag] v2.4.0-rc3 -> v2.4.0-rc3 2025-12-04T09:20:36.1353263Z * [new tag] v2.4.0-rc4 -> v2.4.0-rc4 2025-12-04T09:20:36.1354650Z * [new tag] v2.4.0-rc5 -> v2.4.0-rc5 2025-12-04T09:20:36.1356052Z * [new tag] v2.4.0-rc6 -> v2.4.0-rc6 2025-12-04T09:20:36.1357437Z * [new tag] v2.4.0-rc7 -> v2.4.0-rc7 2025-12-04T09:20:36.1358732Z * [new tag] v2.4.0-rc8 -> v2.4.0-rc8 2025-12-04T09:20:36.1360222Z * [new tag] v2.4.0-rc9 -> v2.4.0-rc9 2025-12-04T09:20:36.1361362Z * [new tag] v2.4.1 -> v2.4.1 2025-12-04T09:20:36.1362774Z * [new tag] v2.4.1-rc1 -> v2.4.1-rc1 2025-12-04T09:20:36.1364169Z * [new tag] v2.4.1-rc2 -> v2.4.1-rc2 2025-12-04T09:20:36.1365527Z * [new tag] v2.4.1-rc3 -> v2.4.1-rc3 2025-12-04T09:20:36.1366881Z * [new tag] v2.5.0 -> v2.5.0 2025-12-04T09:20:36.1368237Z * [new tag] v2.5.0-rc1 -> v2.5.0-rc1 2025-12-04T09:20:36.1369408Z * [new tag] v2.5.0-rc10 -> v2.5.0-rc10 2025-12-04T09:20:36.1370777Z * [new tag] v2.5.0-rc2 -> v2.5.0-rc2 2025-12-04T09:20:36.1372354Z * [new tag] v2.5.0-rc3 -> v2.5.0-rc3 2025-12-04T09:20:36.1373777Z * [new tag] v2.5.0-rc4 -> v2.5.0-rc4 2025-12-04T09:20:36.1375098Z * [new tag] v2.5.0-rc5 -> v2.5.0-rc5 2025-12-04T09:20:36.1376644Z * [new tag] v2.5.0-rc6 -> v2.5.0-rc6 2025-12-04T09:20:36.1378010Z * [new tag] v2.5.0-rc7 -> v2.5.0-rc7 2025-12-04T09:20:36.1379402Z * [new tag] v2.5.0-rc8 -> v2.5.0-rc8 2025-12-04T09:20:36.1381061Z * [new tag] v2.5.0-rc9 -> v2.5.0-rc9 2025-12-04T09:20:36.1381904Z * [new tag] v2.5.1 -> v2.5.1 2025-12-04T09:20:36.1383199Z * [new tag] v2.5.1-rc1 -> v2.5.1-rc1 2025-12-04T09:20:36.1384321Z * [new tag] v2.6.0 -> v2.6.0 2025-12-04T09:20:36.1385792Z * [new tag] v2.6.0-rc1 -> v2.6.0-rc1 2025-12-04T09:20:36.1387173Z * [new tag] v2.6.0-rc2 -> v2.6.0-rc2 2025-12-04T09:20:36.1388540Z * [new tag] v2.6.0-rc3 -> v2.6.0-rc3 2025-12-04T09:20:36.1389874Z * [new tag] v2.6.0-rc4 -> v2.6.0-rc4 2025-12-04T09:20:36.1391465Z * [new tag] v2.6.0-rc5 -> v2.6.0-rc5 2025-12-04T09:20:36.1392831Z * [new tag] v2.6.0-rc6 -> v2.6.0-rc6 2025-12-04T09:20:36.1394180Z * [new tag] v2.6.0-rc7 -> v2.6.0-rc7 2025-12-04T09:20:36.1395725Z * [new tag] v2.6.0-rc8 -> v2.6.0-rc8 2025-12-04T09:20:36.1397103Z * [new tag] v2.6.0-rc9 -> v2.6.0-rc9 2025-12-04T09:20:36.1398618Z * [new tag] v2.7.0 -> v2.7.0 2025-12-04T09:20:36.1399966Z * [new tag] v2.7.0-rc1 -> v2.7.0-rc1 2025-12-04T09:20:36.1401141Z * [new tag] v2.7.0-rc10 -> v2.7.0-rc10 2025-12-04T09:20:36.1402630Z * [new tag] v2.7.0-rc2 -> v2.7.0-rc2 2025-12-04T09:20:36.1404094Z * [new tag] v2.7.0-rc3 -> v2.7.0-rc3 2025-12-04T09:20:36.1405448Z * [new tag] v2.7.0-rc4 -> v2.7.0-rc4 2025-12-04T09:20:36.1406779Z * [new tag] v2.7.0-rc5 -> v2.7.0-rc5 2025-12-04T09:20:36.1408140Z * [new tag] v2.7.0-rc6 -> v2.7.0-rc6 2025-12-04T09:20:36.1409689Z * [new tag] v2.7.0-rc7 -> v2.7.0-rc7 2025-12-04T09:20:36.1411062Z * [new tag] v2.7.0-rc8 -> v2.7.0-rc8 2025-12-04T09:20:36.1412583Z * [new tag] v2.7.0-rc9 -> v2.7.0-rc9 2025-12-04T09:20:36.1413760Z * [new tag] v2.7.1 -> v2.7.1 2025-12-04T09:20:36.1415224Z * [new tag] v2.7.1-rc1 -> v2.7.1-rc1 2025-12-04T09:20:36.1416663Z * [new tag] v2.7.1-rc2 -> v2.7.1-rc2 2025-12-04T09:20:36.1418082Z * [new tag] v2.7.1-rc3 -> v2.7.1-rc3 2025-12-04T09:20:36.1419598Z * [new tag] v2.7.1-rc4 -> v2.7.1-rc4 2025-12-04T09:20:36.1420980Z * [new tag] v2.7.1-rc5 -> v2.7.1-rc5 2025-12-04T09:20:36.1422204Z * [new tag] v2.8.0 -> v2.8.0 2025-12-04T09:20:36.1423643Z * [new tag] v2.8.0-rc1 -> v2.8.0-rc1 2025-12-04T09:20:36.1425035Z * [new tag] v2.8.0-rc2 -> v2.8.0-rc2 2025-12-04T09:20:36.1426554Z * [new tag] v2.8.0-rc3 -> v2.8.0-rc3 2025-12-04T09:20:36.1427945Z * [new tag] v2.8.0-rc4 -> v2.8.0-rc4 2025-12-04T09:20:36.1429322Z * [new tag] v2.8.0-rc5 -> v2.8.0-rc5 2025-12-04T09:20:36.1430803Z * [new tag] v2.8.0-rc6 -> v2.8.0-rc6 2025-12-04T09:20:36.1432175Z * [new tag] v2.8.0-rc7 -> v2.8.0-rc7 2025-12-04T09:20:36.1433543Z * [new tag] v2.8.0-rc8 -> v2.8.0-rc8 2025-12-04T09:20:36.1435461Z * [new tag] v2.9.0 -> v2.9.0 2025-12-04T09:20:36.1436873Z * [new tag] v2.9.0-rc1 -> v2.9.0-rc1 2025-12-04T09:20:36.1438533Z * [new tag] v2.9.0-rc10 -> v2.9.0-rc10 2025-12-04T09:20:36.1439852Z * [new tag] v2.9.0-rc11 -> v2.9.0-rc11 2025-12-04T09:20:36.1441459Z * [new tag] v2.9.0-rc2 -> v2.9.0-rc2 2025-12-04T09:20:36.1442827Z * [new tag] v2.9.0-rc3 -> v2.9.0-rc3 2025-12-04T09:20:36.1444296Z * [new tag] v2.9.0-rc4 -> v2.9.0-rc4 2025-12-04T09:20:36.1445698Z * [new tag] v2.9.0-rc5 -> v2.9.0-rc5 2025-12-04T09:20:36.1447285Z * [new tag] v2.9.0-rc6 -> v2.9.0-rc6 2025-12-04T09:20:36.1448710Z * [new tag] v2.9.0-rc7 -> v2.9.0-rc7 2025-12-04T09:20:36.1450346Z * [new tag] v2.9.0-rc8 -> v2.9.0-rc8 2025-12-04T09:20:36.1451527Z * [new tag] v2.9.0-rc9 -> v2.9.0-rc9 2025-12-04T09:20:36.1452823Z * [new tag] v2.9.1 -> v2.9.1 2025-12-04T09:20:36.1454265Z * [new tag] v2.9.1-rc1 -> v2.9.1-rc1 2025-12-04T09:20:36.1455707Z * [new tag] v2.9.1-rc2 -> v2.9.1-rc2 2025-12-04T09:20:36.1457597Z * [new tag] viable/strict/1759343184 -> viable/strict/1759343184 2025-12-04T09:20:36.1458964Z * [new tag] viable/strict/1759346540 -> viable/strict/1759346540 2025-12-04T09:20:36.1460279Z * [new tag] viable/strict/1759348181 -> viable/strict/1759348181 2025-12-04T09:20:36.1461680Z * [new tag] viable/strict/1759350324 -> viable/strict/1759350324 2025-12-04T09:20:36.1462905Z * [new tag] viable/strict/1759351793 -> viable/strict/1759351793 2025-12-04T09:20:36.1464225Z * [new tag] viable/strict/1759353844 -> viable/strict/1759353844 2025-12-04T09:20:36.1465518Z * [new tag] viable/strict/1759355374 -> viable/strict/1759355374 2025-12-04T09:20:36.1466860Z * [new tag] viable/strict/1759357472 -> viable/strict/1759357472 2025-12-04T09:20:36.1468149Z * [new tag] viable/strict/1759361002 -> viable/strict/1759361002 2025-12-04T09:20:36.1469767Z * [new tag] viable/strict/1759362585 -> viable/strict/1759362585 2025-12-04T09:20:36.1471353Z * [new tag] viable/strict/1759365359 -> viable/strict/1759365359 2025-12-04T09:20:36.1474647Z * [new tag] viable/strict/1759370089 -> viable/strict/1759370089 2025-12-04T09:20:36.1475989Z * [new tag] viable/strict/1759377554 -> viable/strict/1759377554 2025-12-04T09:20:36.1477902Z * [new tag] viable/strict/1759379133 -> viable/strict/1759379133 2025-12-04T09:20:36.1479335Z * [new tag] viable/strict/1759389871 -> viable/strict/1759389871 2025-12-04T09:20:36.1480758Z * [new tag] viable/strict/1759393562 -> viable/strict/1759393562 2025-12-04T09:20:36.1482114Z * [new tag] viable/strict/1759395076 -> viable/strict/1759395076 2025-12-04T09:20:36.1483528Z * [new tag] viable/strict/1759398579 -> viable/strict/1759398579 2025-12-04T09:20:36.1484963Z * [new tag] viable/strict/1759404142 -> viable/strict/1759404142 2025-12-04T09:20:36.1486326Z * [new tag] viable/strict/1759405773 -> viable/strict/1759405773 2025-12-04T09:20:36.1487713Z * [new tag] viable/strict/1759408041 -> viable/strict/1759408041 2025-12-04T09:20:36.1489071Z * [new tag] viable/strict/1759411593 -> viable/strict/1759411593 2025-12-04T09:20:36.1490456Z * [new tag] viable/strict/1759427395 -> viable/strict/1759427395 2025-12-04T09:20:36.1491814Z * [new tag] viable/strict/1759434582 -> viable/strict/1759434582 2025-12-04T09:20:36.1493114Z * [new tag] viable/strict/1759436720 -> viable/strict/1759436720 2025-12-04T09:20:36.1494753Z * [new tag] viable/strict/1759440219 -> viable/strict/1759440219 2025-12-04T09:20:36.1496045Z * [new tag] viable/strict/1759441948 -> viable/strict/1759441948 2025-12-04T09:20:36.1497417Z * [new tag] viable/strict/1759443860 -> viable/strict/1759443860 2025-12-04T09:20:36.1498775Z * [new tag] viable/strict/1759445377 -> viable/strict/1759445377 2025-12-04T09:20:36.1500313Z * [new tag] viable/strict/1759447415 -> viable/strict/1759447415 2025-12-04T09:20:36.1501670Z * [new tag] viable/strict/1759451750 -> viable/strict/1759451750 2025-12-04T09:20:36.1503079Z * [new tag] viable/strict/1759453910 -> viable/strict/1759453910 2025-12-04T09:20:36.1504502Z * [new tag] viable/strict/1759456483 -> viable/strict/1759456483 2025-12-04T09:20:36.1505944Z * [new tag] viable/strict/1759459279 -> viable/strict/1759459279 2025-12-04T09:20:36.1507392Z * [new tag] viable/strict/1759460742 -> viable/strict/1759460742 2025-12-04T09:20:36.1508747Z * [new tag] viable/strict/1759462025 -> viable/strict/1759462025 2025-12-04T09:20:36.1510241Z * [new tag] viable/strict/1759469086 -> viable/strict/1759469086 2025-12-04T09:20:36.1511640Z * [new tag] viable/strict/1759470581 -> viable/strict/1759470581 2025-12-04T09:20:36.1513019Z * [new tag] viable/strict/1759472786 -> viable/strict/1759472786 2025-12-04T09:20:36.1514402Z * [new tag] viable/strict/1759476294 -> viable/strict/1759476294 2025-12-04T09:20:36.1515775Z * [new tag] viable/strict/1759479963 -> viable/strict/1759479963 2025-12-04T09:20:36.1517159Z * [new tag] viable/strict/1759492177 -> viable/strict/1759492177 2025-12-04T09:20:36.1518550Z * [new tag] viable/strict/1759519278 -> viable/strict/1759519278 2025-12-04T09:20:36.1519904Z * [new tag] viable/strict/1759524580 -> viable/strict/1759524580 2025-12-04T09:20:36.1521229Z * [new tag] viable/strict/1759528193 -> viable/strict/1759528193 2025-12-04T09:20:36.1522811Z * [new tag] viable/strict/1759533797 -> viable/strict/1759533797 2025-12-04T09:20:36.1524196Z * [new tag] viable/strict/1759542780 -> viable/strict/1759542780 2025-12-04T09:20:36.1525605Z * [new tag] viable/strict/1759549779 -> viable/strict/1759549779 2025-12-04T09:20:36.1527023Z * [new tag] viable/strict/1759555455 -> viable/strict/1759555455 2025-12-04T09:20:36.1528437Z * [new tag] viable/strict/1759559176 -> viable/strict/1759559176 2025-12-04T09:20:36.1529832Z * [new tag] viable/strict/1759560629 -> viable/strict/1759560629 2025-12-04T09:20:36.1531182Z * [new tag] viable/strict/1759569848 -> viable/strict/1759569848 2025-12-04T09:20:36.1532802Z * [new tag] viable/strict/1759571382 -> viable/strict/1759571382 2025-12-04T09:20:36.1534225Z * [new tag] viable/strict/1759573474 -> viable/strict/1759573474 2025-12-04T09:20:36.1535565Z * [new tag] viable/strict/1759618187 -> viable/strict/1759618187 2025-12-04T09:20:36.1536950Z * [new tag] viable/strict/1759626742 -> viable/strict/1759626742 2025-12-04T09:20:36.1538389Z * [new tag] viable/strict/1759632427 -> viable/strict/1759632427 2025-12-04T09:20:36.1540023Z * [new tag] viable/strict/1759634971 -> viable/strict/1759634971 2025-12-04T09:20:36.1541340Z * [new tag] viable/strict/1759661382 -> viable/strict/1759661382 2025-12-04T09:20:36.1542635Z * [new tag] viable/strict/1759663294 -> viable/strict/1759663294 2025-12-04T09:20:36.1543900Z * [new tag] viable/strict/1759708178 -> viable/strict/1759708178 2025-12-04T09:20:36.1545482Z * [new tag] viable/strict/1759715695 -> viable/strict/1759715695 2025-12-04T09:20:36.1546726Z * [new tag] viable/strict/1759728293 -> viable/strict/1759728293 2025-12-04T09:20:36.1548087Z * [new tag] viable/strict/1759735513 -> viable/strict/1759735513 2025-12-04T09:20:36.1549485Z * [new tag] viable/strict/1759739177 -> viable/strict/1759739177 2025-12-04T09:20:36.1550927Z * [new tag] viable/strict/1759758635 -> viable/strict/1759758635 2025-12-04T09:20:36.1552349Z * [new tag] viable/strict/1759765784 -> viable/strict/1759765784 2025-12-04T09:20:36.1553791Z * [new tag] viable/strict/1759767948 -> viable/strict/1759767948 2025-12-04T09:20:36.1555169Z * [new tag] viable/strict/1759771461 -> viable/strict/1759771461 2025-12-04T09:20:36.1556466Z * [new tag] viable/strict/1759776706 -> viable/strict/1759776706 2025-12-04T09:20:36.1557827Z * [new tag] viable/strict/1759782317 -> viable/strict/1759782317 2025-12-04T09:20:36.1559283Z * [new tag] viable/strict/1759783777 -> viable/strict/1759783777 2025-12-04T09:20:36.1560759Z * [new tag] viable/strict/1759785815 -> viable/strict/1759785815 2025-12-04T09:20:36.1562242Z * [new tag] viable/strict/1759789459 -> viable/strict/1759789459 2025-12-04T09:20:36.1563512Z * [new tag] viable/strict/1759790974 -> viable/strict/1759790974 2025-12-04T09:20:36.1564774Z * [new tag] viable/strict/1759794583 -> viable/strict/1759794583 2025-12-04T09:20:36.1566120Z * [new tag] viable/strict/1759797408 -> viable/strict/1759797408 2025-12-04T09:20:36.1567990Z * [new tag] viable/strict/1759799518 -> viable/strict/1759799518 2025-12-04T09:20:36.1569344Z * [new tag] viable/strict/1759804909 -> viable/strict/1759804909 2025-12-04T09:20:36.1570746Z * [new tag] viable/strict/1759807643 -> viable/strict/1759807643 2025-12-04T09:20:36.1572445Z * [new tag] viable/strict/1759809089 -> viable/strict/1759809089 2025-12-04T09:20:36.1573802Z * [new tag] viable/strict/1759811145 -> viable/strict/1759811145 2025-12-04T09:20:36.1575210Z * [new tag] viable/strict/1759812581 -> viable/strict/1759812581 2025-12-04T09:20:36.1576704Z * [new tag] viable/strict/1759814683 -> viable/strict/1759814683 2025-12-04T09:20:36.1578091Z * [new tag] viable/strict/1759821889 -> viable/strict/1759821889 2025-12-04T09:20:36.1579601Z * [new tag] viable/strict/1759823376 -> viable/strict/1759823376 2025-12-04T09:20:36.1581031Z * [new tag] viable/strict/1759827107 -> viable/strict/1759827107 2025-12-04T09:20:36.1582354Z * [new tag] viable/strict/1759830577 -> viable/strict/1759830577 2025-12-04T09:20:36.1583846Z * [new tag] viable/strict/1759832720 -> viable/strict/1759832720 2025-12-04T09:20:36.1585321Z * [new tag] viable/strict/1759842063 -> viable/strict/1759842063 2025-12-04T09:20:36.1586676Z * [new tag] viable/strict/1759847121 -> viable/strict/1759847121 2025-12-04T09:20:36.1588286Z * [new tag] viable/strict/1759850721 -> viable/strict/1759850721 2025-12-04T09:20:36.1589646Z * [new tag] viable/strict/1759857870 -> viable/strict/1759857870 2025-12-04T09:20:36.1591055Z * [new tag] viable/strict/1759863143 -> viable/strict/1759863143 2025-12-04T09:20:36.1592473Z * [new tag] viable/strict/1759875874 -> viable/strict/1759875874 2025-12-04T09:20:36.1593731Z * [new tag] viable/strict/1759877385 -> viable/strict/1759877385 2025-12-04T09:20:36.1595132Z * [new tag] viable/strict/1759883801 -> viable/strict/1759883801 2025-12-04T09:20:36.1596658Z * [new tag] viable/strict/1759885922 -> viable/strict/1759885922 2025-12-04T09:20:36.1597912Z * [new tag] viable/strict/1759888488 -> viable/strict/1759888488 2025-12-04T09:20:36.1599243Z * [new tag] viable/strict/1759895471 -> viable/strict/1759895471 2025-12-04T09:20:36.1600615Z * [new tag] viable/strict/1759904803 -> viable/strict/1759904803 2025-12-04T09:20:36.1602129Z * [new tag] viable/strict/1759908300 -> viable/strict/1759908300 2025-12-04T09:20:36.1603545Z * [new tag] viable/strict/1759915520 -> viable/strict/1759915520 2025-12-04T09:20:36.1604936Z * [new tag] viable/strict/1759916978 -> viable/strict/1759916978 2025-12-04T09:20:36.1606190Z * [new tag] viable/strict/1759930024 -> viable/strict/1759930024 2025-12-04T09:20:36.1607586Z * [new tag] viable/strict/1759948122 -> viable/strict/1759948122 2025-12-04T09:20:36.1608955Z * [new tag] viable/strict/1759952983 -> viable/strict/1759952983 2025-12-04T09:20:36.1610446Z * [new tag] viable/strict/1759955121 -> viable/strict/1759955121 2025-12-04T09:20:36.1611820Z * [new tag] viable/strict/1759962298 -> viable/strict/1759962298 2025-12-04T09:20:36.1613245Z * [new tag] viable/strict/1759965837 -> viable/strict/1759965837 2025-12-04T09:20:36.1614672Z * [new tag] viable/strict/1759970213 -> viable/strict/1759970213 2025-12-04T09:20:36.1616048Z * [new tag] viable/strict/1759974894 -> viable/strict/1759974894 2025-12-04T09:20:36.1617413Z * [new tag] viable/strict/1759977763 -> viable/strict/1759977763 2025-12-04T09:20:36.1618864Z * [new tag] viable/strict/1759979241 -> viable/strict/1759979241 2025-12-04T09:20:36.1620354Z * [new tag] viable/strict/1759985417 -> viable/strict/1759985417 2025-12-04T09:20:36.1621705Z * [new tag] viable/strict/1759987490 -> viable/strict/1759987490 2025-12-04T09:20:36.1623117Z * [new tag] viable/strict/1759996180 -> viable/strict/1759996180 2025-12-04T09:20:36.1624475Z * [new tag] viable/strict/1760065682 -> viable/strict/1760065682 2025-12-04T09:20:36.1625991Z * [new tag] viable/strict/1760066894 -> viable/strict/1760066894 2025-12-04T09:20:36.1627405Z * [new tag] viable/strict/1760070345 -> viable/strict/1760070345 2025-12-04T09:20:36.1628820Z * [new tag] viable/strict/1760089782 -> viable/strict/1760089782 2025-12-04T09:20:36.1630260Z * [new tag] viable/strict/1760091921 -> viable/strict/1760091921 2025-12-04T09:20:36.1631682Z * [new tag] viable/strict/1760127924 -> viable/strict/1760127924 2025-12-04T09:20:36.1633019Z * [new tag] viable/strict/1760129489 -> viable/strict/1760129489 2025-12-04T09:20:36.1634533Z * [new tag] viable/strict/1760132980 -> viable/strict/1760132980 2025-12-04T09:20:36.1635886Z * [new tag] viable/strict/1760135060 -> viable/strict/1760135060 2025-12-04T09:20:36.1637306Z * [new tag] viable/strict/1760215782 -> viable/strict/1760215782 2025-12-04T09:20:36.1638689Z * [new tag] viable/strict/1760273849 -> viable/strict/1760273849 2025-12-04T09:20:36.1640096Z * [new tag] viable/strict/1760275517 -> viable/strict/1760275517 2025-12-04T09:20:36.1641458Z * [new tag] viable/strict/1760276979 -> viable/strict/1760276979 2025-12-04T09:20:36.1642845Z * [new tag] viable/strict/1760279007 -> viable/strict/1760279007 2025-12-04T09:20:36.1644100Z * [new tag] viable/strict/1760286328 -> viable/strict/1760286328 2025-12-04T09:20:36.1645438Z * [new tag] viable/strict/1760493304 -> viable/strict/1760493304 2025-12-04T09:20:36.1646871Z * [new tag] viable/strict/1760496298 -> viable/strict/1760496298 2025-12-04T09:20:36.1648254Z * [new tag] viable/strict/1760518396 -> viable/strict/1760518396 2025-12-04T09:20:36.1649638Z * [new tag] viable/strict/1760534864 -> viable/strict/1760534864 2025-12-04T09:20:36.1651082Z * [new tag] viable/strict/1760549062 -> viable/strict/1760549062 2025-12-04T09:20:36.1652589Z * [new tag] viable/strict/1760552799 -> viable/strict/1760552799 2025-12-04T09:20:36.1654011Z * [new tag] viable/strict/1760554355 -> viable/strict/1760554355 2025-12-04T09:20:36.1655473Z * [new tag] viable/strict/1760556275 -> viable/strict/1760556275 2025-12-04T09:20:36.1656884Z * [new tag] viable/strict/1760564979 -> viable/strict/1760564979 2025-12-04T09:20:36.1658653Z * [new tag] viable/strict/1760567049 -> viable/strict/1760567049 2025-12-04T09:20:36.1660534Z * [new tag] viable/strict/1760568585 -> viable/strict/1760568585 2025-12-04T09:20:36.1661926Z * [new tag] viable/strict/1760570630 -> viable/strict/1760570630 2025-12-04T09:20:36.1663230Z * [new tag] viable/strict/1760572180 -> viable/strict/1760572180 2025-12-04T09:20:36.1664592Z * [new tag] viable/strict/1760575094 -> viable/strict/1760575094 2025-12-04T09:20:36.1666072Z * [new tag] viable/strict/1760579709 -> viable/strict/1760579709 2025-12-04T09:20:36.1667877Z * [new tag] viable/strict/1760582614 -> viable/strict/1760582614 2025-12-04T09:20:36.1669321Z * [new tag] viable/strict/1760586815 -> viable/strict/1760586815 2025-12-04T09:20:36.1670578Z * [new tag] viable/strict/1760588829 -> viable/strict/1760588829 2025-12-04T09:20:36.1672210Z * [new tag] viable/strict/1760590200 -> viable/strict/1760590200 2025-12-04T09:20:36.1673668Z * [new tag] viable/strict/1760592311 -> viable/strict/1760592311 2025-12-04T09:20:36.1675091Z * [new tag] viable/strict/1760619733 -> viable/strict/1760619733 2025-12-04T09:20:36.1676444Z * [new tag] viable/strict/1760628335 -> viable/strict/1760628335 2025-12-04T09:20:36.1677813Z * [new tag] viable/strict/1760635490 -> viable/strict/1760635490 2025-12-04T09:20:36.1679179Z * [new tag] viable/strict/1760640743 -> viable/strict/1760640743 2025-12-04T09:20:36.1680593Z * [new tag] viable/strict/1760642528 -> viable/strict/1760642528 2025-12-04T09:20:36.1681964Z * [new tag] viable/strict/1760646330 -> viable/strict/1760646330 2025-12-04T09:20:36.1683248Z * [new tag] viable/strict/1760666101 -> viable/strict/1760666101 2025-12-04T09:20:36.1684646Z * [new tag] viable/strict/1760668990 -> viable/strict/1760668990 2025-12-04T09:20:36.1686008Z * [new tag] viable/strict/1760670600 -> viable/strict/1760670600 2025-12-04T09:20:36.1687418Z * [new tag] viable/strict/1760671704 -> viable/strict/1760671704 2025-12-04T09:20:36.1688828Z * [new tag] viable/strict/1760673121 -> viable/strict/1760673121 2025-12-04T09:20:36.1690215Z * [new tag] viable/strict/1760675352 -> viable/strict/1760675352 2025-12-04T09:20:36.1691605Z * [new tag] viable/strict/1760696731 -> viable/strict/1760696731 2025-12-04T09:20:36.1694120Z * [new tag] viable/strict/1760723515 -> viable/strict/1760723515 2025-12-04T09:20:36.1695528Z * [new tag] viable/strict/1760727234 -> viable/strict/1760727234 2025-12-04T09:20:36.1696949Z * [new tag] viable/strict/1760730578 -> viable/strict/1760730578 2025-12-04T09:20:36.1698359Z * [new tag] viable/strict/1760732726 -> viable/strict/1760732726 2025-12-04T09:20:36.1699978Z * [new tag] viable/strict/1760734180 -> viable/strict/1760734180 2025-12-04T09:20:36.1701375Z * [new tag] viable/strict/1760736251 -> viable/strict/1760736251 2025-12-04T09:20:36.1702857Z * [new tag] viable/strict/1760737772 -> viable/strict/1760737772 2025-12-04T09:20:36.1704254Z * [new tag] viable/strict/1760758005 -> viable/strict/1760758005 2025-12-04T09:20:36.1705676Z * [new tag] viable/strict/1760761532 -> viable/strict/1760761532 2025-12-04T09:20:36.1707085Z * [new tag] viable/strict/1760802581 -> viable/strict/1760802581 2025-12-04T09:20:36.1708492Z * [new tag] viable/strict/1760827772 -> viable/strict/1760827772 2025-12-04T09:20:36.1709849Z * [new tag] viable/strict/1760834524 -> viable/strict/1760834524 2025-12-04T09:20:36.1711257Z * [new tag] viable/strict/1760845009 -> viable/strict/1760845009 2025-12-04T09:20:36.1712809Z * [new tag] viable/strict/1760876836 -> viable/strict/1760876836 2025-12-04T09:20:36.1714168Z * [new tag] viable/strict/1760880329 -> viable/strict/1760880329 2025-12-04T09:20:36.1721293Z * [new tag] viable/strict/1760888987 -> viable/strict/1760888987 2025-12-04T09:20:36.1721613Z * [new tag] viable/strict/1760912664 -> viable/strict/1760912664 2025-12-04T09:20:36.1721777Z * [new tag] viable/strict/1760925321 -> viable/strict/1760925321 2025-12-04T09:20:36.1721925Z * [new tag] viable/strict/1760931488 -> viable/strict/1760931488 2025-12-04T09:20:36.1722065Z * [new tag] viable/strict/1760932693 -> viable/strict/1760932693 2025-12-04T09:20:36.1722298Z * [new tag] viable/strict/1761004184 -> viable/strict/1761004184 2025-12-04T09:20:36.1723892Z * [new tag] viable/strict/1761014748 -> viable/strict/1761014748 2025-12-04T09:20:36.1725319Z * [new tag] viable/strict/1761017491 -> viable/strict/1761017491 2025-12-04T09:20:36.1726718Z * [new tag] viable/strict/1761018806 -> viable/strict/1761018806 2025-12-04T09:20:36.1728151Z * [new tag] viable/strict/1761020754 -> viable/strict/1761020754 2025-12-04T09:20:36.1729563Z * [new tag] viable/strict/1761024303 -> viable/strict/1761024303 2025-12-04T09:20:36.1730950Z * [new tag] viable/strict/1761029582 -> viable/strict/1761029582 2025-12-04T09:20:36.1732327Z * [new tag] viable/strict/1761031535 -> viable/strict/1761031535 2025-12-04T09:20:36.1733745Z * [new tag] viable/strict/1761035196 -> viable/strict/1761035196 2025-12-04T09:20:36.1735181Z * [new tag] viable/strict/1761045825 -> viable/strict/1761045825 2025-12-04T09:20:36.1736647Z * [new tag] viable/strict/1761054796 -> viable/strict/1761054796 2025-12-04T09:20:36.1738036Z * [new tag] viable/strict/1761060314 -> viable/strict/1761060314 2025-12-04T09:20:36.1739550Z * [new tag] viable/strict/1761071198 -> viable/strict/1761071198 2025-12-04T09:20:36.1740990Z * [new tag] viable/strict/1761074628 -> viable/strict/1761074628 2025-12-04T09:20:36.1742403Z * [new tag] viable/strict/1761078351 -> viable/strict/1761078351 2025-12-04T09:20:36.1743755Z * [new tag] viable/strict/1761079822 -> viable/strict/1761079822 2025-12-04T09:20:36.1745111Z * [new tag] viable/strict/1761081873 -> viable/strict/1761081873 2025-12-04T09:20:36.1746512Z * [new tag] viable/strict/1761083392 -> viable/strict/1761083392 2025-12-04T09:20:36.1747961Z * [new tag] viable/strict/1761085465 -> viable/strict/1761085465 2025-12-04T09:20:36.1749390Z * [new tag] viable/strict/1761089099 -> viable/strict/1761089099 2025-12-04T09:20:36.1751291Z * [new tag] viable/strict/1761095535 -> viable/strict/1761095535 2025-12-04T09:20:36.1752677Z * [new tag] viable/strict/1761098119 -> viable/strict/1761098119 2025-12-04T09:20:36.1754415Z * [new tag] viable/strict/1761101330 -> viable/strict/1761101330 2025-12-04T09:20:36.1755813Z * [new tag] viable/strict/1761114425 -> viable/strict/1761114425 2025-12-04T09:20:36.1757201Z * [new tag] viable/strict/1761116036 -> viable/strict/1761116036 2025-12-04T09:20:36.1758583Z * [new tag] viable/strict/1761119379 -> viable/strict/1761119379 2025-12-04T09:20:36.1760002Z * [new tag] viable/strict/1761121601 -> viable/strict/1761121601 2025-12-04T09:20:36.1761372Z * [new tag] viable/strict/1761123234 -> viable/strict/1761123234 2025-12-04T09:20:36.1762637Z * [new tag] viable/strict/1761126621 -> viable/strict/1761126621 2025-12-04T09:20:36.1764076Z * [new tag] viable/strict/1761132259 -> viable/strict/1761132259 2025-12-04T09:20:36.1765498Z * [new tag] viable/strict/1761146746 -> viable/strict/1761146746 2025-12-04T09:20:36.1766879Z * [new tag] viable/strict/1761164752 -> viable/strict/1761164752 2025-12-04T09:20:36.1768296Z * [new tag] viable/strict/1761166198 -> viable/strict/1761166198 2025-12-04T09:20:36.1769735Z * [new tag] viable/strict/1761175424 -> viable/strict/1761175424 2025-12-04T09:20:36.1771289Z * [new tag] viable/strict/1761176983 -> viable/strict/1761176983 2025-12-04T09:20:36.1772913Z * [new tag] viable/strict/1761179891 -> viable/strict/1761179891 2025-12-04T09:20:36.1774373Z * [new tag] viable/strict/1761181930 -> viable/strict/1761181930 2025-12-04T09:20:36.1775789Z * [new tag] viable/strict/1761184516 -> viable/strict/1761184516 2025-12-04T09:20:36.1777252Z * [new tag] viable/strict/1761190179 -> viable/strict/1761190179 2025-12-04T09:20:36.1778618Z * [new tag] viable/strict/1761193558 -> viable/strict/1761193558 2025-12-04T09:20:36.1780109Z * [new tag] viable/strict/1761207990 -> viable/strict/1761207990 2025-12-04T09:20:36.1781573Z * [new tag] viable/strict/1761229539 -> viable/strict/1761229539 2025-12-04T09:20:36.1783146Z * [new tag] viable/strict/1761244031 -> viable/strict/1761244031 2025-12-04T09:20:36.1784542Z * [new tag] viable/strict/1761248986 -> viable/strict/1761248986 2025-12-04T09:20:36.1785917Z * [new tag] viable/strict/1761259791 -> viable/strict/1761259791 2025-12-04T09:20:36.1787287Z * [new tag] viable/strict/1761266139 -> viable/strict/1761266139 2025-12-04T09:20:36.1788798Z * [new tag] viable/strict/1761268316 -> viable/strict/1761268316 2025-12-04T09:20:36.1790164Z * [new tag] viable/strict/1761273805 -> viable/strict/1761273805 2025-12-04T09:20:36.1791536Z * [new tag] viable/strict/1761275261 -> viable/strict/1761275261 2025-12-04T09:20:36.1792953Z * [new tag] viable/strict/1761277913 -> viable/strict/1761277913 2025-12-04T09:20:36.1794445Z * [new tag] viable/strict/1761290701 -> viable/strict/1761290701 2025-12-04T09:20:36.1795840Z * [new tag] viable/strict/1761294396 -> viable/strict/1761294396 2025-12-04T09:20:36.1797231Z * [new tag] viable/strict/1761303047 -> viable/strict/1761303047 2025-12-04T09:20:36.1798608Z * [new tag] viable/strict/1761335388 -> viable/strict/1761335388 2025-12-04T09:20:36.1799995Z * [new tag] viable/strict/1761337551 -> viable/strict/1761337551 2025-12-04T09:20:36.1801630Z * [new tag] viable/strict/1761339007 -> viable/strict/1761339007 2025-12-04T09:20:36.1802927Z * [new tag] viable/strict/1761341050 -> viable/strict/1761341050 2025-12-04T09:20:36.1804264Z * [new tag] viable/strict/1761346188 -> viable/strict/1761346188 2025-12-04T09:20:36.1805753Z * [new tag] viable/strict/1761349792 -> viable/strict/1761349792 2025-12-04T09:20:36.1807129Z * [new tag] viable/strict/1761352620 -> viable/strict/1761352620 2025-12-04T09:20:36.1808523Z * [new tag] viable/strict/1761354730 -> viable/strict/1761354730 2025-12-04T09:20:36.1809986Z * [new tag] viable/strict/1761357298 -> viable/strict/1761357298 2025-12-04T09:20:36.1811355Z * [new tag] viable/strict/1761360201 -> viable/strict/1761360201 2025-12-04T09:20:36.1812737Z * [new tag] viable/strict/1761361753 -> viable/strict/1761361753 2025-12-04T09:20:36.1814210Z * [new tag] viable/strict/1761364351 -> viable/strict/1761364351 2025-12-04T09:20:36.1815619Z * [new tag] viable/strict/1761366338 -> viable/strict/1761366338 2025-12-04T09:20:36.1817107Z * [new tag] viable/strict/1761367802 -> viable/strict/1761367802 2025-12-04T09:20:36.1818509Z * [new tag] viable/strict/1761369889 -> viable/strict/1761369889 2025-12-04T09:20:36.1820047Z * [new tag] viable/strict/1761371385 -> viable/strict/1761371385 2025-12-04T09:20:36.1821630Z * [new tag] viable/strict/1761373581 -> viable/strict/1761373581 2025-12-04T09:20:36.1823065Z * [new tag] viable/strict/1761375054 -> viable/strict/1761375054 2025-12-04T09:20:36.1824500Z * [new tag] viable/strict/1761421785 -> viable/strict/1761421785 2025-12-04T09:20:36.1825961Z * [new tag] viable/strict/1761434614 -> viable/strict/1761434614 2025-12-04T09:20:36.1827688Z * [new tag] viable/strict/1761439254 -> viable/strict/1761439254 2025-12-04T09:20:36.1829082Z * [new tag] viable/strict/1761454187 -> viable/strict/1761454187 2025-12-04T09:20:36.1830519Z * [new tag] viable/strict/1761459991 -> viable/strict/1761459991 2025-12-04T09:20:36.1832250Z * [new tag] viable/strict/1761470668 -> viable/strict/1761470668 2025-12-04T09:20:36.1834027Z * [new tag] viable/strict/1761472188 -> viable/strict/1761472188 2025-12-04T09:20:36.1835409Z * [new tag] viable/strict/1761503178 -> viable/strict/1761503178 2025-12-04T09:20:36.1836846Z * [new tag] viable/strict/1761517492 -> viable/strict/1761517492 2025-12-04T09:20:36.1838265Z * [new tag] viable/strict/1761518981 -> viable/strict/1761518981 2025-12-04T09:20:36.1839814Z * [new tag] viable/strict/1761533609 -> viable/strict/1761533609 2025-12-04T09:20:36.1841056Z * [new tag] viable/strict/1761546438 -> viable/strict/1761546438 2025-12-04T09:20:36.1842523Z * [new tag] viable/strict/1761548133 -> viable/strict/1761548133 2025-12-04T09:20:36.1844668Z * [new tag] viable/strict/1761555186 -> viable/strict/1761555186 2025-12-04T09:20:36.1846154Z * [new tag] viable/strict/1761557178 -> viable/strict/1761557178 2025-12-04T09:20:36.1847549Z * [new tag] viable/strict/1761560772 -> viable/strict/1761560772 2025-12-04T09:20:36.1849029Z * [new tag] viable/strict/1761562266 -> viable/strict/1761562266 2025-12-04T09:20:36.1850559Z * [new tag] viable/strict/1761564260 -> viable/strict/1761564260 2025-12-04T09:20:36.1852010Z * [new tag] viable/strict/1761568072 -> viable/strict/1761568072 2025-12-04T09:20:36.1853447Z * [new tag] viable/strict/1761571683 -> viable/strict/1761571683 2025-12-04T09:20:36.1854809Z * [new tag] viable/strict/1761580199 -> viable/strict/1761580199 2025-12-04T09:20:36.1856193Z * [new tag] viable/strict/1761587383 -> viable/strict/1761587383 2025-12-04T09:20:36.1857718Z * [new tag] viable/strict/1761591165 -> viable/strict/1761591165 2025-12-04T09:20:36.1859171Z * [new tag] viable/strict/1761594575 -> viable/strict/1761594575 2025-12-04T09:20:36.1860665Z * [new tag] viable/strict/1761596710 -> viable/strict/1761596710 2025-12-04T09:20:36.1862118Z * [new tag] viable/strict/1761598189 -> viable/strict/1761598189 2025-12-04T09:20:36.1863581Z * [new tag] viable/strict/1761600254 -> viable/strict/1761600254 2025-12-04T09:20:36.1865009Z * [new tag] viable/strict/1761603879 -> viable/strict/1761603879 2025-12-04T09:20:36.1866417Z * [new tag] viable/strict/1761605429 -> viable/strict/1761605429 2025-12-04T09:20:36.1867948Z * [new tag] viable/strict/1761607468 -> viable/strict/1761607468 2025-12-04T09:20:36.1869365Z * [new tag] viable/strict/1761608983 -> viable/strict/1761608983 2025-12-04T09:20:36.1870790Z * [new tag] viable/strict/1761611846 -> viable/strict/1761611846 2025-12-04T09:20:36.1874205Z * [new tag] viable/strict/1761613922 -> viable/strict/1761613922 2025-12-04T09:20:36.1875532Z * [new tag] viable/strict/1761616504 -> viable/strict/1761616504 2025-12-04T09:20:36.1876780Z * [new tag] viable/strict/1761619599 -> viable/strict/1761619599 2025-12-04T09:20:36.1878196Z * [new tag] viable/strict/1761686693 -> viable/strict/1761686693 2025-12-04T09:20:36.1879615Z * [new tag] viable/strict/1761688179 -> viable/strict/1761688179 2025-12-04T09:20:36.1881067Z * [new tag] viable/strict/1761691973 -> viable/strict/1761691973 2025-12-04T09:20:36.1882607Z * [new tag] viable/strict/1761693884 -> viable/strict/1761693884 2025-12-04T09:20:36.1884083Z * [new tag] viable/strict/1761695389 -> viable/strict/1761695389 2025-12-04T09:20:36.1885490Z * [new tag] viable/strict/1761698408 -> viable/strict/1761698408 2025-12-04T09:20:36.1886988Z * [new tag] viable/strict/1761702931 -> viable/strict/1761702931 2025-12-04T09:20:36.1888357Z * [new tag] viable/strict/1761706307 -> viable/strict/1761706307 2025-12-04T09:20:36.1889821Z * [new tag] viable/strict/1761709065 -> viable/strict/1761709065 2025-12-04T09:20:36.1891297Z * [new tag] viable/strict/1761710285 -> viable/strict/1761710285 2025-12-04T09:20:36.1892846Z * [new tag] viable/strict/1761711983 -> viable/strict/1761711983 2025-12-04T09:20:36.1894325Z * [new tag] viable/strict/1761713514 -> viable/strict/1761713514 2025-12-04T09:20:36.1895874Z * [new tag] viable/strict/1761715523 -> viable/strict/1761715523 2025-12-04T09:20:36.1897365Z * [new tag] viable/strict/1761727973 -> viable/strict/1761727973 2025-12-04T09:20:36.1898910Z * [new tag] viable/strict/1761751558 -> viable/strict/1761751558 2025-12-04T09:20:36.1900451Z * [new tag] viable/strict/1761755187 -> viable/strict/1761755187 2025-12-04T09:20:36.1901895Z * [new tag] viable/strict/1761756826 -> viable/strict/1761756826 2025-12-04T09:20:36.1903278Z * [new tag] viable/strict/1761769551 -> viable/strict/1761769551 2025-12-04T09:20:36.1904882Z * [new tag] viable/strict/1761771032 -> viable/strict/1761771032 2025-12-04T09:20:36.1906168Z * [new tag] viable/strict/1761773101 -> viable/strict/1761773101 2025-12-04T09:20:36.1907657Z * [new tag] viable/strict/1761781792 -> viable/strict/1761781792 2025-12-04T09:20:36.1909334Z * [new tag] viable/strict/1761784788 -> viable/strict/1761784788 2025-12-04T09:20:36.1910795Z * [new tag] viable/strict/1761786740 -> viable/strict/1761786740 2025-12-04T09:20:36.1912227Z * [new tag] viable/strict/1761789332 -> viable/strict/1761789332 2025-12-04T09:20:36.1914059Z * [new tag] viable/strict/1761792569 -> viable/strict/1761792569 2025-12-04T09:20:36.1915519Z * [new tag] viable/strict/1761795289 -> viable/strict/1761795289 2025-12-04T09:20:36.1916973Z * [new tag] viable/strict/1761798345 -> viable/strict/1761798345 2025-12-04T09:20:36.1918470Z * [new tag] viable/strict/1761799827 -> viable/strict/1761799827 2025-12-04T09:20:36.1919991Z * [new tag] viable/strict/1761805604 -> viable/strict/1761805604 2025-12-04T09:20:36.1921427Z * [new tag] viable/strict/1761807202 -> viable/strict/1761807202 2025-12-04T09:20:36.1922887Z * [new tag] viable/strict/1761809094 -> viable/strict/1761809094 2025-12-04T09:20:36.1924330Z * [new tag] viable/strict/1761810576 -> viable/strict/1761810576 2025-12-04T09:20:36.1925892Z * [new tag] viable/strict/1761812771 -> viable/strict/1761812771 2025-12-04T09:20:36.1927367Z * [new tag] viable/strict/1761814363 -> viable/strict/1761814363 2025-12-04T09:20:36.1928831Z * [new tag] viable/strict/1761857410 -> viable/strict/1761857410 2025-12-04T09:20:36.1930313Z * [new tag] viable/strict/1761860985 -> viable/strict/1761860985 2025-12-04T09:20:36.1931797Z * [new tag] viable/strict/1761863094 -> viable/strict/1761863094 2025-12-04T09:20:36.1933227Z * [new tag] viable/strict/1761864590 -> viable/strict/1761864590 2025-12-04T09:20:36.1934690Z * [new tag] viable/strict/1761866675 -> viable/strict/1761866675 2025-12-04T09:20:36.1936288Z * [new tag] viable/strict/1761868178 -> viable/strict/1761868178 2025-12-04T09:20:36.1937818Z * [new tag] viable/strict/1761871111 -> viable/strict/1761871111 2025-12-04T09:20:36.1939320Z * [new tag] viable/strict/1761873126 -> viable/strict/1761873126 2025-12-04T09:20:36.1941316Z * [new tag] viable/strict/1761875714 -> viable/strict/1761875714 2025-12-04T09:20:36.1943194Z * [new tag] viable/strict/1761878924 -> viable/strict/1761878924 2025-12-04T09:20:36.1944558Z * [new tag] viable/strict/1761881727 -> viable/strict/1761881727 2025-12-04T09:20:36.1945740Z * [new tag] viable/strict/1761882959 -> viable/strict/1761882959 2025-12-04T09:20:36.1947218Z * [new tag] viable/strict/1761886268 -> viable/strict/1761886268 2025-12-04T09:20:36.1948628Z * [new tag] viable/strict/1761893641 -> viable/strict/1761893641 2025-12-04T09:20:36.1950137Z * [new tag] viable/strict/1761931517 -> viable/strict/1761931517 2025-12-04T09:20:36.1951577Z * [new tag] viable/strict/1761933080 -> viable/strict/1761933080 2025-12-04T09:20:36.1953123Z * [new tag] viable/strict/1761935217 -> viable/strict/1761935217 2025-12-04T09:20:36.1954567Z * [new tag] viable/strict/1761938533 -> viable/strict/1761938533 2025-12-04T09:20:36.1956087Z * [new tag] viable/strict/1761940184 -> viable/strict/1761940184 2025-12-04T09:20:36.1957542Z * [new tag] viable/strict/1761942338 -> viable/strict/1761942338 2025-12-04T09:20:36.1959089Z * [new tag] viable/strict/1761946100 -> viable/strict/1761946100 2025-12-04T09:20:36.1960501Z * [new tag] viable/strict/1761947374 -> viable/strict/1761947374 2025-12-04T09:20:36.1961946Z * [new tag] viable/strict/1761950978 -> viable/strict/1761950978 2025-12-04T09:20:36.1963570Z * [new tag] viable/strict/1761957727 -> viable/strict/1761957727 2025-12-04T09:20:36.1964987Z * [new tag] viable/strict/1761959532 -> viable/strict/1761959532 2025-12-04T09:20:36.1966394Z * [new tag] viable/strict/1761965366 -> viable/strict/1761965366 2025-12-04T09:20:36.1967960Z * [new tag] viable/strict/1761968066 -> viable/strict/1761968066 2025-12-04T09:20:36.1969407Z * [new tag] viable/strict/1761969322 -> viable/strict/1761969322 2025-12-04T09:20:36.1970915Z * [new tag] viable/strict/1761974723 -> viable/strict/1761974723 2025-12-04T09:20:36.1972670Z * [new tag] viable/strict/1761981837 -> viable/strict/1761981837 2025-12-04T09:20:36.1974188Z * [new tag] viable/strict/1761985546 -> viable/strict/1761985546 2025-12-04T09:20:36.1975665Z * [new tag] viable/strict/1761987030 -> viable/strict/1761987030 2025-12-04T09:20:36.1977240Z * [new tag] viable/strict/1762003554 -> viable/strict/1762003554 2025-12-04T09:20:36.1978652Z * [new tag] viable/strict/1762021560 -> viable/strict/1762021560 2025-12-04T09:20:36.1980204Z * [new tag] viable/strict/1762032190 -> viable/strict/1762032190 2025-12-04T09:20:36.1981714Z * [new tag] viable/strict/1762040981 -> viable/strict/1762040981 2025-12-04T09:20:36.1983187Z * [new tag] viable/strict/1762048525 -> viable/strict/1762048525 2025-12-04T09:20:36.1984701Z * [new tag] viable/strict/1762104223 -> viable/strict/1762104223 2025-12-04T09:20:36.1986161Z * [new tag] viable/strict/1762105778 -> viable/strict/1762105778 2025-12-04T09:20:36.1987615Z * [new tag] viable/strict/1762115109 -> viable/strict/1762115109 2025-12-04T09:20:36.1989075Z * [new tag] viable/strict/1762125840 -> viable/strict/1762125840 2025-12-04T09:20:36.1990413Z * [new tag] viable/strict/1762127377 -> viable/strict/1762127377 2025-12-04T09:20:36.1992188Z * [new tag] viable/strict/1762134925 -> viable/strict/1762134925 2025-12-04T09:20:36.1993578Z * [new tag] viable/strict/1762138338 -> viable/strict/1762138338 2025-12-04T09:20:36.1995083Z * [new tag] viable/strict/1762148993 -> viable/strict/1762148993 2025-12-04T09:20:36.1996529Z * [new tag] viable/strict/1762152871 -> viable/strict/1762152871 2025-12-04T09:20:36.1997997Z * [new tag] viable/strict/1762156183 -> viable/strict/1762156183 2025-12-04T09:20:36.1999642Z * [new tag] viable/strict/1762163457 -> viable/strict/1762163457 2025-12-04T09:20:36.2001160Z * [new tag] viable/strict/1762165569 -> viable/strict/1762165569 2025-12-04T09:20:36.2002733Z * [new tag] viable/strict/1762169035 -> viable/strict/1762169035 2025-12-04T09:20:36.2004252Z * [new tag] viable/strict/1762174936 -> viable/strict/1762174936 2025-12-04T09:20:36.2005704Z * [new tag] viable/strict/1762194412 -> viable/strict/1762194412 2025-12-04T09:20:36.2007219Z * [new tag] viable/strict/1762195876 -> viable/strict/1762195876 2025-12-04T09:20:36.2008675Z * [new tag] viable/strict/1762197788 -> viable/strict/1762197788 2025-12-04T09:20:36.2010174Z * [new tag] viable/strict/1762199389 -> viable/strict/1762199389 2025-12-04T09:20:36.2011837Z * [new tag] viable/strict/1762206585 -> viable/strict/1762206585 2025-12-04T09:20:36.2013309Z * [new tag] viable/strict/1762210184 -> viable/strict/1762210184 2025-12-04T09:20:36.2014635Z * [new tag] viable/strict/1762218736 -> viable/strict/1762218736 2025-12-04T09:20:36.2016123Z * [new tag] viable/strict/1762224529 -> viable/strict/1762224529 2025-12-04T09:20:36.2017861Z * [new tag] viable/strict/1762227253 -> viable/strict/1762227253 2025-12-04T09:20:36.2019166Z * [new tag] viable/strict/1762228515 -> viable/strict/1762228515 2025-12-04T09:20:36.2020764Z * [new tag] viable/strict/1762230349 -> viable/strict/1762230349 2025-12-04T09:20:36.2022199Z * [new tag] viable/strict/1762231859 -> viable/strict/1762231859 2025-12-04T09:20:36.2023886Z * [new tag] viable/strict/1762233925 -> viable/strict/1762233925 2025-12-04T09:20:36.2025507Z * [new tag] viable/strict/1762237630 -> viable/strict/1762237630 2025-12-04T09:20:36.2026871Z * [new tag] viable/strict/1762253522 -> viable/strict/1762253522 2025-12-04T09:20:36.2028450Z * [new tag] viable/strict/1762278588 -> viable/strict/1762278588 2025-12-04T09:20:36.2029881Z * [new tag] viable/strict/1762284203 -> viable/strict/1762284203 2025-12-04T09:20:36.2031406Z * [new tag] viable/strict/1762289446 -> viable/strict/1762289446 2025-12-04T09:20:36.2032890Z * [new tag] viable/strict/1762291515 -> viable/strict/1762291515 2025-12-04T09:20:36.2034394Z * [new tag] viable/strict/1762295100 -> viable/strict/1762295100 2025-12-04T09:20:36.2035747Z * [new tag] viable/strict/1762296590 -> viable/strict/1762296590 2025-12-04T09:20:36.2037488Z * [new tag] viable/strict/1762300179 -> viable/strict/1762300179 2025-12-04T09:20:36.2038879Z * [new tag] viable/strict/1762303207 -> viable/strict/1762303207 2025-12-04T09:20:36.2040352Z * [new tag] viable/strict/1762386584 -> viable/strict/1762386584 2025-12-04T09:20:36.2041892Z * [new tag] viable/strict/1762391537 -> viable/strict/1762391537 2025-12-04T09:20:36.2043097Z * [new tag] viable/strict/1762394119 -> viable/strict/1762394119 2025-12-04T09:20:36.2044843Z * [new tag] viable/strict/1762397437 -> viable/strict/1762397437 2025-12-04T09:20:36.2046360Z * [new tag] viable/strict/1762400256 -> viable/strict/1762400256 2025-12-04T09:20:36.2047868Z * [new tag] viable/strict/1762401469 -> viable/strict/1762401469 2025-12-04T09:20:36.2049394Z * [new tag] viable/strict/1762408195 -> viable/strict/1762408195 2025-12-04T09:20:36.2050890Z * [new tag] viable/strict/1762410411 -> viable/strict/1762410411 2025-12-04T09:20:36.2052474Z * [new tag] viable/strict/1762417613 -> viable/strict/1762417613 2025-12-04T09:20:36.2054073Z * [new tag] viable/strict/1762419198 -> viable/strict/1762419198 2025-12-04T09:20:36.2055565Z * [new tag] viable/strict/1762422656 -> viable/strict/1762422656 2025-12-04T09:20:36.2057373Z * [new tag] viable/strict/1762424746 -> viable/strict/1762424746 2025-12-04T09:20:36.2058902Z * [new tag] viable/strict/1762446386 -> viable/strict/1762446386 2025-12-04T09:20:36.2060536Z * [new tag] viable/strict/1762449912 -> viable/strict/1762449912 2025-12-04T09:20:36.2062005Z * [new tag] viable/strict/1762457031 -> viable/strict/1762457031 2025-12-04T09:20:36.2063430Z * [new tag] viable/strict/1762462441 -> viable/strict/1762462441 2025-12-04T09:20:36.2064888Z * [new tag] viable/strict/1762467909 -> viable/strict/1762467909 2025-12-04T09:20:36.2066452Z * [new tag] viable/strict/1762471493 -> viable/strict/1762471493 2025-12-04T09:20:36.2067888Z * [new tag] viable/strict/1762475990 -> viable/strict/1762475990 2025-12-04T09:20:36.2069524Z * [new tag] viable/strict/1762477933 -> viable/strict/1762477933 2025-12-04T09:20:36.2071106Z * [new tag] viable/strict/1762491053 -> viable/strict/1762491053 2025-12-04T09:20:36.2072960Z * [new tag] viable/strict/1762493118 -> viable/strict/1762493118 2025-12-04T09:20:36.2074254Z * [new tag] viable/strict/1762498442 -> viable/strict/1762498442 2025-12-04T09:20:36.2075757Z * [new tag] viable/strict/1762501778 -> viable/strict/1762501778 2025-12-04T09:20:36.2077254Z * [new tag] viable/strict/1762504001 -> viable/strict/1762504001 2025-12-04T09:20:36.2078853Z * [new tag] viable/strict/1762505583 -> viable/strict/1762505583 2025-12-04T09:20:36.2080375Z * [new tag] viable/strict/1762507523 -> viable/strict/1762507523 2025-12-04T09:20:36.2081891Z * [new tag] viable/strict/1762511140 -> viable/strict/1762511140 2025-12-04T09:20:36.2083451Z * [new tag] viable/strict/1762512632 -> viable/strict/1762512632 2025-12-04T09:20:36.2085069Z * [new tag] viable/strict/1762520467 -> viable/strict/1762520467 2025-12-04T09:20:36.2086515Z * [new tag] viable/strict/1762522016 -> viable/strict/1762522016 2025-12-04T09:20:36.2087939Z * [new tag] viable/strict/1762530591 -> viable/strict/1762530591 2025-12-04T09:20:36.2089450Z * [new tag] viable/strict/1762543405 -> viable/strict/1762543405 2025-12-04T09:20:36.2090814Z * [new tag] viable/strict/1762544998 -> viable/strict/1762544998 2025-12-04T09:20:36.2092216Z * [new tag] viable/strict/1762552182 -> viable/strict/1762552182 2025-12-04T09:20:36.2093787Z * [new tag] viable/strict/1762554297 -> viable/strict/1762554297 2025-12-04T09:20:36.2095133Z * [new tag] viable/strict/1762559381 -> viable/strict/1762559381 2025-12-04T09:20:36.2096618Z * [new tag] viable/strict/1762562222 -> viable/strict/1762562222 2025-12-04T09:20:36.2098109Z * [new tag] viable/strict/1762564319 -> viable/strict/1762564319 2025-12-04T09:20:36.2099548Z * [new tag] viable/strict/1762566904 -> viable/strict/1762566904 2025-12-04T09:20:36.2101076Z * [new tag] viable/strict/1762569781 -> viable/strict/1762569781 2025-12-04T09:20:36.2102467Z * [new tag] viable/strict/1762575940 -> viable/strict/1762575940 2025-12-04T09:20:36.2104056Z * [new tag] viable/strict/1762580974 -> viable/strict/1762580974 2025-12-04T09:20:36.2105622Z * [new tag] viable/strict/1762583185 -> viable/strict/1762583185 2025-12-04T09:20:36.2107110Z * [new tag] viable/strict/1762586647 -> viable/strict/1762586647 2025-12-04T09:20:36.2108569Z * [new tag] viable/strict/1762588183 -> viable/strict/1762588183 2025-12-04T09:20:36.2110077Z * [new tag] viable/strict/1762593886 -> viable/strict/1762593886 2025-12-04T09:20:36.2111550Z * [new tag] viable/strict/1762650743 -> viable/strict/1762650743 2025-12-04T09:20:36.2113290Z * [new tag] viable/strict/1762653328 -> viable/strict/1762653328 2025-12-04T09:20:36.2114756Z * [new tag] viable/strict/1762659342 -> viable/strict/1762659342 2025-12-04T09:20:36.2116216Z * [new tag] viable/strict/1762662360 -> viable/strict/1762662360 2025-12-04T09:20:36.2117689Z * [new tag] viable/strict/1762667377 -> viable/strict/1762667377 2025-12-04T09:20:36.2119174Z * [new tag] viable/strict/1762671090 -> viable/strict/1762671090 2025-12-04T09:20:36.2120654Z * [new tag] viable/strict/1762680284 -> viable/strict/1762680284 2025-12-04T09:20:36.2122172Z * [new tag] viable/strict/1762683900 -> viable/strict/1762683900 2025-12-04T09:20:36.2123626Z * [new tag] viable/strict/1762705541 -> viable/strict/1762705541 2025-12-04T09:20:36.2125120Z * [new tag] viable/strict/1762709004 -> viable/strict/1762709004 2025-12-04T09:20:36.2126734Z * [new tag] viable/strict/1762746004 -> viable/strict/1762746004 2025-12-04T09:20:36.2128217Z * [new tag] viable/strict/1762748799 -> viable/strict/1762748799 2025-12-04T09:20:36.2129674Z * [new tag] viable/strict/1762759504 -> viable/strict/1762759504 2025-12-04T09:20:36.2131276Z * [new tag] viable/strict/1762760973 -> viable/strict/1762760973 2025-12-04T09:20:36.2132875Z * [new tag] viable/strict/1762775374 -> viable/strict/1762775374 2025-12-04T09:20:36.2134714Z * [new tag] viable/strict/1762777661 -> viable/strict/1762777661 2025-12-04T09:20:36.2136177Z * [new tag] viable/strict/1762779774 -> viable/strict/1762779774 2025-12-04T09:20:36.2137891Z * [new tag] viable/strict/1762781259 -> viable/strict/1762781259 2025-12-04T09:20:36.2139423Z * [new tag] viable/strict/1762793628 -> viable/strict/1762793628 2025-12-04T09:20:36.2141001Z * [new tag] viable/strict/1762800711 -> viable/strict/1762800711 2025-12-04T09:20:36.2142422Z * [new tag] viable/strict/1762809894 -> viable/strict/1762809894 2025-12-04T09:20:36.2143928Z * [new tag] viable/strict/1762811384 -> viable/strict/1762811384 2025-12-04T09:20:36.2145439Z * [new tag] viable/strict/1762813841 -> viable/strict/1762813841 2025-12-04T09:20:36.2146888Z * [new tag] viable/strict/1762815047 -> viable/strict/1762815047 2025-12-04T09:20:36.2148536Z * [new tag] viable/strict/1762817094 -> viable/strict/1762817094 2025-12-04T09:20:36.2150041Z * [new tag] viable/strict/1762818582 -> viable/strict/1762818582 2025-12-04T09:20:36.2151507Z * [new tag] viable/strict/1762821623 -> viable/strict/1762821623 2025-12-04T09:20:36.2153009Z * [new tag] viable/strict/1762823531 -> viable/strict/1762823531 2025-12-04T09:20:36.2154597Z * [new tag] viable/strict/1762849583 -> viable/strict/1762849583 2025-12-04T09:20:36.2156183Z * [new tag] viable/strict/1762851200 -> viable/strict/1762851200 2025-12-04T09:20:36.2157546Z * [new tag] viable/strict/1762854603 -> viable/strict/1762854603 2025-12-04T09:20:36.2159054Z * [new tag] viable/strict/1762858276 -> viable/strict/1762858276 2025-12-04T09:20:36.2160571Z * [new tag] viable/strict/1762860891 -> viable/strict/1762860891 2025-12-04T09:20:36.2162621Z * [new tag] viable/strict/1762866174 -> viable/strict/1762866174 2025-12-04T09:20:36.2164045Z * [new tag] viable/strict/1762867653 -> viable/strict/1762867653 2025-12-04T09:20:36.2165526Z * [new tag] viable/strict/1762872669 -> viable/strict/1762872669 2025-12-04T09:20:36.2166877Z * [new tag] viable/strict/1762878380 -> viable/strict/1762878380 2025-12-04T09:20:36.2168441Z * [new tag] viable/strict/1762889003 -> viable/strict/1762889003 2025-12-04T09:20:36.2169881Z * [new tag] viable/strict/1762890589 -> viable/strict/1762890589 2025-12-04T09:20:36.2171331Z * [new tag] viable/strict/1762892743 -> viable/strict/1762892743 2025-12-04T09:20:36.2173026Z * [new tag] viable/strict/1762894271 -> viable/strict/1762894271 2025-12-04T09:20:36.2174413Z * [new tag] viable/strict/1762896287 -> viable/strict/1762896287 2025-12-04T09:20:36.2175845Z * [new tag] viable/strict/1762915871 -> viable/strict/1762915871 2025-12-04T09:20:36.2177429Z * [new tag] viable/strict/1762918569 -> viable/strict/1762918569 2025-12-04T09:20:36.2178765Z * [new tag] viable/strict/1762919776 -> viable/strict/1762919776 2025-12-04T09:20:36.2180434Z * [new tag] viable/strict/1762923072 -> viable/strict/1762923072 2025-12-04T09:20:36.2182031Z * [new tag] viable/strict/1762928826 -> viable/strict/1762928826 2025-12-04T09:20:36.2183517Z * [new tag] viable/strict/1762930451 -> viable/strict/1762930451 2025-12-04T09:20:36.2185060Z * [new tag] viable/strict/1762933780 -> viable/strict/1762933780 2025-12-04T09:20:36.2186543Z * [new tag] viable/strict/1762937638 -> viable/strict/1762937638 2025-12-04T09:20:36.2188160Z * [new tag] viable/strict/1762939545 -> viable/strict/1762939545 2025-12-04T09:20:36.2189719Z * [new tag] viable/strict/1762962692 -> viable/strict/1762962692 2025-12-04T09:20:36.2191286Z * [new tag] viable/strict/1762979143 -> viable/strict/1762979143 2025-12-04T09:20:36.2192714Z * [new tag] viable/strict/1762984188 -> viable/strict/1762984188 2025-12-04T09:20:36.2194055Z * [new tag] viable/strict/1762986306 -> viable/strict/1762986306 2025-12-04T09:20:36.2195568Z * [new tag] viable/strict/1762989903 -> viable/strict/1762989903 2025-12-04T09:20:36.2197059Z * [new tag] viable/strict/1762991377 -> viable/strict/1762991377 2025-12-04T09:20:36.2198587Z * [new tag] viable/strict/1762998921 -> viable/strict/1762998921 2025-12-04T09:20:36.2200125Z * [new tag] viable/strict/1763002287 -> viable/strict/1763002287 2025-12-04T09:20:36.2201743Z * [new tag] viable/strict/1763016840 -> viable/strict/1763016840 2025-12-04T09:20:36.2203127Z * [new tag] viable/strict/1763020180 -> viable/strict/1763020180 2025-12-04T09:20:36.2204705Z * [new tag] viable/strict/1763027421 -> viable/strict/1763027421 2025-12-04T09:20:36.2206178Z * [new tag] viable/strict/1763031120 -> viable/strict/1763031120 2025-12-04T09:20:36.2207677Z * [new tag] viable/strict/1763036861 -> viable/strict/1763036861 2025-12-04T09:20:36.2209148Z * [new tag] viable/strict/1763038993 -> viable/strict/1763038993 2025-12-04T09:20:36.2210795Z * [new tag] viable/strict/1763054703 -> viable/strict/1763054703 2025-12-04T09:20:36.2212170Z * [new tag] viable/strict/1763067061 -> viable/strict/1763067061 2025-12-04T09:20:36.2213676Z * [new tag] viable/strict/1763070847 -> viable/strict/1763070847 2025-12-04T09:20:36.2215176Z * [new tag] viable/strict/1763072706 -> viable/strict/1763072706 2025-12-04T09:20:36.2216750Z * [new tag] viable/strict/1763076302 -> viable/strict/1763076302 2025-12-04T09:20:36.2218294Z * [new tag] viable/strict/1763080816 -> viable/strict/1763080816 2025-12-04T09:20:36.2219814Z * [new tag] viable/strict/1763082732 -> viable/strict/1763082732 2025-12-04T09:20:36.2221356Z * [new tag] viable/strict/1763085329 -> viable/strict/1763085329 2025-12-04T09:20:36.2222932Z * [new tag] viable/strict/1763088623 -> viable/strict/1763088623 2025-12-04T09:20:36.2224484Z * [new tag] viable/strict/1763091402 -> viable/strict/1763091402 2025-12-04T09:20:36.2225958Z * [new tag] viable/strict/1763092602 -> viable/strict/1763092602 2025-12-04T09:20:36.2227586Z * [new tag] viable/strict/1763094355 -> viable/strict/1763094355 2025-12-04T09:20:36.2229029Z * [new tag] viable/strict/1763099390 -> viable/strict/1763099390 2025-12-04T09:20:36.2230488Z * [new tag] viable/strict/1763101608 -> viable/strict/1763101608 2025-12-04T09:20:36.2232377Z * [new tag] viable/strict/1763105102 -> viable/strict/1763105102 2025-12-04T09:20:36.2233916Z * [new tag] viable/strict/1763112347 -> viable/strict/1763112347 2025-12-04T09:20:36.2235405Z * [new tag] viable/strict/1763119471 -> viable/strict/1763119471 2025-12-04T09:20:36.2236882Z * [new tag] viable/strict/1763126835 -> viable/strict/1763126835 2025-12-04T09:20:36.2238113Z * [new tag] viable/strict/1763149779 -> viable/strict/1763149779 2025-12-04T09:20:36.2239629Z * [new tag] viable/strict/1763164178 -> viable/strict/1763164178 2025-12-04T09:20:36.2241054Z * [new tag] viable/strict/1763167104 -> viable/strict/1763167104 2025-12-04T09:20:36.2242638Z * [new tag] viable/strict/1763169132 -> viable/strict/1763169132 2025-12-04T09:20:36.2244110Z * [new tag] viable/strict/1763171708 -> viable/strict/1763171708 2025-12-04T09:20:36.2245669Z * [new tag] viable/strict/1763174759 -> viable/strict/1763174759 2025-12-04T09:20:36.2247153Z * [new tag] viable/strict/1763180744 -> viable/strict/1763180744 2025-12-04T09:20:36.2248600Z * [new tag] viable/strict/1763182227 -> viable/strict/1763182227 2025-12-04T09:20:36.2250171Z * [new tag] viable/strict/1763184309 -> viable/strict/1763184309 2025-12-04T09:20:36.2252184Z * [new tag] viable/strict/1763187991 -> viable/strict/1763187991 2025-12-04T09:20:36.2253593Z * [new tag] viable/strict/1763191445 -> viable/strict/1763191445 2025-12-04T09:20:36.2255145Z * [new tag] viable/strict/1763195152 -> viable/strict/1763195152 2025-12-04T09:20:36.2256511Z * [new tag] viable/strict/1763205769 -> viable/strict/1763205769 2025-12-04T09:20:36.2258047Z * [new tag] viable/strict/1763246990 -> viable/strict/1763246990 2025-12-04T09:20:36.2259664Z * [new tag] viable/strict/1763261578 -> viable/strict/1763261578 2025-12-04T09:20:36.2261097Z * [new tag] viable/strict/1763286573 -> viable/strict/1763286573 2025-12-04T09:20:36.2262507Z * [new tag] viable/strict/1763292167 -> viable/strict/1763292167 2025-12-04T09:20:36.2264068Z * [new tag] viable/strict/1763333386 -> viable/strict/1763333386 2025-12-04T09:20:36.2265466Z * [new tag] viable/strict/1763340082 -> viable/strict/1763340082 2025-12-04T09:20:36.2267568Z * [new tag] viable/strict/1763364324 -> viable/strict/1763364324 2025-12-04T09:20:36.2269056Z * [new tag] viable/strict/1763371569 -> viable/strict/1763371569 2025-12-04T09:20:36.2270577Z * [new tag] viable/strict/1763373067 -> viable/strict/1763373067 2025-12-04T09:20:36.2273817Z * [new tag] viable/strict/1763375157 -> viable/strict/1763375157 2025-12-04T09:20:36.2275334Z * [new tag] viable/strict/1763382462 -> viable/strict/1763382462 2025-12-04T09:20:36.2276932Z * [new tag] viable/strict/1763394661 -> viable/strict/1763394661 2025-12-04T09:20:36.2278535Z * [new tag] viable/strict/1763396797 -> viable/strict/1763396797 2025-12-04T09:20:36.2280065Z * [new tag] viable/strict/1763398542 -> viable/strict/1763398542 2025-12-04T09:20:36.2281554Z * [new tag] viable/strict/1763401807 -> viable/strict/1763401807 2025-12-04T09:20:36.2282950Z * [new tag] viable/strict/1763414698 -> viable/strict/1763414698 2025-12-04T09:20:36.2284488Z * [new tag] viable/strict/1763419807 -> viable/strict/1763419807 2025-12-04T09:20:36.2286007Z * [new tag] viable/strict/1763426369 -> viable/strict/1763426369 2025-12-04T09:20:36.2287518Z * [new tag] viable/strict/1763428331 -> viable/strict/1763428331 2025-12-04T09:20:36.2289045Z * [new tag] viable/strict/1763430922 -> viable/strict/1763430922 2025-12-04T09:20:36.2290389Z * [new tag] viable/strict/1763434184 -> viable/strict/1763434184 2025-12-04T09:20:36.2291877Z * [new tag] viable/strict/1763439973 -> viable/strict/1763439973 2025-12-04T09:20:36.2293526Z * [new tag] viable/strict/1763444995 -> viable/strict/1763444995 2025-12-04T09:20:36.2294982Z * [new tag] viable/strict/1763447206 -> viable/strict/1763447206 2025-12-04T09:20:36.2296458Z * [new tag] viable/strict/1763448826 -> viable/strict/1763448826 2025-12-04T09:20:36.2297932Z * [new tag] viable/strict/1763450717 -> viable/strict/1763450717 2025-12-04T09:20:36.2299605Z * [new tag] viable/strict/1763452183 -> viable/strict/1763452183 2025-12-04T09:20:36.2301273Z * [new tag] viable/strict/1763457945 -> viable/strict/1763457945 2025-12-04T09:20:36.2302862Z * [new tag] viable/strict/1763459439 -> viable/strict/1763459439 2025-12-04T09:20:36.2304224Z * [new tag] viable/strict/1763461556 -> viable/strict/1763461556 2025-12-04T09:20:36.2305736Z * [new tag] viable/strict/1763463103 -> viable/strict/1763463103 2025-12-04T09:20:36.2307256Z * [new tag] viable/strict/1763465100 -> viable/strict/1763465100 2025-12-04T09:20:36.2308623Z * [new tag] viable/strict/1763468866 -> viable/strict/1763468866 2025-12-04T09:20:36.2310105Z * [new tag] viable/strict/1763493823 -> viable/strict/1763493823 2025-12-04T09:20:36.2311459Z * [new tag] viable/strict/1763496249 -> viable/strict/1763496249 2025-12-04T09:20:36.2312881Z * [new tag] viable/strict/1763502620 -> viable/strict/1763502620 2025-12-04T09:20:36.2314453Z * [new tag] viable/strict/1763504715 -> viable/strict/1763504715 2025-12-04T09:20:36.2315943Z * [new tag] viable/strict/1763506208 -> viable/strict/1763506208 2025-12-04T09:20:36.2317439Z * [new tag] viable/strict/1763520590 -> viable/strict/1763520590 2025-12-04T09:20:36.2319022Z * [new tag] viable/strict/1763523357 -> viable/strict/1763523357 2025-12-04T09:20:36.2320621Z * [new tag] viable/strict/1763529922 -> viable/strict/1763529922 2025-12-04T09:20:36.2322173Z * [new tag] viable/strict/1763531408 -> viable/strict/1763531408 2025-12-04T09:20:36.2323722Z * [new tag] viable/strict/1763533622 -> viable/strict/1763533622 2025-12-04T09:20:36.2325228Z * [new tag] viable/strict/1763538576 -> viable/strict/1763538576 2025-12-04T09:20:36.2326766Z * [new tag] viable/strict/1763545823 -> viable/strict/1763545823 2025-12-04T09:20:36.2328104Z * [new tag] viable/strict/1763547951 -> viable/strict/1763547951 2025-12-04T09:20:36.2329644Z * [new tag] viable/strict/1763551477 -> viable/strict/1763551477 2025-12-04T09:20:36.2331634Z * [new tag] viable/strict/1763552982 -> viable/strict/1763552982 2025-12-04T09:20:36.2333045Z * [new tag] viable/strict/1763594698 -> viable/strict/1763594698 2025-12-04T09:20:36.2334539Z * [new tag] viable/strict/1763596178 -> viable/strict/1763596178 2025-12-04T09:20:36.2336067Z * [new tag] viable/strict/1763599155 -> viable/strict/1763599155 2025-12-04T09:20:36.2337593Z * [new tag] viable/strict/1763603717 -> viable/strict/1763603717 2025-12-04T09:20:36.2339071Z * [new tag] viable/strict/1763606923 -> viable/strict/1763606923 2025-12-04T09:20:36.2340679Z * [new tag] viable/strict/1763609715 -> viable/strict/1763609715 2025-12-04T09:20:36.2342231Z * [new tag] viable/strict/1763612757 -> viable/strict/1763612757 2025-12-04T09:20:36.2343798Z * [new tag] viable/strict/1763616325 -> viable/strict/1763616325 2025-12-04T09:20:36.2345255Z * [new tag] viable/strict/1763623509 -> viable/strict/1763623509 2025-12-04T09:20:36.2346901Z * [new tag] viable/strict/1763624984 -> viable/strict/1763624984 2025-12-04T09:20:36.2348472Z * [new tag] viable/strict/1763628796 -> viable/strict/1763628796 2025-12-04T09:20:36.2349965Z * [new tag] viable/strict/1763634343 -> viable/strict/1763634343 2025-12-04T09:20:36.2351391Z * [new tag] viable/strict/1763635867 -> viable/strict/1763635867 2025-12-04T09:20:36.2352921Z * [new tag] viable/strict/1763639382 -> viable/strict/1763639382 2025-12-04T09:20:36.2354411Z * [new tag] viable/strict/1763646626 -> viable/strict/1763646626 2025-12-04T09:20:36.2356039Z * [new tag] viable/strict/1763655997 -> viable/strict/1763655997 2025-12-04T09:20:36.2357548Z * [new tag] viable/strict/1763659444 -> viable/strict/1763659444 2025-12-04T09:20:36.2359034Z * [new tag] viable/strict/1763660992 -> viable/strict/1763660992 2025-12-04T09:20:36.2360496Z * [new tag] viable/strict/1763663201 -> viable/strict/1763663201 2025-12-04T09:20:36.2362027Z * [new tag] viable/strict/1763670362 -> viable/strict/1763670362 2025-12-04T09:20:36.2363368Z * [new tag] viable/strict/1763675378 -> viable/strict/1763675378 2025-12-04T09:20:36.2364839Z * [new tag] viable/strict/1763693343 -> viable/strict/1763693343 2025-12-04T09:20:36.2366302Z * [new tag] viable/strict/1763696088 -> viable/strict/1763696088 2025-12-04T09:20:36.2367945Z * [new tag] viable/strict/1763697343 -> viable/strict/1763697343 2025-12-04T09:20:36.2369452Z * [new tag] viable/strict/1763699165 -> viable/strict/1763699165 2025-12-04T09:20:36.2371091Z * [new tag] viable/strict/1763700660 -> viable/strict/1763700660 2025-12-04T09:20:36.2372672Z * [new tag] viable/strict/1763704209 -> viable/strict/1763704209 2025-12-04T09:20:36.2374234Z * [new tag] viable/strict/1763706411 -> viable/strict/1763706411 2025-12-04T09:20:36.2375722Z * [new tag] viable/strict/1763708082 -> viable/strict/1763708082 2025-12-04T09:20:36.2377114Z * [new tag] viable/strict/1763711381 -> viable/strict/1763711381 2025-12-04T09:20:36.2378587Z * [new tag] viable/strict/1763713593 -> viable/strict/1763713593 2025-12-04T09:20:36.2380187Z * [new tag] viable/strict/1763715201 -> viable/strict/1763715201 2025-12-04T09:20:36.2381634Z * [new tag] viable/strict/1763733017 -> viable/strict/1763733017 2025-12-04T09:20:36.2383115Z * [new tag] viable/strict/1763735108 -> viable/strict/1763735108 2025-12-04T09:20:36.2384622Z * [new tag] viable/strict/1763749579 -> viable/strict/1763749579 2025-12-04T09:20:36.2386133Z * [new tag] viable/strict/1763751113 -> viable/strict/1763751113 2025-12-04T09:20:36.2387599Z * [new tag] viable/strict/1763753035 -> viable/strict/1763753035 2025-12-04T09:20:36.2389212Z * [new tag] viable/strict/1763754578 -> viable/strict/1763754578 2025-12-04T09:20:36.2390771Z * [new tag] viable/strict/1763756748 -> viable/strict/1763756748 2025-12-04T09:20:36.2392255Z * [new tag] viable/strict/1763758205 -> viable/strict/1763758205 2025-12-04T09:20:36.2393643Z * [new tag] viable/strict/1763764050 -> viable/strict/1763764050 2025-12-04T09:20:36.2395103Z * [new tag] viable/strict/1763771887 -> viable/strict/1763771887 2025-12-04T09:20:36.2396748Z * [new tag] viable/strict/1763773920 -> viable/strict/1763773920 2025-12-04T09:20:36.2398293Z * [new tag] viable/strict/1763776501 -> viable/strict/1763776501 2025-12-04T09:20:36.2399750Z * [new tag] viable/strict/1763779437 -> viable/strict/1763779437 2025-12-04T09:20:36.2401366Z * [new tag] viable/strict/1763781038 -> viable/strict/1763781038 2025-12-04T09:20:36.2402809Z * [new tag] viable/strict/1763782245 -> viable/strict/1763782245 2025-12-04T09:20:36.2404210Z * [new tag] viable/strict/1763785568 -> viable/strict/1763785568 2025-12-04T09:20:36.2405712Z * [new tag] viable/strict/1763787006 -> viable/strict/1763787006 2025-12-04T09:20:36.2407205Z * [new tag] viable/strict/1763789103 -> viable/strict/1763789103 2025-12-04T09:20:36.2408731Z * [new tag] viable/strict/1763790578 -> viable/strict/1763790578 2025-12-04T09:20:36.2410379Z * [new tag] viable/strict/1763796275 -> viable/strict/1763796275 2025-12-04T09:20:36.2412034Z * [new tag] viable/strict/1763801465 -> viable/strict/1763801465 2025-12-04T09:20:36.2414000Z * [new tag] viable/strict/1763803522 -> viable/strict/1763803522 2025-12-04T09:20:36.2414974Z * [new tag] viable/strict/1763808581 -> viable/strict/1763808581 2025-12-04T09:20:36.2416344Z * [new tag] viable/strict/1763840977 -> viable/strict/1763840977 2025-12-04T09:20:36.2417963Z * [new tag] viable/strict/1763846659 -> viable/strict/1763846659 2025-12-04T09:20:36.2419413Z * [new tag] viable/strict/1763872065 -> viable/strict/1763872065 2025-12-04T09:20:36.2421193Z * [new tag] viable/strict/1763873648 -> viable/strict/1763873648 2025-12-04T09:20:36.2422421Z * [new tag] viable/strict/1763875506 -> viable/strict/1763875506 2025-12-04T09:20:36.2423906Z * [new tag] viable/strict/1763889904 -> viable/strict/1763889904 2025-12-04T09:20:36.2425417Z * [new tag] viable/strict/1763930999 -> viable/strict/1763930999 2025-12-04T09:20:36.2426932Z * [new tag] viable/strict/1763944964 -> viable/strict/1763944964 2025-12-04T09:20:36.2428688Z * [new tag] viable/strict/1763958474 -> viable/strict/1763958474 2025-12-04T09:20:36.2430151Z * [new tag] viable/strict/1763967263 -> viable/strict/1763967263 2025-12-04T09:20:36.2431789Z * [new tag] viable/strict/1763972803 -> viable/strict/1763972803 2025-12-04T09:20:36.2433233Z * [new tag] viable/strict/1763976376 -> viable/strict/1763976376 2025-12-04T09:20:36.2434659Z * [new tag] viable/strict/1763989404 -> viable/strict/1763989404 2025-12-04T09:20:36.2436097Z * [new tag] viable/strict/1763990887 -> viable/strict/1763990887 2025-12-04T09:20:36.2437563Z * [new tag] viable/strict/1764019919 -> viable/strict/1764019919 2025-12-04T09:20:36.2439096Z * [new tag] viable/strict/1764023134 -> viable/strict/1764023134 2025-12-04T09:20:36.2440420Z * [new tag] viable/strict/1764024593 -> viable/strict/1764024593 2025-12-04T09:20:36.2441917Z * [new tag] viable/strict/1764026706 -> viable/strict/1764026706 2025-12-04T09:20:36.2443555Z * [new tag] viable/strict/1764031139 -> viable/strict/1764031139 2025-12-04T09:20:36.2445142Z * [new tag] viable/strict/1764033131 -> viable/strict/1764033131 2025-12-04T09:20:36.2446554Z * [new tag] viable/strict/1764035725 -> viable/strict/1764035725 2025-12-04T09:20:36.2447866Z * [new tag] viable/strict/1764624265 -> viable/strict/1764624265 2025-12-04T09:20:36.2449286Z * [new tag] viable/strict/1764631514 -> viable/strict/1764631514 2025-12-04T09:20:36.2450575Z * [new tag] viable/strict/1764632987 -> viable/strict/1764632987 2025-12-04T09:20:36.2452158Z * [new tag] viable/strict/1764636063 -> viable/strict/1764636063 2025-12-04T09:20:36.2453205Z * [new tag] viable/strict/1764643975 -> viable/strict/1764643975 2025-12-04T09:20:36.2454879Z * [new tag] viable/strict/1764646859 -> viable/strict/1764646859 2025-12-04T09:20:36.2456505Z * [new tag] viable/strict/1764653120 -> viable/strict/1764653120 2025-12-04T09:20:36.2457419Z * [new tag] viable/strict/1764654632 -> viable/strict/1764654632 2025-12-04T09:20:36.2458899Z * [new tag] viable/strict/1764656821 -> viable/strict/1764656821 2025-12-04T09:20:36.2460368Z * [new tag] viable/strict/1764658557 -> viable/strict/1764658557 2025-12-04T09:20:36.2461725Z * [new tag] viable/strict/1764660333 -> viable/strict/1764660333 2025-12-04T09:20:36.2463017Z * [new tag] viable/strict/1764661812 -> viable/strict/1764661812 2025-12-04T09:20:36.2464202Z * [new tag] viable/strict/1764664023 -> viable/strict/1764664023 2025-12-04T09:20:36.2465700Z * [new tag] viable/strict/1764669150 -> viable/strict/1764669150 2025-12-04T09:20:36.2467038Z * [new tag] viable/strict/1764680709 -> viable/strict/1764680709 2025-12-04T09:20:36.2468371Z * [new tag] viable/strict/1764687619 -> viable/strict/1764687619 2025-12-04T09:20:36.2469698Z * [new tag] viable/strict/1764696355 -> viable/strict/1764696355 2025-12-04T09:20:36.2471155Z * [new tag] viable/strict/1764701767 -> viable/strict/1764701767 2025-12-04T09:20:36.2472667Z * [new tag] viable/strict/1764710768 -> viable/strict/1764710768 2025-12-04T09:20:36.2473976Z * [new tag] viable/strict/1764716202 -> viable/strict/1764716202 2025-12-04T09:20:36.2475316Z * [new tag] viable/strict/1764793566 -> viable/strict/1764793566 2025-12-04T09:20:36.2476642Z * [new tag] viable/strict/1764797093 -> viable/strict/1764797093 2025-12-04T09:20:36.2477970Z * [new tag] viable/strict/1764800729 -> viable/strict/1764800729 2025-12-04T09:20:36.2479412Z * [new tag] whc_flight_1 -> whc_flight_1 2025-12-04T09:20:36.2480859Z * [new tag] whc_flight_2 -> whc_flight_2 2025-12-04T09:20:36.2482474Z * [new tag] whc_flight_4 -> whc_flight_4 2025-12-04T09:20:36.3533557Z [command]/usr/bin/git rev-parse --verify --quiet ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32^{object} 2025-12-04T09:20:36.3559797Z ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:20:36.3564434Z ##[endgroup] 2025-12-04T09:20:36.3564854Z ##[group]Determining the checkout info 2025-12-04T09:20:36.3565660Z ##[endgroup] 2025-12-04T09:20:36.3569544Z [command]/usr/bin/git sparse-checkout disable 2025-12-04T09:20:36.3610276Z [command]/usr/bin/git config --local --unset-all extensions.worktreeConfig 2025-12-04T09:20:36.3640107Z ##[group]Checking out the ref 2025-12-04T09:20:36.3643623Z [command]/usr/bin/git checkout --progress --force ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:20:37.3953055Z Updating files: 73% (14795/20121) 2025-12-04T09:20:37.4199447Z Updating files: 74% (14890/20121) 2025-12-04T09:20:37.4669587Z Updating files: 75% (15091/20121) 2025-12-04T09:20:37.4829398Z Updating files: 76% (15292/20121) 2025-12-04T09:20:37.4984485Z Updating files: 77% (15494/20121) 2025-12-04T09:20:37.5198563Z Updating files: 78% (15695/20121) 2025-12-04T09:20:37.5464194Z Updating files: 79% (15896/20121) 2025-12-04T09:20:37.5777701Z Updating files: 80% (16097/20121) 2025-12-04T09:20:37.6062080Z Updating files: 81% (16299/20121) 2025-12-04T09:20:37.6289842Z Updating files: 82% (16500/20121) 2025-12-04T09:20:37.6463931Z Updating files: 83% (16701/20121) 2025-12-04T09:20:37.6624319Z Updating files: 84% (16902/20121) 2025-12-04T09:20:37.6803016Z Updating files: 85% (17103/20121) 2025-12-04T09:20:37.6978517Z Updating files: 86% (17305/20121) 2025-12-04T09:20:37.7142574Z Updating files: 87% (17506/20121) 2025-12-04T09:20:37.7281081Z Updating files: 88% (17707/20121) 2025-12-04T09:20:37.7439322Z Updating files: 89% (17908/20121) 2025-12-04T09:20:37.7627635Z Updating files: 90% (18109/20121) 2025-12-04T09:20:37.7765640Z Updating files: 91% (18311/20121) 2025-12-04T09:20:37.7939474Z Updating files: 92% (18512/20121) 2025-12-04T09:20:37.8138009Z Updating files: 93% (18713/20121) 2025-12-04T09:20:37.8351633Z Updating files: 94% (18914/20121) 2025-12-04T09:20:37.8543200Z Updating files: 95% (19115/20121) 2025-12-04T09:20:37.8721315Z Updating files: 96% (19317/20121) 2025-12-04T09:20:37.8902906Z Updating files: 97% (19518/20121) 2025-12-04T09:20:37.9186463Z Updating files: 98% (19719/20121) 2025-12-04T09:20:37.9377709Z Updating files: 99% (19920/20121) 2025-12-04T09:20:37.9377997Z Updating files: 100% (20121/20121) 2025-12-04T09:20:37.9378588Z Updating files: 100% (20121/20121), done. 2025-12-04T09:20:37.9612876Z Note: switching to 'ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32'. 2025-12-04T09:20:37.9613170Z 2025-12-04T09:20:37.9613370Z You are in 'detached HEAD' state. You can look around, make experimental 2025-12-04T09:20:37.9613880Z changes and commit them, and you can discard any commits you make in this 2025-12-04T09:20:37.9614355Z state without impacting any branches by switching back to a branch. 2025-12-04T09:20:37.9614641Z 2025-12-04T09:20:37.9614823Z If you want to create a new branch to retain commits you create, you may 2025-12-04T09:20:37.9615267Z do so (now or later) by using -c with the switch command. Example: 2025-12-04T09:20:37.9615524Z 2025-12-04T09:20:37.9615646Z git switch -c 2025-12-04T09:20:37.9615824Z 2025-12-04T09:20:37.9615918Z Or undo this operation with: 2025-12-04T09:20:37.9616080Z 2025-12-04T09:20:37.9616160Z git switch - 2025-12-04T09:20:37.9616282Z 2025-12-04T09:20:37.9616503Z Turn off this advice by setting config variable advice.detachedHead to false 2025-12-04T09:20:37.9616830Z 2025-12-04T09:20:37.9619409Z HEAD is now at ffd9b0fb435 Resolve collective autotuning test failure on arm (#168919) 2025-12-04T09:20:37.9742280Z ##[endgroup] 2025-12-04T09:20:37.9742685Z ##[group]Setting up auth for fetching submodules 2025-12-04T09:20:37.9748796Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-12-04T09:20:37.9807236Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2025-12-04T09:20:37.9840926Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2025-12-04T09:20:37.9873353Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2025-12-04T09:20:37.9899789Z ##[endgroup] 2025-12-04T09:20:37.9900455Z ##[group]Fetching submodules 2025-12-04T09:20:37.9904104Z [command]/usr/bin/git submodule sync --recursive 2025-12-04T09:20:38.0273621Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --recursive 2025-12-04T09:20:38.0626422Z Submodule 'android/libs/fbjni' (https://github.com/facebookincubator/fbjni.git) registered for path 'android/libs/fbjni' 2025-12-04T09:20:38.0629264Z Submodule 'third_party/NNPACK_deps/FP16' (https://github.com/Maratyszcza/FP16.git) registered for path 'third_party/FP16' 2025-12-04T09:20:38.0633004Z Submodule 'third_party/NNPACK_deps/FXdiv' (https://github.com/Maratyszcza/FXdiv.git) registered for path 'third_party/FXdiv' 2025-12-04T09:20:38.0636603Z Submodule 'third_party/NNPACK' (https://github.com/Maratyszcza/NNPACK.git) registered for path 'third_party/NNPACK' 2025-12-04T09:20:38.0640378Z Submodule 'third_party/NVTX' (https://github.com/NVIDIA/NVTX.git) registered for path 'third_party/NVTX' 2025-12-04T09:20:38.0644367Z Submodule 'third_party/VulkanMemoryAllocator' (https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator.git) registered for path 'third_party/VulkanMemoryAllocator' 2025-12-04T09:20:38.0648096Z Submodule 'third_party/XNNPACK' (https://github.com/google/XNNPACK.git) registered for path 'third_party/XNNPACK' 2025-12-04T09:20:38.0652100Z Submodule 'third_party/aiter' (https://github.com/ROCm/aiter.git) registered for path 'third_party/aiter' 2025-12-04T09:20:38.0656485Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/benchmark' 2025-12-04T09:20:38.0661111Z Submodule 'third_party/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/composable_kernel' 2025-12-04T09:20:38.0665053Z Submodule 'third_party/cpp-httplib' (https://github.com/yhirose/cpp-httplib.git) registered for path 'third_party/cpp-httplib' 2025-12-04T09:20:38.0669349Z Submodule 'third_party/cpuinfo' (https://github.com/pytorch/cpuinfo.git) registered for path 'third_party/cpuinfo' 2025-12-04T09:20:38.0674255Z Submodule 'third_party/cudnn_frontend' (https://github.com/NVIDIA/cudnn-frontend.git) registered for path 'third_party/cudnn_frontend' 2025-12-04T09:20:38.0678689Z Submodule 'third_party/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/cutlass' 2025-12-04T09:20:38.0683221Z Submodule 'third_party/fbgemm' (https://github.com/pytorch/fbgemm) registered for path 'third_party/fbgemm' 2025-12-04T09:20:38.0688527Z Submodule 'third_party/flash-attention' (https://github.com/Dao-AILab/flash-attention.git) registered for path 'third_party/flash-attention' 2025-12-04T09:20:38.0696113Z Submodule 'third_party/flatbuffers' (https://github.com/google/flatbuffers.git) registered for path 'third_party/flatbuffers' 2025-12-04T09:20:38.0701141Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/fmt' 2025-12-04T09:20:38.0705896Z Submodule 'third_party/gemmlowp/gemmlowp' (https://github.com/google/gemmlowp.git) registered for path 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:20:38.0710732Z Submodule 'third_party/gloo' (https://github.com/pytorch/gloo) registered for path 'third_party/gloo' 2025-12-04T09:20:38.0715864Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/googletest' 2025-12-04T09:20:38.0720946Z Submodule 'third_party/ideep' (https://github.com/intel/ideep) registered for path 'third_party/ideep' 2025-12-04T09:20:38.0726103Z Submodule 'third_party/ittapi' (https://github.com/intel/ittapi.git) registered for path 'third_party/ittapi' 2025-12-04T09:20:38.0731364Z Submodule 'third_party/kineto' (https://github.com/pytorch/kineto) registered for path 'third_party/kineto' 2025-12-04T09:20:38.0736811Z Submodule 'third_party/kleidiai' (https://github.com/ARM-software/kleidiai.git) registered for path 'third_party/kleidiai' 2025-12-04T09:20:38.0742266Z Submodule 'third_party/mimalloc' (https://github.com/microsoft/mimalloc.git) registered for path 'third_party/mimalloc' 2025-12-04T09:20:38.0747733Z Submodule 'third_party/nlohmann' (https://github.com/nlohmann/json.git) registered for path 'third_party/nlohmann' 2025-12-04T09:20:38.0753457Z Submodule 'third_party/onnx' (https://github.com/onnx/onnx.git) registered for path 'third_party/onnx' 2025-12-04T09:20:38.0760362Z Submodule 'third_party/opentelemetry-cpp' (https://github.com/open-telemetry/opentelemetry-cpp.git) registered for path 'third_party/opentelemetry-cpp' 2025-12-04T09:20:38.0765309Z Submodule 'third_party/pocketfft' (https://github.com/mreineck/pocketfft) registered for path 'third_party/pocketfft' 2025-12-04T09:20:38.0771567Z Submodule 'third_party/protobuf' (https://github.com/protocolbuffers/protobuf.git) registered for path 'third_party/protobuf' 2025-12-04T09:20:38.0777732Z Submodule 'third_party/NNPACK_deps/psimd' (https://github.com/Maratyszcza/psimd.git) registered for path 'third_party/psimd' 2025-12-04T09:20:38.0783999Z Submodule 'third_party/NNPACK_deps/pthreadpool' (https://github.com/Maratyszcza/pthreadpool.git) registered for path 'third_party/pthreadpool' 2025-12-04T09:20:38.0792731Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/pybind11' 2025-12-04T09:20:38.0798979Z Submodule 'third_party/python-peachpy' (https://github.com/malfet/PeachPy.git) registered for path 'third_party/python-peachpy' 2025-12-04T09:20:38.0805377Z Submodule 'third_party/sleef' (https://github.com/shibatch/sleef) registered for path 'third_party/sleef' 2025-12-04T09:20:38.0811940Z Submodule 'third_party/tensorpipe' (https://github.com/pytorch/tensorpipe.git) registered for path 'third_party/tensorpipe' 2025-12-04T09:20:38.0848689Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/android/libs/fbjni'... 2025-12-04T09:20:38.3251631Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FXdiv'... 2025-12-04T09:20:38.3252594Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FP16'... 2025-12-04T09:20:38.3253438Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NNPACK'... 2025-12-04T09:20:38.3292023Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fmt'... 2025-12-04T09:20:40.8387914Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/benchmark'... 2025-12-04T09:20:40.8389279Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NVTX'... 2025-12-04T09:20:40.8390444Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gloo'... 2025-12-04T09:20:40.8391668Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention'... 2025-12-04T09:20:40.8392736Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpuinfo'... 2025-12-04T09:20:40.8393755Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gemmlowp/gemmlowp'... 2025-12-04T09:20:40.8394391Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpp-httplib'... 2025-12-04T09:20:40.8395073Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep'... 2025-12-04T09:20:40.8395790Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ittapi'... 2025-12-04T09:20:40.8396541Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kleidiai'... 2025-12-04T09:20:40.8397318Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pocketfft'... 2025-12-04T09:20:40.8397847Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/googletest'... 2025-12-04T09:20:40.8398371Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cudnn_frontend'... 2025-12-04T09:20:40.8399174Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/psimd'... 2025-12-04T09:20:40.8478682Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flatbuffers'... 2025-12-04T09:20:41.1330181Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/mimalloc'... 2025-12-04T09:20:41.1331280Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/VulkanMemoryAllocator'... 2025-12-04T09:20:41.1332289Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pthreadpool'... 2025-12-04T09:20:41.2331385Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/sleef'... 2025-12-04T09:20:42.1054265Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/python-peachpy'... 2025-12-04T09:20:42.1055702Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe'... 2025-12-04T09:20:42.1056845Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto'... 2025-12-04T09:20:42.2055590Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx'... 2025-12-04T09:20:43.8823698Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pybind11'... 2025-12-04T09:20:43.8825068Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cutlass'... 2025-12-04T09:20:43.8826279Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm'... 2025-12-04T09:20:43.9824775Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/XNNPACK'... 2025-12-04T09:20:59.9610097Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/composable_kernel'... 2025-12-04T09:20:59.9610888Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/nlohmann'... 2025-12-04T09:20:59.9611588Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp'... 2025-12-04T09:20:59.9612270Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter'... 2025-12-04T09:20:59.9612885Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf'... 2025-12-04T09:20:59.9811020Z Submodule path 'android/libs/fbjni': checked out '7e1e1fe3858c63c251c637ae41a20de425dde96f' 2025-12-04T09:20:59.9972750Z Submodule path 'third_party/FP16': checked out '4dfe081cf6bcd15db339cf2680b9281b8451eeb3' 2025-12-04T09:21:00.0103630Z Submodule path 'third_party/FXdiv': checked out 'b408327ac2a15ec3e43352421954f5b1967701d1' 2025-12-04T09:21:00.0420049Z Submodule path 'third_party/NNPACK': checked out 'c07e3a0400713d546e0dea2d5466dd22ea389c73' 2025-12-04T09:21:00.1373882Z Submodule path 'third_party/NVTX': checked out '3ebbc93ded7285963bff932c678fa367eb393ba6' 2025-12-04T09:21:00.1918777Z Submodule path 'third_party/VulkanMemoryAllocator': checked out '1d8f600fd424278486eade7ed3e877c99f0846b1' 2025-12-04T09:21:01.1199469Z Submodule path 'third_party/XNNPACK': checked out '51a0103656eff6fc9bfd39a4597923c4b542c883' 2025-12-04T09:21:01.3180127Z Submodule path 'third_party/aiter': checked out '01aae101b9e5e94d6c16a9514c9fb8df99c93150' 2025-12-04T09:21:01.3204966Z Submodule '3rdparty/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:21:01.3240846Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter/3rdparty/composable_kernel'... 2025-12-04T09:21:05.6182624Z Submodule path 'third_party/aiter/3rdparty/composable_kernel': checked out 'cffe8fa2a442ac8e80dd236a1a5d24fe3d7e0cbf' 2025-12-04T09:21:05.6485228Z Submodule path 'third_party/benchmark': checked out '299e5928955cc62af9968370293b916f5130916f' 2025-12-04T09:21:06.0817587Z Submodule path 'third_party/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-12-04T09:21:06.1354010Z Submodule path 'third_party/cpp-httplib': checked out '89c932f313c6437c38f2982869beacc89c2f2246' 2025-12-04T09:21:06.2382613Z Submodule path 'third_party/cpuinfo': checked out 'f858c30bcb16f8effd5ff46996f0514539e17abc' 2025-12-04T09:21:06.2915665Z Submodule path 'third_party/cudnn_frontend': checked out '0b1577c8c83401237d601d0d0db5210506705396' 2025-12-04T09:21:07.0403454Z Submodule path 'third_party/cutlass': checked out 'f88806b1e31dfa579842638740216dd41fc6c588' 2025-12-04T09:21:07.2292817Z Submodule path 'third_party/fbgemm': checked out 'c0b988d39a9e47c794d699f29930ed4d7c7e13a4' 2025-12-04T09:21:07.2319279Z Submodule 'external/asmjit' (https://github.com/asmjit/asmjit.git) registered for path 'third_party/fbgemm/external/asmjit' 2025-12-04T09:21:07.2322730Z Submodule 'external/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:21:07.2326465Z Submodule 'external/cpuinfo' (https://github.com/pytorch/cpuinfo) registered for path 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:21:07.2330199Z Submodule 'external/cutlass' (https://github.com/jwfromm/cutlass) registered for path 'third_party/fbgemm/external/cutlass' 2025-12-04T09:21:07.2334038Z Submodule 'external/googletest' (https://github.com/google/googletest) registered for path 'third_party/fbgemm/external/googletest' 2025-12-04T09:21:07.2338088Z Submodule 'external/hipify_torch' (https://github.com/ROCmSoftwarePlatform/hipify_torch.git) registered for path 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:21:07.2342462Z Submodule 'external/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/fbgemm/external/json' 2025-12-04T09:21:07.2378294Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/asmjit'... 2025-12-04T09:21:08.3846418Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/hipify_torch'... 2025-12-04T09:21:08.3847120Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cpuinfo'... 2025-12-04T09:21:08.3847762Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/googletest'... 2025-12-04T09:21:08.4847655Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/composable_kernel'... 2025-12-04T09:21:11.3406174Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cutlass'... 2025-12-04T09:21:11.4407700Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/json'... 2025-12-04T09:21:13.6025741Z Submodule path 'third_party/fbgemm/external/asmjit': checked out 'a3199e8857792cd10b7589ff5d58343d2c9008ea' 2025-12-04T09:21:14.0382150Z Submodule path 'third_party/fbgemm/external/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-12-04T09:21:14.1454429Z Submodule path 'third_party/fbgemm/external/cpuinfo': checked out '6543fec09b2f04ac4a666882998b534afc9c1349' 2025-12-04T09:21:14.8828357Z Submodule path 'third_party/fbgemm/external/cutlass': checked out '98125ce499b0fdf7ffbe0e3052f5b8709f4840f8' 2025-12-04T09:21:14.9345923Z Submodule path 'third_party/fbgemm/external/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T09:21:14.9504998Z Submodule path 'third_party/fbgemm/external/hipify_torch': checked out '63b6a7b541fa7f08f8475ca7d74054db36ff2691' 2025-12-04T09:21:15.0716894Z Submodule path 'third_party/fbgemm/external/json': checked out '9cca280a4d0ccf0c08f47a99aa71d1b0e52f8d03' 2025-12-04T09:21:15.1579147Z Submodule path 'third_party/flash-attention': checked out '979702c87a8713a8e0a5e9fee122b90d2ef13be5' 2025-12-04T09:21:15.1602623Z Submodule 'csrc/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:21:15.1606293Z Submodule 'csrc/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:21:15.1638959Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/composable_kernel'... 2025-12-04T09:21:19.1558393Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/cutlass'... 2025-12-04T09:21:19.4682325Z Submodule path 'third_party/flash-attention/csrc/composable_kernel': checked out '888317e698e9803c62bd38568abc9e05d7709f33' 2025-12-04T09:21:20.1191735Z Submodule path 'third_party/flash-attention/csrc/cutlass': checked out 'c506e16788cb08416a4a57e11a9067beeee29420' 2025-12-04T09:21:20.2932079Z Submodule path 'third_party/flatbuffers': checked out 'a2cd1ea3b6d3fee220106b5fed3f7ce8da9eb757' 2025-12-04T09:21:20.3275448Z Submodule path 'third_party/fmt': checked out '407c905e45ad75fc29bf0f9bb7c5c2fd3475976f' 2025-12-04T09:21:20.3706387Z Submodule path 'third_party/gemmlowp/gemmlowp': checked out '3fb5c176c17c765a3492cd2f0321b0dab712f350' 2025-12-04T09:21:20.4032454Z Submodule path 'third_party/gloo': checked out '54cbae0d3a67fa890b4c3d9ee162b7860315e341' 2025-12-04T09:21:20.4535831Z Submodule path 'third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T09:21:20.5013141Z Submodule path 'third_party/ideep': checked out '719d8e6cd7f7a0e01b155657526d693acf97c2b3' 2025-12-04T09:21:20.5034856Z Submodule 'mkl-dnn' (https://github.com/intel/mkl-dnn.git) registered for path 'third_party/ideep/mkl-dnn' 2025-12-04T09:21:20.5067429Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep/mkl-dnn'... 2025-12-04T09:21:34.7156962Z Submodule path 'third_party/ideep/mkl-dnn': checked out '8d263e693366ef8db40acc569cc7d8edf644556d' 2025-12-04T09:21:34.7412582Z Submodule path 'third_party/ittapi': checked out 'dec1d23ca65ab069d225dfe40dea14f455170959' 2025-12-04T09:21:34.8302391Z Submodule path 'third_party/kineto': checked out '31f85df8fbd89c188f14ef10f1ec65379786b943' 2025-12-04T09:21:34.8326352Z Submodule 'libkineto/third_party/dynolog' (https://github.com/facebookincubator/dynolog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:21:34.8329889Z Submodule 'libkineto/third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:21:34.8333891Z Submodule 'libkineto/third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:21:34.8366915Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog'... 2025-12-04T09:21:35.7715073Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/fmt'... 2025-12-04T09:21:36.1892046Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/googletest'... 2025-12-04T09:21:36.2858654Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog': checked out 'd2ffe0a4e3acace628db49974246b66fc3e85fb1' 2025-12-04T09:21:36.2881113Z Submodule 'third_party/DCGM' (https://github.com/NVIDIA/DCGM.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:21:36.2884773Z Submodule 'third_party/cpr' (https://github.com/libcpr/cpr.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:21:36.2888576Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:21:36.2892511Z Submodule 'third_party/gflags' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:21:36.2896462Z Submodule 'third_party/glog' (https://github.com/google/glog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:21:36.2900714Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:21:36.2904882Z Submodule 'third_party/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:21:36.2909340Z Submodule 'third_party/pfs' (https://github.com/dtrugman/pfs.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:21:36.2913876Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:21:36.2950025Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM'... 2025-12-04T09:21:38.2314580Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/pfs'... 2025-12-04T09:21:38.2315713Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp'... 2025-12-04T09:21:38.2316770Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags'... 2025-12-04T09:21:38.2317723Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/cpr'... 2025-12-04T09:21:38.2318511Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/glog'... 2025-12-04T09:21:38.2319561Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/googletest'... 2025-12-04T09:21:38.2320372Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/fmt'... 2025-12-04T09:21:38.3315314Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/json'... 2025-12-04T09:21:42.6711388Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM': checked out 'ffde4e54bc7249a6039a5e6b45b395141e1217f9' 2025-12-04T09:21:42.6948618Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr': checked out '871ed52d350214a034f6ef8a3b8f51c5ce1bd400' 2025-12-04T09:21:42.7359836Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt': checked out 'cd4af11efc9c622896a3e4cb599fa28668ca3d05' 2025-12-04T09:21:42.7534363Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags': checked out 'e171aa2d15ed9eb17054558e0b3a6a413bb01067' 2025-12-04T09:21:42.7555841Z Submodule 'doc' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:21:42.7588688Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc'... 2025-12-04T09:21:43.0245258Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc': checked out '8411df715cf522606e3b1aca386ddfc0b63d34b4' 2025-12-04T09:21:43.0474637Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog': checked out 'b33e3bad4c46c8a6345525fd822af355e5ef9446' 2025-12-04T09:21:43.0974129Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T09:21:43.2120335Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json': checked out '4f8fba14066156b73f1189a2b8bd568bde5284c5' 2025-12-04T09:21:43.2337119Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs': checked out 'f68a2fa8ea36c783bdd760371411fcb495aa3150' 2025-12-04T09:21:43.2566014Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp': checked out 'b1234816facfdda29845c46696a02998a4af115a' 2025-12-04T09:21:43.2586933Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:21:43.2590797Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:21:43.2623779Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-12-04T09:21:45.4955624Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest'... 2025-12-04T09:21:45.7487306Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'd7ba35bbb649209c66e582d5a0244ba988a15159' 2025-12-04T09:21:45.8008444Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-12-04T09:21:45.8365011Z Submodule path 'third_party/kineto/libkineto/third_party/fmt': checked out '40626af88bd7df9a5fb80be7b25ac85b122d6c21' 2025-12-04T09:21:45.8854777Z Submodule path 'third_party/kineto/libkineto/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T09:21:45.9501523Z Submodule path 'third_party/kleidiai': checked out 'd7770c89632329a9914ef1a90289917597639cbe' 2025-12-04T09:21:45.9954407Z Submodule path 'third_party/mimalloc': checked out 'fbd8b99c2b828428947d70fdc046bb55609be93e' 2025-12-04T09:21:46.1191731Z Submodule path 'third_party/nlohmann': checked out '55f93686c01528224f448c19128836e7df245f72' 2025-12-04T09:21:46.7005372Z Submodule path 'third_party/onnx': checked out 'e709452ef2bbc1d113faf678c24e6d3467696e83' 2025-12-04T09:21:46.7049240Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/onnx/third_party/pybind11' 2025-12-04T09:21:46.7083076Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx/third_party/pybind11'... 2025-12-04T09:21:47.4423717Z Submodule path 'third_party/onnx/third_party/pybind11': checked out 'a2e59f0e7065404b44dfe92a28aca47ba1378dc4' 2025-12-04T09:21:47.5313489Z Submodule path 'third_party/opentelemetry-cpp': checked out 'a799f4aed9c94b765dcdaabaeab7d5e7e2310878' 2025-12-04T09:21:47.5337292Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark) registered for path 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:21:47.5341022Z Submodule 'third_party/googletest' (https://github.com/google/googletest) registered for path 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:21:47.5344794Z Submodule 'third_party/ms-gsl' (https://github.com/microsoft/GSL) registered for path 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:21:47.5348776Z Submodule 'third_party/nlohmann-json' (https://github.com/nlohmann/json) registered for path 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:21:47.5353546Z Submodule 'third_party/opentelemetry-proto' (https://github.com/open-telemetry/opentelemetry-proto) registered for path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:21:47.5357197Z Submodule 'third_party/opentracing-cpp' (https://github.com/opentracing/opentracing-cpp.git) registered for path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:21:47.5361252Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:21:47.5365442Z Submodule 'tools/vcpkg' (https://github.com/Microsoft/vcpkg) registered for path 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:21:47.5400108Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/benchmark'... 2025-12-04T09:21:47.9371460Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentracing-cpp'... 2025-12-04T09:21:47.9373039Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentelemetry-proto'... 2025-12-04T09:21:47.9374381Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/ms-gsl'... 2025-12-04T09:21:47.9375689Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp'... 2025-12-04T09:21:48.0372050Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/googletest'... 2025-12-04T09:21:48.5526925Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/nlohmann-json'... 2025-12-04T09:21:54.6212002Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/tools/vcpkg'... 2025-12-04T09:21:55.1445520Z Submodule path 'third_party/opentelemetry-cpp/third_party/benchmark': checked out 'd572f4777349d43653b21d6c2fc63020ab326db2' 2025-12-04T09:21:55.1904411Z Submodule path 'third_party/opentelemetry-cpp/third_party/googletest': checked out 'b796f7d44681514f58a683a3a71ff17c94edb0c1' 2025-12-04T09:21:55.2103171Z Submodule path 'third_party/opentelemetry-cpp/third_party/ms-gsl': checked out '6f4529395c5b7c2d661812257cd6780c67e54afa' 2025-12-04T09:21:55.3334325Z Submodule path 'third_party/opentelemetry-cpp/third_party/nlohmann-json': checked out 'bc889afb4c5bf1c0d8ee29ef35eaaf4c8bef8a5d' 2025-12-04T09:21:55.3512716Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto': checked out '4ca4f0335c63cda7ab31ea7ed70d6553aee14dce' 2025-12-04T09:21:55.3715319Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp': checked out '06b57f48ded1fa3bdd3d4346f6ef29e40e08eaf5' 2025-12-04T09:21:55.3935694Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp': checked out 'c9ffcdda9086ffd9e1283ea7a0276d831f3c8a8d' 2025-12-04T09:21:55.3957098Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:21:55.3960890Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:21:55.3993495Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-12-04T09:21:57.3391821Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest'... 2025-12-04T09:21:57.5917830Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'eefb26f82b233268fc98577d265352720d477ba4' 2025-12-04T09:21:57.6428000Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-12-04T09:21:58.3269900Z Submodule path 'third_party/opentelemetry-cpp/tools/vcpkg': checked out '8eb57355a4ffb410a2e94c07b4dca2dffbee8e50' 2025-12-04T09:21:58.3414554Z Submodule path 'third_party/pocketfft': checked out '0fa0ef591e38c2758e3184c6c23e497b9f732ffa' 2025-12-04T09:21:58.6448536Z Submodule path 'third_party/protobuf': checked out 'd1eca4e4b421cd2997495c4b4e65cea6be4e9b8a' 2025-12-04T09:21:58.6475715Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:21:58.6479344Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/protobuf/third_party/googletest' 2025-12-04T09:21:58.6512693Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/benchmark'... 2025-12-04T09:21:59.1558826Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/googletest'... 2025-12-04T09:21:59.4447788Z Submodule path 'third_party/protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' 2025-12-04T09:21:59.5198151Z Submodule path 'third_party/protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' 2025-12-04T09:21:59.5322025Z Submodule path 'third_party/psimd': checked out '072586a71b55b7f8c584153d223e95687148a900' 2025-12-04T09:21:59.5475568Z Submodule path 'third_party/pthreadpool': checked out '4fe0e1e183925bf8cfa6aae24237e724a96479b8' 2025-12-04T09:21:59.5975262Z Submodule path 'third_party/pybind11': checked out 'f5fbe867d2d26e4a0a9177a51f6e568868ad3dc8' 2025-12-04T09:21:59.6309297Z Submodule path 'third_party/python-peachpy': checked out 'f45429b087dd7d5bc78bb40dc7cf06425c252d67' 2025-12-04T09:21:59.6789159Z Submodule path 'third_party/sleef': checked out '5a1d179df9cf652951b59010a2d2075372d67f68' 2025-12-04T09:21:59.7138304Z Submodule path 'third_party/tensorpipe': checked out '2b4cd91092d335a697416b2a3cb398283246849d' 2025-12-04T09:21:59.7160865Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:21:59.7164701Z Submodule 'third_party/libnop' (https://github.com/google/libnop.git) registered for path 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:21:59.7168253Z Submodule 'third_party/libuv' (https://github.com/libuv/libuv.git) registered for path 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:21:59.7172436Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:21:59.7205220Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/googletest'... 2025-12-04T09:22:00.6567064Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libnop'... 2025-12-04T09:22:00.6568237Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11'... 2025-12-04T09:22:00.7540205Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libuv'... 2025-12-04T09:22:00.8143576Z Submodule path 'third_party/tensorpipe/third_party/googletest': checked out 'aee0f9d9b5b87796ee8a0ab26b7587ec30e8858e' 2025-12-04T09:22:00.8348482Z Submodule path 'third_party/tensorpipe/third_party/libnop': checked out '910b55815be16109f04f4180e9adee14fb4ce281' 2025-12-04T09:22:00.9152255Z Submodule path 'third_party/tensorpipe/third_party/libuv': checked out '5152db2cbfeb5582e9c27c5ea1dba2cd9e10759b' 2025-12-04T09:22:00.9502417Z Submodule path 'third_party/tensorpipe/third_party/pybind11': checked out 'a23996fce38ff6ccfbcdc09f1e63f2c4be5ea2ef' 2025-12-04T09:22:00.9522995Z Submodule 'tools/clang' (https://github.com/wjakob/clang-cindex-python3) registered for path 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:22:00.9554970Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11/tools/clang'... 2025-12-04T09:22:01.1586154Z Submodule path 'third_party/tensorpipe/third_party/pybind11/tools/clang': checked out '6a00cbc4a9b8e68b71caf7f774b3f9c753ae84d5' 2025-12-04T09:22:01.1632696Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2025-12-04T09:22:01.1997521Z Entering 'android/libs/fbjni' 2025-12-04T09:22:01.2050532Z Entering 'third_party/FP16' 2025-12-04T09:22:01.2101657Z Entering 'third_party/FXdiv' 2025-12-04T09:22:01.2154683Z Entering 'third_party/NNPACK' 2025-12-04T09:22:01.2207835Z Entering 'third_party/NVTX' 2025-12-04T09:22:01.2263637Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:22:01.2317622Z Entering 'third_party/XNNPACK' 2025-12-04T09:22:01.2381742Z Entering 'third_party/aiter' 2025-12-04T09:22:01.2434939Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:22:01.2494492Z Entering 'third_party/benchmark' 2025-12-04T09:22:01.2548557Z Entering 'third_party/composable_kernel' 2025-12-04T09:22:01.2609098Z Entering 'third_party/cpp-httplib' 2025-12-04T09:22:01.2661196Z Entering 'third_party/cpuinfo' 2025-12-04T09:22:01.2714528Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:22:01.2770614Z Entering 'third_party/cutlass' 2025-12-04T09:22:01.2824087Z Entering 'third_party/fbgemm' 2025-12-04T09:22:01.2876091Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:22:01.2928357Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:22:01.2988967Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:22:01.3041052Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:22:01.3100161Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:22:01.3150928Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:22:01.3202688Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:22:01.3262646Z Entering 'third_party/flash-attention' 2025-12-04T09:22:01.3317011Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:22:01.3375335Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:22:01.3437710Z Entering 'third_party/flatbuffers' 2025-12-04T09:22:01.3496222Z Entering 'third_party/fmt' 2025-12-04T09:22:01.3549474Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:22:01.3601941Z Entering 'third_party/gloo' 2025-12-04T09:22:01.3656648Z Entering 'third_party/googletest' 2025-12-04T09:22:01.3715055Z Entering 'third_party/ideep' 2025-12-04T09:22:01.3769746Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:22:01.3830124Z Entering 'third_party/ittapi' 2025-12-04T09:22:01.3883231Z Entering 'third_party/kineto' 2025-12-04T09:22:01.3936678Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:22:01.3989356Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:22:01.4042189Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:22:01.4094253Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:22:01.4147925Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:22:01.4197293Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:22:01.4252857Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:22:01.4309342Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:22:01.4361258Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:22:01.4412718Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:22:01.4467889Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:22:01.4519169Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:01.4574063Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:01.4632190Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:22:01.4683852Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:22:01.4744571Z Entering 'third_party/kleidiai' 2025-12-04T09:22:01.4799367Z Entering 'third_party/mimalloc' 2025-12-04T09:22:01.4851450Z Entering 'third_party/nlohmann' 2025-12-04T09:22:01.4909504Z Entering 'third_party/onnx' 2025-12-04T09:22:01.4975987Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:22:01.5033820Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:22:01.5087442Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:22:01.5138772Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:22:01.5190695Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:22:01.5240604Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:22:01.5292568Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:22:01.5342797Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:22:01.5392570Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:22:01.5441221Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:01.5494090Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:01.5551192Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:22:01.5621966Z Entering 'third_party/pocketfft' 2025-12-04T09:22:01.5675255Z Entering 'third_party/protobuf' 2025-12-04T09:22:01.5728135Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:22:01.5780782Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:22:01.5833948Z Entering 'third_party/psimd' 2025-12-04T09:22:01.5885133Z Entering 'third_party/pthreadpool' 2025-12-04T09:22:01.5942319Z Entering 'third_party/pybind11' 2025-12-04T09:22:01.5993171Z Entering 'third_party/python-peachpy' 2025-12-04T09:22:01.6043505Z Entering 'third_party/sleef' 2025-12-04T09:22:01.6093447Z Entering 'third_party/tensorpipe' 2025-12-04T09:22:01.6145027Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:22:01.6197122Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:22:01.6251147Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:22:01.6301884Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:22:01.6349637Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:22:01.6417203Z ##[endgroup] 2025-12-04T09:22:01.6417628Z ##[group]Persisting credentials for submodules 2025-12-04T09:22:01.6423890Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || :" 2025-12-04T09:22:01.6785811Z Entering 'android/libs/fbjni' 2025-12-04T09:22:01.6856613Z Entering 'third_party/FP16' 2025-12-04T09:22:01.6927893Z Entering 'third_party/FXdiv' 2025-12-04T09:22:01.7001001Z Entering 'third_party/NNPACK' 2025-12-04T09:22:01.7069874Z Entering 'third_party/NVTX' 2025-12-04T09:22:01.7143093Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:22:01.7211477Z Entering 'third_party/XNNPACK' 2025-12-04T09:22:01.7294535Z Entering 'third_party/aiter' 2025-12-04T09:22:01.7367350Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:22:01.7449800Z Entering 'third_party/benchmark' 2025-12-04T09:22:01.7521185Z Entering 'third_party/composable_kernel' 2025-12-04T09:22:01.7599355Z Entering 'third_party/cpp-httplib' 2025-12-04T09:22:01.7673896Z Entering 'third_party/cpuinfo' 2025-12-04T09:22:01.7742004Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:22:01.7810575Z Entering 'third_party/cutlass' 2025-12-04T09:22:01.7893135Z Entering 'third_party/fbgemm' 2025-12-04T09:22:01.7962335Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:22:01.8030876Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:22:01.8108570Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:22:01.8180232Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:22:01.8257089Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:22:01.8328176Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:22:01.8399366Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:22:01.8476311Z Entering 'third_party/flash-attention' 2025-12-04T09:22:01.8545612Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:22:01.8621605Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:22:01.8700334Z Entering 'third_party/flatbuffers' 2025-12-04T09:22:01.8772554Z Entering 'third_party/fmt' 2025-12-04T09:22:01.8841676Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:22:01.8910838Z Entering 'third_party/gloo' 2025-12-04T09:22:01.8984478Z Entering 'third_party/googletest' 2025-12-04T09:22:01.9050983Z Entering 'third_party/ideep' 2025-12-04T09:22:01.9117533Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:22:01.9197899Z Entering 'third_party/ittapi' 2025-12-04T09:22:01.9270918Z Entering 'third_party/kineto' 2025-12-04T09:22:01.9341562Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:22:01.9409472Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:22:01.9486097Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:22:01.9563350Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:22:01.9632268Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:22:01.9701501Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:22:01.9775113Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:22:01.9850509Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:22:01.9921279Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:22:01.9992329Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:22:02.0061308Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:22:02.0129195Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:02.0203552Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:02.0278621Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:22:02.0348897Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:22:02.0423258Z Entering 'third_party/kleidiai' 2025-12-04T09:22:02.0496104Z Entering 'third_party/mimalloc' 2025-12-04T09:22:02.0568055Z Entering 'third_party/nlohmann' 2025-12-04T09:22:02.0641801Z Entering 'third_party/onnx' 2025-12-04T09:22:02.0725649Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:22:02.0796698Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:22:02.0873132Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:22:02.0940869Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:22:02.1007990Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:22:02.1084814Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:22:02.1152732Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:22:02.1221160Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:22:02.1289274Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:22:02.1356982Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:02.1430354Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:02.1503477Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:22:02.1590533Z Entering 'third_party/pocketfft' 2025-12-04T09:22:02.1662561Z Entering 'third_party/protobuf' 2025-12-04T09:22:02.1732944Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:22:02.1801527Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:22:02.1874710Z Entering 'third_party/psimd' 2025-12-04T09:22:02.1941390Z Entering 'third_party/pthreadpool' 2025-12-04T09:22:02.2010515Z Entering 'third_party/pybind11' 2025-12-04T09:22:02.2084822Z Entering 'third_party/python-peachpy' 2025-12-04T09:22:02.2158611Z Entering 'third_party/sleef' 2025-12-04T09:22:02.2232066Z Entering 'third_party/tensorpipe' 2025-12-04T09:22:02.2299187Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:22:02.2368003Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:22:02.2440528Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:22:02.2509108Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:22:02.2581491Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:22:02.2676507Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url" 2025-12-04T09:22:02.3026764Z Entering 'android/libs/fbjni' 2025-12-04T09:22:02.3103326Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-12-04T09:22:02.3120218Z Entering 'third_party/FP16' 2025-12-04T09:22:02.3187110Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-12-04T09:22:02.3208489Z Entering 'third_party/FXdiv' 2025-12-04T09:22:02.3277859Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-12-04T09:22:02.3298981Z Entering 'third_party/NNPACK' 2025-12-04T09:22:02.3361990Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-12-04T09:22:02.3382004Z Entering 'third_party/NVTX' 2025-12-04T09:22:02.3446025Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-12-04T09:22:02.3468388Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:22:02.3535377Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-12-04T09:22:02.3557944Z Entering 'third_party/XNNPACK' 2025-12-04T09:22:02.3626633Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-12-04T09:22:02.3661579Z Entering 'third_party/aiter' 2025-12-04T09:22:02.3722884Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-12-04T09:22:02.3743626Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:22:02.3806841Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-12-04T09:22:02.3838318Z Entering 'third_party/benchmark' 2025-12-04T09:22:02.3908299Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:22:02.3929500Z Entering 'third_party/composable_kernel' 2025-12-04T09:22:02.3991246Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-12-04T09:22:02.4019925Z Entering 'third_party/cpp-httplib' 2025-12-04T09:22:02.4086635Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-12-04T09:22:02.4108224Z Entering 'third_party/cpuinfo' 2025-12-04T09:22:02.4178447Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-12-04T09:22:02.4200184Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:22:02.4265369Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-12-04T09:22:02.4288523Z Entering 'third_party/cutlass' 2025-12-04T09:22:02.4356817Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-12-04T09:22:02.4388335Z Entering 'third_party/fbgemm' 2025-12-04T09:22:02.4455478Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-12-04T09:22:02.4482881Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:22:02.4546387Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-12-04T09:22:02.4568075Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:22:02.4637560Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-12-04T09:22:02.4666070Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:22:02.4729284Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-12-04T09:22:02.4750559Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:22:02.4818679Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-12-04T09:22:02.4848260Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:22:02.4909676Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-12-04T09:22:02.4930632Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:22:02.4996000Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-12-04T09:22:02.5017364Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:22:02.5081181Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-12-04T09:22:02.5105585Z Entering 'third_party/flash-attention' 2025-12-04T09:22:02.5169264Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-12-04T09:22:02.5189683Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:22:02.5255762Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-12-04T09:22:02.5283328Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:22:02.5346379Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-12-04T09:22:02.5377812Z Entering 'third_party/flatbuffers' 2025-12-04T09:22:02.5442338Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-12-04T09:22:02.5466161Z Entering 'third_party/fmt' 2025-12-04T09:22:02.5528855Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-12-04T09:22:02.5550437Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:22:02.5616399Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-12-04T09:22:02.5642727Z Entering 'third_party/gloo' 2025-12-04T09:22:02.5706351Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-12-04T09:22:02.5728424Z Entering 'third_party/googletest' 2025-12-04T09:22:02.5795924Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:22:02.5818096Z Entering 'third_party/ideep' 2025-12-04T09:22:02.5886726Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-12-04T09:22:02.5906858Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:22:02.5969335Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-12-04T09:22:02.5997900Z Entering 'third_party/ittapi' 2025-12-04T09:22:02.6062050Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-12-04T09:22:02.6083551Z Entering 'third_party/kineto' 2025-12-04T09:22:02.6146302Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-12-04T09:22:02.6166808Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:22:02.6236823Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-12-04T09:22:02.6261819Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:22:02.6329554Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-12-04T09:22:02.6352538Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:22:02.6417684Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-12-04T09:22:02.6439521Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:22:02.6508157Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-12-04T09:22:02.6530090Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:22:02.6598666Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-12-04T09:22:02.6616546Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:22:02.6679988Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-12-04T09:22:02.6703968Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:22:02.6768478Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-12-04T09:22:02.6791242Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:22:02.6856819Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:22:02.6879012Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:22:02.6949171Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-12-04T09:22:02.6972484Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:22:02.7037219Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-12-04T09:22:02.7058867Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:22:02.7125674Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T09:22:02.7146286Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:02.7211006Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T09:22:02.7234127Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:02.7300613Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T09:22:02.7326930Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:22:02.7396733Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-12-04T09:22:02.7418811Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:22:02.7486802Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-12-04T09:22:02.7514218Z Entering 'third_party/kleidiai' 2025-12-04T09:22:02.7578466Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-12-04T09:22:02.7600854Z Entering 'third_party/mimalloc' 2025-12-04T09:22:02.7660327Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-12-04T09:22:02.7682204Z Entering 'third_party/nlohmann' 2025-12-04T09:22:02.7745858Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-12-04T09:22:02.7769078Z Entering 'third_party/onnx' 2025-12-04T09:22:02.7837031Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-12-04T09:22:02.7874278Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:22:02.7937564Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:22:02.7963751Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:22:02.8027670Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-12-04T09:22:02.8049263Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:22:02.8116805Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:22:02.8136150Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:22:02.8201903Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:22:02.8222646Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:22:02.8287817Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-12-04T09:22:02.8308939Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:22:02.8377361Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-12-04T09:22:02.8398233Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:22:02.8466917Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-12-04T09:22:02.8488656Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:22:02.8556963Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-12-04T09:22:02.8576643Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:22:02.8640671Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T09:22:02.8659773Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:02.8724022Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T09:22:02.8746251Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:02.8809302Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T09:22:02.8832965Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:22:02.8897672Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-12-04T09:22:02.8935451Z Entering 'third_party/pocketfft' 2025-12-04T09:22:02.8999585Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-12-04T09:22:02.9020798Z Entering 'third_party/protobuf' 2025-12-04T09:22:02.9085563Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-12-04T09:22:02.9109483Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:22:02.9173712Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:22:02.9193930Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:22:02.9256671Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:22:02.9282180Z Entering 'third_party/psimd' 2025-12-04T09:22:02.9346278Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-12-04T09:22:02.9367788Z Entering 'third_party/pthreadpool' 2025-12-04T09:22:02.9436659Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-12-04T09:22:02.9458746Z Entering 'third_party/pybind11' 2025-12-04T09:22:02.9526688Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:22:02.9551661Z Entering 'third_party/python-peachpy' 2025-12-04T09:22:02.9616513Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-12-04T09:22:02.9638204Z Entering 'third_party/sleef' 2025-12-04T09:22:02.9706499Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-12-04T09:22:02.9728309Z Entering 'third_party/tensorpipe' 2025-12-04T09:22:02.9794988Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-12-04T09:22:02.9815663Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:22:02.9878201Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:22:02.9899514Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:22:02.9963766Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-12-04T09:22:02.9985400Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:22:03.0048122Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-12-04T09:22:03.0069663Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:22:03.0136831Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:22:03.0156733Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:22:03.0227374Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-12-04T09:22:03.1299352Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2025-12-04T09:22:03.1656592Z Entering 'android/libs/fbjni' 2025-12-04T09:22:03.1710447Z Entering 'third_party/FP16' 2025-12-04T09:22:03.1763143Z Entering 'third_party/FXdiv' 2025-12-04T09:22:03.1814297Z Entering 'third_party/NNPACK' 2025-12-04T09:22:03.1869032Z Entering 'third_party/NVTX' 2025-12-04T09:22:03.1922928Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:22:03.1975131Z Entering 'third_party/XNNPACK' 2025-12-04T09:22:03.2039636Z Entering 'third_party/aiter' 2025-12-04T09:22:03.2098001Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:22:03.2160188Z Entering 'third_party/benchmark' 2025-12-04T09:22:03.2221299Z Entering 'third_party/composable_kernel' 2025-12-04T09:22:03.2282628Z Entering 'third_party/cpp-httplib' 2025-12-04T09:22:03.2333904Z Entering 'third_party/cpuinfo' 2025-12-04T09:22:03.2390506Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:22:03.2442690Z Entering 'third_party/cutlass' 2025-12-04T09:22:03.2503615Z Entering 'third_party/fbgemm' 2025-12-04T09:22:03.2556530Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:22:03.2610380Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:22:03.2672425Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:22:03.2723231Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:22:03.2784101Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:22:03.2834625Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:22:03.2886992Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:22:03.2945773Z Entering 'third_party/flash-attention' 2025-12-04T09:22:03.2997913Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:22:03.3053041Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:22:03.3118704Z Entering 'third_party/flatbuffers' 2025-12-04T09:22:03.3173657Z Entering 'third_party/fmt' 2025-12-04T09:22:03.3228466Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:22:03.3281787Z Entering 'third_party/gloo' 2025-12-04T09:22:03.3332635Z Entering 'third_party/googletest' 2025-12-04T09:22:03.3386388Z Entering 'third_party/ideep' 2025-12-04T09:22:03.3437694Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:22:03.3499730Z Entering 'third_party/ittapi' 2025-12-04T09:22:03.3552447Z Entering 'third_party/kineto' 2025-12-04T09:22:03.3606134Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:22:03.3660333Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:22:03.3714092Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:22:03.3765005Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:22:03.3818140Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:22:03.3870140Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:22:03.3924262Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:22:03.3976086Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:22:03.4028662Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:22:03.4081595Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:22:03.4133081Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:22:03.4183784Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:03.4236945Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:03.4294555Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:22:03.4344561Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:22:03.4402717Z Entering 'third_party/kleidiai' 2025-12-04T09:22:03.4458615Z Entering 'third_party/mimalloc' 2025-12-04T09:22:03.4510357Z Entering 'third_party/nlohmann' 2025-12-04T09:22:03.4565520Z Entering 'third_party/onnx' 2025-12-04T09:22:03.4632051Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:22:03.4689598Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:22:03.4743117Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:22:03.4798266Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:22:03.4851574Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:22:03.4903495Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:22:03.4955256Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:22:03.5004483Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:22:03.5055194Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:22:03.5106250Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:03.5160949Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:03.5215281Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:22:03.5289955Z Entering 'third_party/pocketfft' 2025-12-04T09:22:03.5342358Z Entering 'third_party/protobuf' 2025-12-04T09:22:03.5397310Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:22:03.5452649Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:22:03.5507923Z Entering 'third_party/psimd' 2025-12-04T09:22:03.5561608Z Entering 'third_party/pthreadpool' 2025-12-04T09:22:03.5614546Z Entering 'third_party/pybind11' 2025-12-04T09:22:03.5666252Z Entering 'third_party/python-peachpy' 2025-12-04T09:22:03.5718418Z Entering 'third_party/sleef' 2025-12-04T09:22:03.5771147Z Entering 'third_party/tensorpipe' 2025-12-04T09:22:03.5822630Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:22:03.5873919Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:22:03.5924074Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:22:03.5972945Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:22:03.6022150Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:22:03.6095783Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2025-12-04T09:22:03.6449033Z Entering 'android/libs/fbjni' 2025-12-04T09:22:03.6502092Z Entering 'third_party/FP16' 2025-12-04T09:22:03.6553608Z Entering 'third_party/FXdiv' 2025-12-04T09:22:03.6603690Z Entering 'third_party/NNPACK' 2025-12-04T09:22:03.6655107Z Entering 'third_party/NVTX' 2025-12-04T09:22:03.6712513Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:22:03.6763148Z Entering 'third_party/XNNPACK' 2025-12-04T09:22:03.6827446Z Entering 'third_party/aiter' 2025-12-04T09:22:03.6881284Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:22:03.6941719Z Entering 'third_party/benchmark' 2025-12-04T09:22:03.6992972Z Entering 'third_party/composable_kernel' 2025-12-04T09:22:03.7051670Z Entering 'third_party/cpp-httplib' 2025-12-04T09:22:03.7104354Z Entering 'third_party/cpuinfo' 2025-12-04T09:22:03.7155625Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:22:03.7210488Z Entering 'third_party/cutlass' 2025-12-04T09:22:03.7270056Z Entering 'third_party/fbgemm' 2025-12-04T09:22:03.7324690Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:22:03.7374524Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:22:03.7432092Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:22:03.7483529Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:22:03.7542277Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:22:03.7593474Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:22:03.7643993Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:22:03.7699226Z Entering 'third_party/flash-attention' 2025-12-04T09:22:03.7751628Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:22:03.7809281Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:22:03.7870496Z Entering 'third_party/flatbuffers' 2025-12-04T09:22:03.7925569Z Entering 'third_party/fmt' 2025-12-04T09:22:03.7979350Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:22:03.8032327Z Entering 'third_party/gloo' 2025-12-04T09:22:03.8081980Z Entering 'third_party/googletest' 2025-12-04T09:22:03.8133087Z Entering 'third_party/ideep' 2025-12-04T09:22:03.8189588Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:22:03.8249154Z Entering 'third_party/ittapi' 2025-12-04T09:22:03.8306787Z Entering 'third_party/kineto' 2025-12-04T09:22:03.8359326Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:22:03.8411045Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:22:03.8464636Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:22:03.8515602Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:22:03.8566191Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:22:03.8617579Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:22:03.8676541Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:22:03.8726212Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:22:03.8779973Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:22:03.8834438Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:22:03.8886530Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:22:03.8937085Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:03.8991515Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:03.9049071Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:22:03.9101864Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:22:03.9156031Z Entering 'third_party/kleidiai' 2025-12-04T09:22:03.9206881Z Entering 'third_party/mimalloc' 2025-12-04T09:22:03.9260836Z Entering 'third_party/nlohmann' 2025-12-04T09:22:03.9314519Z Entering 'third_party/onnx' 2025-12-04T09:22:03.9385104Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:22:03.9439117Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:22:03.9493176Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:22:03.9543989Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:22:03.9595370Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:22:03.9645338Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:22:03.9695904Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:22:03.9747580Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:22:03.9800456Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:22:03.9852652Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:03.9905679Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:03.9958058Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:22:04.0031647Z Entering 'third_party/pocketfft' 2025-12-04T09:22:04.0084179Z Entering 'third_party/protobuf' 2025-12-04T09:22:04.0140470Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:22:04.0192486Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:22:04.0246431Z Entering 'third_party/psimd' 2025-12-04T09:22:04.0296575Z Entering 'third_party/pthreadpool' 2025-12-04T09:22:04.0349457Z Entering 'third_party/pybind11' 2025-12-04T09:22:04.0402011Z Entering 'third_party/python-peachpy' 2025-12-04T09:22:04.0453415Z Entering 'third_party/sleef' 2025-12-04T09:22:04.0504691Z Entering 'third_party/tensorpipe' 2025-12-04T09:22:04.0559749Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:22:04.0611863Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:22:04.0662859Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:22:04.0714608Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:22:04.0762747Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:22:04.0832448Z ##[endgroup] 2025-12-04T09:22:04.0872264Z [command]/usr/bin/git log -1 --format=%H 2025-12-04T09:22:04.0896786Z ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:22:04.1002602Z ##[group]Run cd "${GITHUB_WORKSPACE}" 2025-12-04T09:22:04.1002880Z cd "${GITHUB_WORKSPACE}" 2025-12-04T09:22:04.1003093Z # Clean stale submodule dirs 2025-12-04T09:22:04.1003309Z if [ -z "${NO_SUDO}" ]; then 2025-12-04T09:22:04.1003584Z  sudo git submodule foreach --recursive git clean -ffdx 2025-12-04T09:22:04.1003854Z else 2025-12-04T09:22:04.1004069Z  git submodule foreach --recursive git clean -ffdx 2025-12-04T09:22:04.1004322Z fi 2025-12-04T09:22:04.1013628Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:04.1013896Z env: 2025-12-04T09:22:04.1014054Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:04.1014396Z NO_SUDO: true 2025-12-04T09:22:04.1014553Z ##[endgroup] 2025-12-04T09:22:04.1397078Z Entering 'android/libs/fbjni' 2025-12-04T09:22:04.1438654Z Entering 'third_party/FP16' 2025-12-04T09:22:04.1482066Z Entering 'third_party/FXdiv' 2025-12-04T09:22:04.1521092Z Entering 'third_party/NNPACK' 2025-12-04T09:22:04.1563717Z Entering 'third_party/NVTX' 2025-12-04T09:22:04.1611858Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:22:04.1651626Z Entering 'third_party/XNNPACK' 2025-12-04T09:22:04.1792953Z Entering 'third_party/aiter' 2025-12-04T09:22:04.1844414Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:22:04.1978977Z Entering 'third_party/benchmark' 2025-12-04T09:22:04.2021539Z Entering 'third_party/composable_kernel' 2025-12-04T09:22:04.2166386Z Entering 'third_party/cpp-httplib' 2025-12-04T09:22:04.2208450Z Entering 'third_party/cpuinfo' 2025-12-04T09:22:04.2254349Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:22:04.2296208Z Entering 'third_party/cutlass' 2025-12-04T09:22:04.2413683Z Entering 'third_party/fbgemm' 2025-12-04T09:22:04.2487857Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:22:04.2529459Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:22:04.2672627Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:22:04.2714865Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:22:04.2831273Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:22:04.2873488Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:22:04.2910857Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:22:04.2967515Z Entering 'third_party/flash-attention' 2025-12-04T09:22:04.3018935Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:22:04.3144217Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:22:04.3251861Z Entering 'third_party/flatbuffers' 2025-12-04T09:22:04.3347184Z Entering 'third_party/fmt' 2025-12-04T09:22:04.3389546Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:22:04.3432427Z Entering 'third_party/gloo' 2025-12-04T09:22:04.3476988Z Entering 'third_party/googletest' 2025-12-04T09:22:04.3521367Z Entering 'third_party/ideep' 2025-12-04T09:22:04.3560223Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:22:04.3661020Z Entering 'third_party/ittapi' 2025-12-04T09:22:04.3704620Z Entering 'third_party/kineto' 2025-12-04T09:22:04.3749466Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:22:04.3793846Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:22:04.3851643Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:22:04.3894651Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:22:04.3935479Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:22:04.3973606Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:22:04.4012211Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:22:04.4052721Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:22:04.4095332Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:22:04.4146442Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:22:04.4184849Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:22:04.4224374Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:04.4285797Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:04.4338236Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:22:04.4381434Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:22:04.4426327Z Entering 'third_party/kleidiai' 2025-12-04T09:22:04.4476016Z Entering 'third_party/mimalloc' 2025-12-04T09:22:04.4519645Z Entering 'third_party/nlohmann' 2025-12-04T09:22:04.4576196Z Entering 'third_party/onnx' 2025-12-04T09:22:04.4998165Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:22:04.5046242Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:22:04.5120834Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:22:04.5160826Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:22:04.5203991Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:22:04.5242105Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:22:04.5297004Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:22:04.5339290Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:22:04.5380967Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:22:04.5420726Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:22:04.5481317Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:22:04.5525980Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:22:04.5862041Z Entering 'third_party/pocketfft' 2025-12-04T09:22:04.5901788Z Entering 'third_party/protobuf' 2025-12-04T09:22:04.5997510Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:22:04.6039841Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:22:04.6096341Z Entering 'third_party/psimd' 2025-12-04T09:22:04.6141904Z Entering 'third_party/pthreadpool' 2025-12-04T09:22:04.6182568Z Entering 'third_party/pybind11' 2025-12-04T09:22:04.6231217Z Entering 'third_party/python-peachpy' 2025-12-04T09:22:04.6273081Z Entering 'third_party/sleef' 2025-12-04T09:22:04.6317699Z Entering 'third_party/tensorpipe' 2025-12-04T09:22:04.6360349Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:22:04.6402237Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:22:04.6443481Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:22:04.6486969Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:22:04.6527231Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:22:04.6680094Z Prepare all required actions 2025-12-04T09:22:04.6680560Z Getting action download info 2025-12-04T09:22:04.8275041Z ##[group]Run ./.github/actions/setup-linux 2025-12-04T09:22:04.8275276Z env: 2025-12-04T09:22:04.8275431Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:04.8275629Z ##[endgroup] 2025-12-04T09:22:04.8306955Z ##[group]Run set -euo pipefail 2025-12-04T09:22:04.8307203Z set -euo pipefail 2025-12-04T09:22:04.8307413Z function get_ec2_metadata() { 2025-12-04T09:22:04.8307692Z  # Pulled from instance metadata endpoint for EC2 2025-12-04T09:22:04.8308150Z  # see https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/instancedata-data-retrieval.html 2025-12-04T09:22:04.8308561Z  category=$1 2025-12-04T09:22:04.8308819Z  # If it is GCP runner (runner name contains gcp), do not run this 2025-12-04T09:22:04.8309117Z  runner_name_str=i-0a8d72b6d567bdb22 2025-12-04T09:22:04.8309390Z  if [[ -f /.inarc ]]; then 2025-12-04T09:22:04.8309633Z  echo "ARC Runner, no info on ec2 metadata" 2025-12-04T09:22:04.8309918Z  elif [[ $runner_name_str == *"gcp"* ]]; then 2025-12-04T09:22:04.8310251Z  echo "Runner is from Google Cloud Platform, No info on ec2 metadata" 2025-12-04T09:22:04.8310550Z  else 2025-12-04T09:22:04.8311168Z  curl -H "X-aws-ec2-metadata-token: $(curl -s -X PUT "http://169.254.169.254/latest/api/token" -H "X-aws-ec2-metadata-token-ttl-seconds: 30")" -fsSL "http://169.254.169.254/latest/meta-data/${category}" 2025-12-04T09:22:04.8311818Z  fi 2025-12-04T09:22:04.8312161Z } 2025-12-04T09:22:04.8312351Z echo "ami-id: $(get_ec2_metadata ami-id)" 2025-12-04T09:22:04.8312650Z echo "instance-id: $(get_ec2_metadata instance-id)" 2025-12-04T09:22:04.8313007Z echo "instance-type: $(get_ec2_metadata instance-type)" 2025-12-04T09:22:04.8313303Z echo "system info $(uname -a)" 2025-12-04T09:22:04.8321942Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:04.8322217Z env: 2025-12-04T09:22:04.8322372Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:04.8322551Z ##[endgroup] 2025-12-04T09:22:04.8478831Z ami-id: ami-08982f1c5bf93d976 2025-12-04T09:22:04.8588769Z instance-id: i-0a8d72b6d567bdb22 2025-12-04T09:22:04.8700562Z instance-type: g6.4xlarge 2025-12-04T09:22:04.8715154Z system info Linux ip-10-0-27-140.ec2.internal 6.1.150-174.273.amzn2023.x86_64 #1 SMP PREEMPT_DYNAMIC Tue Sep 9 12:21:26 UTC 2025 x86_64 x86_64 x86_64 GNU/Linux 2025-12-04T09:22:04.8734558Z ##[group]Run if [ -f /usr/bin/nvidia-smi ]; then nvidia-smi; fi 2025-12-04T09:22:04.8734930Z if [ -f /usr/bin/nvidia-smi ]; then nvidia-smi; fi 2025-12-04T09:22:04.8742520Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:04.8742789Z env: 2025-12-04T09:22:04.8742945Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:04.8743132Z ##[endgroup] 2025-12-04T09:22:06.3372284Z Thu Dec 4 09:22:06 2025 2025-12-04T09:22:06.3373685Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:22:06.3375484Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-12-04T09:22:06.3377117Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:22:06.3378222Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T09:22:06.3378845Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-12-04T09:22:06.3379467Z | | | MIG M. | 2025-12-04T09:22:06.3379862Z |=========================================+========================+======================| 2025-12-04T09:22:06.3452069Z | 0 NVIDIA L4 Off | 00000000:35:00.0 Off | 0 | 2025-12-04T09:22:06.3452885Z | N/A 41C P0 29W / 72W | 0MiB / 23034MiB | 4% Default | 2025-12-04T09:22:06.3453391Z | | | N/A | 2025-12-04T09:22:06.3453871Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:22:06.3454282Z 2025-12-04T09:22:06.3454472Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:22:06.3454934Z | Processes: | 2025-12-04T09:22:06.3455487Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T09:22:06.3456049Z | ID ID Usage | 2025-12-04T09:22:06.3456507Z |=========================================================================================| 2025-12-04T09:22:06.3457320Z | No running processes found | 2025-12-04T09:22:06.3457863Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:22:06.6707690Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:22:06.6708382Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:22:06.6720321Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:06.6720782Z env: 2025-12-04T09:22:06.6720938Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:06.6721128Z ##[endgroup] 2025-12-04T09:22:06.6780513Z ##[group]Run if systemctl is-active --quiet docker; then 2025-12-04T09:22:06.6780911Z if systemctl is-active --quiet docker; then 2025-12-04T09:22:06.6781337Z  echo "Docker daemon is running..."; 2025-12-04T09:22:06.6781675Z else 2025-12-04T09:22:06.6781988Z  echo "Starting docker daemon..." && sudo systemctl start docker; 2025-12-04T09:22:06.6782370Z fi 2025-12-04T09:22:06.6790017Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:06.6790299Z env: 2025-12-04T09:22:06.6790462Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:06.6790652Z ##[endgroup] 2025-12-04T09:22:06.6885115Z Docker daemon is running... 2025-12-04T09:22:06.6923153Z ##[group]Run nick-fields/retry@v3.0.0 2025-12-04T09:22:06.6923374Z with: 2025-12-04T09:22:06.6923519Z shell: bash 2025-12-04T09:22:06.6923679Z timeout_minutes: 5 2025-12-04T09:22:06.6923863Z max_attempts: 3 2025-12-04T09:22:06.6924029Z retry_wait_seconds: 30 2025-12-04T09:22:06.6925654Z command: AWS_ACCOUNT_ID=$(aws sts get-caller-identity|grep Account|cut -f4 -d\") aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" # For LF Runners we need to make sure we also login to Meta's ECR docker registry too. META_AWS_ACCOUNT_ID=308535385114 if [ "$AWS_ACCOUNT_ID" != "$META_AWS_ACCOUNT_ID" ] ; then aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$META_AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" fi 2025-12-04T09:22:06.6927398Z polling_interval_seconds: 1 2025-12-04T09:22:06.6927606Z warning_on_retry: true 2025-12-04T09:22:06.6927791Z continue_on_error: false 2025-12-04T09:22:06.6927963Z env: 2025-12-04T09:22:06.6928111Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:06.6928302Z AWS_RETRY_MODE: standard 2025-12-04T09:22:06.6928478Z AWS_MAX_ATTEMPTS: 5 2025-12-04T09:22:06.6928668Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:22:06.6928864Z ##[endgroup] 2025-12-04T09:22:07.7420354Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T09:22:07.7421136Z Configure a credential helper to remove this warning. See 2025-12-04T09:22:07.7421576Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T09:22:07.7422034Z 2025-12-04T09:22:07.7422125Z Login Succeeded 2025-12-04T09:22:07.7690178Z Command completed after 1 attempt(s). 2025-12-04T09:22:07.7749697Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:22:07.7750071Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:22:07.7750400Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:22:07.7758609Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:07.7758893Z env: 2025-12-04T09:22:07.7759051Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:07.7759242Z ##[endgroup] 2025-12-04T09:22:07.7860347Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T09:22:07.7860780Z # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T09:22:07.7861110Z # shellcheck disable=SC2046 2025-12-04T09:22:07.7861361Z docker stop $(docker ps -q) || true 2025-12-04T09:22:07.7861618Z # Prune all of the docker images 2025-12-04T09:22:07.7861856Z docker system prune -af 2025-12-04T09:22:07.7869654Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:07.7869937Z env: 2025-12-04T09:22:07.7870099Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:07.7870293Z ##[endgroup] 2025-12-04T09:22:07.8146763Z "docker stop" requires at least 1 argument. 2025-12-04T09:22:07.8147167Z See 'docker stop --help'. 2025-12-04T09:22:07.8147368Z 2025-12-04T09:22:07.8147782Z Usage: docker stop [OPTIONS] CONTAINER [CONTAINER...] 2025-12-04T09:22:07.8148055Z 2025-12-04T09:22:07.8148235Z Stop one or more running containers 2025-12-04T09:22:07.8462963Z Total reclaimed space: 0B 2025-12-04T09:22:07.8611970Z ##[group]Run pytorch/test-infra/.github/actions/calculate-docker-image@main 2025-12-04T09:22:07.8612345Z with: 2025-12-04T09:22:07.8612944Z docker-image-name: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:07.8613589Z use-custom-docker-registry: true 2025-12-04T09:22:07.8613821Z docker-build-dir: .ci/docker 2025-12-04T09:22:07.8614047Z docker-build-script: ./build.sh 2025-12-04T09:22:07.8614256Z working-directory: . 2025-12-04T09:22:07.8614499Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:07.8614781Z force-push: false 2025-12-04T09:22:07.8614948Z env: 2025-12-04T09:22:07.8615091Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:07.8615292Z ##[endgroup] 2025-12-04T09:22:07.8631925Z ##[group]Run set -ex 2025-12-04T09:22:07.8632135Z set -ex 2025-12-04T09:22:07.8632290Z  2025-12-04T09:22:07.8632600Z # If the docker build directory or the build script doesn't exist, the action will 2025-12-04T09:22:07.8633065Z # gracefully return the docker image name as it is. Pulling docker image in Linux 2025-12-04T09:22:07.8633476Z # job could then download the pre-built image as usual 2025-12-04T09:22:07.8634049Z if [[ -d "${DOCKER_BUILD_DIR}" ]] && [[ -f "${DOCKER_BUILD_DIR}/${DOCKER_BUILD_SCRIPT}" ]] && [[ "${USE_CUSTOM_DOCKER_REGISTRY}" == "true" ]]; then 2025-12-04T09:22:07.8634595Z  echo "skip=false" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:07.8634885Z else 2025-12-04T09:22:07.8635108Z  echo "skip=true" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:07.8635475Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:07.8635768Z  2025-12-04T09:22:07.8636161Z  echo "Not using custom ECR registry. Either it was not requested or there is no Docker build script in the ${REPO_NAME} repo..." 2025-12-04T09:22:07.8636608Z  exit 0 2025-12-04T09:22:07.8636754Z fi 2025-12-04T09:22:07.8636899Z  2025-12-04T09:22:07.8637131Z if [[ "${DOCKER_IMAGE_NAME}" == *"${DOCKER_REGISTRY}/${REPO_NAME}"* ]]; then 2025-12-04T09:22:07.8637560Z  # The docker image name already includes the ECR prefix and tag, so we can just 2025-12-04T09:22:07.8637929Z  # use it as it is, but first let's extract the tag 2025-12-04T09:22:07.8638266Z  DOCKER_TAG=$(echo "${DOCKER_IMAGE_NAME}" | awk -F '[:,]' '{print $2}') 2025-12-04T09:22:07.8638625Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:07.8638967Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:07.8639250Z else 2025-12-04T09:22:07.8639435Z  if [[ "${DOCKER_IMAGE_NAME}" == *:* ]]; then 2025-12-04T09:22:07.8639707Z  CUSTOM_TAG_PREFIX=${DOCKER_IMAGE_NAME#*:} 2025-12-04T09:22:07.8639976Z  DOCKER_IMAGE_NAME=${DOCKER_IMAGE_NAME%%:*} 2025-12-04T09:22:07.8640206Z  fi 2025-12-04T09:22:07.8640524Z  DOCKER_TAG=${CUSTOM_TAG_PREFIX:+${CUSTOM_TAG_PREFIX}-}$(git rev-parse HEAD:"${DOCKER_BUILD_DIR}") 2025-12-04T09:22:07.8640950Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:07.8641386Z  echo "docker-image=${DOCKER_REGISTRY}/${REPO_NAME}/${DOCKER_IMAGE_NAME}:${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:07.8641890Z  echo "custom-tag-prefix=${CUSTOM_TAG_PREFIX}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:07.8642193Z fi 2025-12-04T09:22:07.8649918Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:07.8650190Z env: 2025-12-04T09:22:07.8650342Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:07.8650701Z REPO_NAME: pytorch 2025-12-04T09:22:07.8651417Z DOCKER_IMAGE_NAME: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:07.8652060Z DOCKER_BUILD_DIR: .ci/docker 2025-12-04T09:22:07.8652274Z DOCKER_BUILD_SCRIPT: ./build.sh 2025-12-04T09:22:07.8652535Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:07.8652822Z USE_CUSTOM_DOCKER_REGISTRY: true 2025-12-04T09:22:07.8653026Z CUSTOM_TAG_PREFIX: 2025-12-04T09:22:07.8653189Z ##[endgroup] 2025-12-04T09:22:07.8681521Z + [[ -d .ci/docker ]] 2025-12-04T09:22:07.8681816Z + [[ -f .ci/docker/./build.sh ]] 2025-12-04T09:22:07.8682075Z + [[ true == \t\r\u\e ]] 2025-12-04T09:22:07.8682297Z + echo skip=false 2025-12-04T09:22:07.8683306Z + [[ 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a == *\3\0\8\5\3\5\3\8\5\1\1\4\.\d\k\r\.\e\c\r\.\u\s\-\e\a\s\t\-\1\.\a\m\a\z\o\n\a\w\s\.\c\o\m\/\p\y\t\o\r\c\h* ]] 2025-12-04T09:22:07.8689012Z ++ echo 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:07.8689643Z ++ awk -F '[:,]' '{print $2}' 2025-12-04T09:22:07.8714951Z + DOCKER_TAG=pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:07.8715796Z + echo docker-tag=pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:07.8716908Z + echo docker-image=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:07.8738348Z ##[group]Run set +e 2025-12-04T09:22:07.8738559Z set +e 2025-12-04T09:22:07.8738715Z set -x 2025-12-04T09:22:07.8738880Z  2025-12-04T09:22:07.8739030Z login() { 2025-12-04T09:22:07.8739527Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-12-04T09:22:07.8739896Z } 2025-12-04T09:22:07.8740047Z  2025-12-04T09:22:07.8740186Z retry () { 2025-12-04T09:22:07.8740389Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-12-04T09:22:07.8740609Z } 2025-12-04T09:22:07.8740747Z  2025-12-04T09:22:07.8740910Z retry login "${DOCKER_REGISTRY}" 2025-12-04T09:22:07.8741120Z  2025-12-04T09:22:07.8741273Z START_TIME=$(date +%s) 2025-12-04T09:22:07.8741475Z # Wait up to 120 minutes 2025-12-04T09:22:07.8741732Z while [[ $(( $(date +%s) - 7200 )) -lt $START_TIME ]]; do 2025-12-04T09:22:07.8742087Z  # Check if image already exists, if it does then skip building it 2025-12-04T09:22:07.8742424Z  if docker manifest inspect "${DOCKER_IMAGE}"; then 2025-12-04T09:22:07.8742684Z  exit 0 2025-12-04T09:22:07.8742860Z  fi 2025-12-04T09:22:07.8743003Z  2025-12-04T09:22:07.8743274Z  # NB: This flag is used by Docker build workflow to push the image to ECR, so we can 2025-12-04T09:22:07.8743739Z  # use this to differentiate between the Docker build and regular build jobs. For the 2025-12-04T09:22:07.8744190Z  # latter, it will wait for the Docker images to become available before continuing 2025-12-04T09:22:07.8744551Z  if [ "${DOCKER_PUSH:-false}" == "true" ]; then 2025-12-04T09:22:07.8744835Z  # It's a Docker build job, let's build the image 2025-12-04T09:22:07.8745074Z  break 2025-12-04T09:22:07.8745232Z  else 2025-12-04T09:22:07.8745462Z  # It's a regular build job, wait for the image to become available 2025-12-04T09:22:07.8745759Z  sleep 300 2025-12-04T09:22:07.8746117Z  fi 2025-12-04T09:22:07.8746260Z done 2025-12-04T09:22:07.8746410Z  2025-12-04T09:22:07.8746790Z # NB: This part requires a full checkout. Otherwise, the merge base will 2025-12-04T09:22:07.8747192Z # be empty. The default action would be to continue rebuild the image 2025-12-04T09:22:07.8747564Z if [[ "$BASE_REVISION" = "$(git rev-parse HEAD)" ]]; then 2025-12-04T09:22:07.8747880Z  # if we're on the base branch then use the parent commit 2025-12-04T09:22:07.8748159Z  MERGE_BASE=$(git rev-parse HEAD~) 2025-12-04T09:22:07.8748366Z else 2025-12-04T09:22:07.8748591Z  # otherwise we're on a PR, so use the most recent base commit 2025-12-04T09:22:07.8748920Z  MERGE_BASE=$(git merge-base HEAD "$BASE_REVISION") 2025-12-04T09:22:07.8749169Z fi 2025-12-04T09:22:07.8749313Z  2025-12-04T09:22:07.8749478Z if [[ -z "${MERGE_BASE}" ]]; then 2025-12-04T09:22:07.8749737Z  echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:07.8749960Z  2025-12-04T09:22:07.8750296Z  echo "Finding merge base only works with full checkout, please set fetch-depth to 0, continuing ..." 2025-12-04T09:22:07.8750683Z  exit 0 2025-12-04T09:22:07.8750832Z fi 2025-12-04T09:22:07.8750975Z  2025-12-04T09:22:07.8751186Z if ! git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}"; then 2025-12-04T09:22:07.8751667Z  echo "Directory '${DOCKER_BUILD_DIR}' not found in commit $MERGE_BASE, you should rebase onto a more recent commit" 2025-12-04T09:22:07.8752067Z  exit 1 2025-12-04T09:22:07.8752218Z fi 2025-12-04T09:22:07.8752363Z  2025-12-04T09:22:07.8752606Z PREVIOUS_DOCKER_TAG=$(git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}") 2025-12-04T09:22:07.8753070Z # If no image exists but the hash is the same as the previous hash then we should error out here 2025-12-04T09:22:07.8753484Z if [[ "${PREVIOUS_DOCKER_TAG}" == "${DOCKER_TAG}" ]]; then 2025-12-04T09:22:07.8753969Z  echo "WARNING: Something has gone wrong and the previous image isn't available for the merge-base of your branch" 2025-12-04T09:22:07.8754496Z  echo " Will re-build docker image to store in local cache, TTS may be longer" 2025-12-04T09:22:07.8754810Z fi 2025-12-04T09:22:07.8754953Z  2025-12-04T09:22:07.8755127Z echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-12-04T09:22:07.8762203Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:07.8762479Z env: 2025-12-04T09:22:07.8762639Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:07.8762828Z DOCKER_BUILD_DIR: .ci/docker 2025-12-04T09:22:07.8763077Z BASE_REVISION: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:22:07.8763810Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:07.8764819Z DOCKER_TAG: pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:07.8765397Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:07.8765668Z DOCKER_PUSH: 2025-12-04T09:22:07.8765827Z ##[endgroup] 2025-12-04T09:22:07.8792244Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:07.8792661Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:07.8795038Z + aws ecr get-login-password --region us-east-1 2025-12-04T09:22:07.8796343Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:08.3583128Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T09:22:08.3583870Z Configure a credential helper to remove this warning. See 2025-12-04T09:22:08.3584530Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T09:22:08.3585412Z 2025-12-04T09:22:08.3585560Z Login Succeeded 2025-12-04T09:22:08.3605801Z ++ date +%s 2025-12-04T09:22:08.3619326Z + START_TIME=1764840128 2025-12-04T09:22:08.3623131Z ++ date +%s 2025-12-04T09:22:08.3634366Z + [[ 1764832928 -lt 1764840128 ]] 2025-12-04T09:22:08.3635256Z + docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:08.5960054Z { 2025-12-04T09:22:08.5960426Z "schemaVersion": 2, 2025-12-04T09:22:08.5960817Z "mediaType": "application/vnd.docker.distribution.manifest.v2+json", 2025-12-04T09:22:08.5961218Z "config": { 2025-12-04T09:22:08.5961517Z "mediaType": "application/vnd.docker.container.image.v1+json", 2025-12-04T09:22:08.5961879Z "size": 34884, 2025-12-04T09:22:08.5962247Z "digest": "sha256:0ccbf4fab5f55e557ff63d66607494707567ca363d0460db6f3a59a889b43505" 2025-12-04T09:22:08.5962658Z }, 2025-12-04T09:22:08.5962862Z "layers": [ 2025-12-04T09:22:08.5963040Z { 2025-12-04T09:22:08.5963321Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5963707Z "size": 30447951, 2025-12-04T09:22:08.5964089Z "digest": "sha256:63e5bc7682b85ae57a1221210f64d62e7a90b0a30f19af4ca734b8242ae49d63" 2025-12-04T09:22:08.5964498Z }, 2025-12-04T09:22:08.5964627Z { 2025-12-04T09:22:08.5964855Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5965144Z "size": 1554, 2025-12-04T09:22:08.5965430Z "digest": "sha256:67c41d48b15893f99e19b54d49f8def5cf3e2624cc0befd35b66c832ba60cfb5" 2025-12-04T09:22:08.5965760Z }, 2025-12-04T09:22:08.5965902Z { 2025-12-04T09:22:08.5966126Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5966421Z "size": 313275557, 2025-12-04T09:22:08.5966729Z "digest": "sha256:de0c40fd6dde835c1f43badf2ab0799e49860f78d0ea747c631ed1be3e7b21a8" 2025-12-04T09:22:08.5967058Z }, 2025-12-04T09:22:08.5967195Z { 2025-12-04T09:22:08.5967422Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5967706Z "size": 787, 2025-12-04T09:22:08.5968000Z "digest": "sha256:64b33a2ca109dc333701c298b7279bdda9f8027697d1a61bae941994e1a7d000" 2025-12-04T09:22:08.5968332Z }, 2025-12-04T09:22:08.5968463Z { 2025-12-04T09:22:08.5968695Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5968980Z "size": 106, 2025-12-04T09:22:08.5969258Z "digest": "sha256:cc25ea4b715b80ba4075887029594ebe0b9db9d8181373c85f3d0a3a33c3fbde" 2025-12-04T09:22:08.5969581Z }, 2025-12-04T09:22:08.5969715Z { 2025-12-04T09:22:08.5969937Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5970275Z "size": 703, 2025-12-04T09:22:08.5970550Z "digest": "sha256:490bde020ad4ee14be5457382aa1ae0753f581d7b51e2b004041d293d19b56f8" 2025-12-04T09:22:08.5970872Z }, 2025-12-04T09:22:08.5971228Z { 2025-12-04T09:22:08.5971459Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5971741Z "size": 1218, 2025-12-04T09:22:08.5972021Z "digest": "sha256:7790282ab43d0548e33ee30a015d991628ca80d3cb184d49264dcf41cacb8b08" 2025-12-04T09:22:08.5972329Z }, 2025-12-04T09:22:08.5972459Z { 2025-12-04T09:22:08.5972672Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5972944Z "size": 484, 2025-12-04T09:22:08.5973215Z "digest": "sha256:0c05eb2fa7ac6593a5c3585793813f9d541cd6f5bd9c7d5544f384c4d4368289" 2025-12-04T09:22:08.5973531Z }, 2025-12-04T09:22:08.5973651Z { 2025-12-04T09:22:08.5973881Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5974161Z "size": 110362074, 2025-12-04T09:22:08.5974451Z "digest": "sha256:af4393706f6e36dcb99a5de8428c9fc055decf76da09daf6211b87e14a5da44b" 2025-12-04T09:22:08.5974772Z }, 2025-12-04T09:22:08.5974897Z { 2025-12-04T09:22:08.5975110Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5975629Z "size": 4961, 2025-12-04T09:22:08.5975912Z "digest": "sha256:5629cd1d11b1636eb50a386fae3ef3bb0f0b3c0bf16808fd8e5b9279ae5e2c34" 2025-12-04T09:22:08.5976384Z }, 2025-12-04T09:22:08.5976527Z { 2025-12-04T09:22:08.5976745Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5977024Z "size": 1755, 2025-12-04T09:22:08.5977291Z "digest": "sha256:91d2c6dd679ec7e9571d682e3dea091a42753aed9b5e398ae795d813f8e2a171" 2025-12-04T09:22:08.5977606Z }, 2025-12-04T09:22:08.5977737Z { 2025-12-04T09:22:08.5977948Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5978228Z "size": 724, 2025-12-04T09:22:08.5978502Z "digest": "sha256:d2df3c6e9a4296b9974862a939f99537d63e6abf11d3ecd840a51c94e63b5963" 2025-12-04T09:22:08.5978818Z }, 2025-12-04T09:22:08.5978942Z { 2025-12-04T09:22:08.5979280Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5979567Z "size": 543, 2025-12-04T09:22:08.5979826Z "digest": "sha256:a45940460693299cd3020ce863feef894f7b34395a86f9e84011a5de390d2dca" 2025-12-04T09:22:08.5980164Z }, 2025-12-04T09:22:08.5980327Z { 2025-12-04T09:22:08.5980553Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5980837Z "size": 3360558011, 2025-12-04T09:22:08.5981137Z "digest": "sha256:709dbb1dbcab1aa6f4bdb407dd877e412a267b5cc843ae5b0c625f6256b14ca0" 2025-12-04T09:22:08.5981472Z }, 2025-12-04T09:22:08.5981605Z { 2025-12-04T09:22:08.5981821Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5982111Z "size": 32, 2025-12-04T09:22:08.5982398Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:08.5982718Z }, 2025-12-04T09:22:08.5982866Z { 2025-12-04T09:22:08.5983097Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5983382Z "size": 397, 2025-12-04T09:22:08.5983680Z "digest": "sha256:1e5e1c867d5eb3aabbbc645a3a323a4fc2ff701cd5087437f467a836ead2e69d" 2025-12-04T09:22:08.5984010Z }, 2025-12-04T09:22:08.5984138Z { 2025-12-04T09:22:08.5984363Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5984649Z "size": 236835, 2025-12-04T09:22:08.5984952Z "digest": "sha256:708b34a8f89ed0a11f0a3390c3a14eff72a46a528060d5c842b7169b1ff0a398" 2025-12-04T09:22:08.5985269Z }, 2025-12-04T09:22:08.5985395Z { 2025-12-04T09:22:08.5985611Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5985886Z "size": 230, 2025-12-04T09:22:08.5986172Z "digest": "sha256:7fffbfcfa7b160aa1cc64a88b30f50422fbd63c52be7a7a53bb08e55d3dc71cb" 2025-12-04T09:22:08.5986494Z }, 2025-12-04T09:22:08.5986615Z { 2025-12-04T09:22:08.5986830Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5987108Z "size": 3229770, 2025-12-04T09:22:08.5987379Z "digest": "sha256:06d4791ec6fdc098d364a832ba3ad86c947cd381e7b86555839b523015ebeef1" 2025-12-04T09:22:08.5987697Z }, 2025-12-04T09:22:08.5987835Z { 2025-12-04T09:22:08.5988049Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5988330Z "size": 1472, 2025-12-04T09:22:08.5988606Z "digest": "sha256:3ba143c6c6430c82109c0bd184d3f1a69ce53af799f0524c09d53dc762797a0d" 2025-12-04T09:22:08.5988922Z }, 2025-12-04T09:22:08.5989044Z { 2025-12-04T09:22:08.5989262Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5989541Z "size": 481, 2025-12-04T09:22:08.5989810Z "digest": "sha256:c55c8138d4dc07a23fc1b00d9122c5355041a30b24bb1d451d48e4771b922ea0" 2025-12-04T09:22:08.5990125Z }, 2025-12-04T09:22:08.5990252Z { 2025-12-04T09:22:08.5990463Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5990743Z "size": 200, 2025-12-04T09:22:08.5991012Z "digest": "sha256:1057447356027fbb83720d4aa9d3116d62d069d105b5ecc57be8905b0040d52d" 2025-12-04T09:22:08.5991430Z }, 2025-12-04T09:22:08.5991559Z { 2025-12-04T09:22:08.5991778Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5992132Z "size": 606, 2025-12-04T09:22:08.5992417Z "digest": "sha256:a6ad19e5160a5c80cd9c4df174b2ba63119bf07e950cdd6c6e2fc99dd93527f4" 2025-12-04T09:22:08.5992743Z }, 2025-12-04T09:22:08.5992926Z + exit 0 2025-12-04T09:22:08.5993057Z { 2025-12-04T09:22:08.5993287Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5993572Z "size": 7889624449, 2025-12-04T09:22:08.5993854Z "digest": "sha256:7ef54e5262c5271fa4d007b0cf80678aa992de2943cefa2f34805047fffa3fee" 2025-12-04T09:22:08.5994171Z }, 2025-12-04T09:22:08.5994306Z { 2025-12-04T09:22:08.5994516Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5994805Z "size": 829, 2025-12-04T09:22:08.5995077Z "digest": "sha256:293e877559f8e228b620f7099bbfe559502feb685619d33086ecb2e675508620" 2025-12-04T09:22:08.5995394Z }, 2025-12-04T09:22:08.5995526Z { 2025-12-04T09:22:08.5995750Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5996044Z "size": 33451731, 2025-12-04T09:22:08.5996339Z "digest": "sha256:336ce7e0d0b42e5054f2aa92835fc5c0852bd7780f3589505144b4fb3c44ce77" 2025-12-04T09:22:08.5996674Z }, 2025-12-04T09:22:08.5996808Z { 2025-12-04T09:22:08.5997028Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5997314Z "size": 104, 2025-12-04T09:22:08.5997594Z "digest": "sha256:c428e36fc1f0536534367629df6269e11b49e9ea717f08436b156d6af7cd14d5" 2025-12-04T09:22:08.5997904Z }, 2025-12-04T09:22:08.5998032Z { 2025-12-04T09:22:08.5998252Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5998526Z "size": 1495, 2025-12-04T09:22:08.5998807Z "digest": "sha256:abaaf10a99390c20591b868cac8038acc87b48da67250b32c5762aaa836f3ae4" 2025-12-04T09:22:08.5999132Z }, 2025-12-04T09:22:08.5999255Z { 2025-12-04T09:22:08.5999472Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.5999756Z "size": 459040703, 2025-12-04T09:22:08.6000052Z "digest": "sha256:ce1510c422ec113c1742d4ff0b198fd77c6812ae34c0a77a97d0b9870842c8ef" 2025-12-04T09:22:08.6000377Z }, 2025-12-04T09:22:08.6000508Z { 2025-12-04T09:22:08.6000725Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6000998Z "size": 163, 2025-12-04T09:22:08.6001275Z "digest": "sha256:da37c5f3fb2276bbe14274d8bd05990e676965fb2bf1019794112db9b93356d6" 2025-12-04T09:22:08.6001594Z }, 2025-12-04T09:22:08.6001720Z { 2025-12-04T09:22:08.6001936Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6002218Z "size": 346, 2025-12-04T09:22:08.6002492Z "digest": "sha256:e9a896e1b061d247bdb21fc88c2f62dacb45da82845a6e450c67af84c9868855" 2025-12-04T09:22:08.6002814Z }, 2025-12-04T09:22:08.6002959Z { 2025-12-04T09:22:08.6003179Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6003456Z "size": 32, 2025-12-04T09:22:08.6003741Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:08.6004062Z }, 2025-12-04T09:22:08.6004193Z { 2025-12-04T09:22:08.6004424Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6004710Z "size": 106, 2025-12-04T09:22:08.6004985Z "digest": "sha256:7aaa1f2fb91e4996347ce12d78fd22e8baf9d8b8b95b7beb1a6560a01b99ed77" 2025-12-04T09:22:08.6005315Z }, 2025-12-04T09:22:08.6005447Z { 2025-12-04T09:22:08.6005660Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6005943Z "size": 424, 2025-12-04T09:22:08.6006219Z "digest": "sha256:73403aa2879a250702e77af70e9c16de6ef05bb3179a195f3ddd3e9ecbfb35ab" 2025-12-04T09:22:08.6006536Z }, 2025-12-04T09:22:08.6006665Z { 2025-12-04T09:22:08.6006983Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6007268Z "size": 19309373, 2025-12-04T09:22:08.6007670Z "digest": "sha256:c63beb86e833a0ead5e78063c3bf2461b83c30240acbbdd4768ba378358d1ce7" 2025-12-04T09:22:08.6007999Z }, 2025-12-04T09:22:08.6008125Z { 2025-12-04T09:22:08.6008334Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6008610Z "size": 106, 2025-12-04T09:22:08.6008883Z "digest": "sha256:5b35e7308c16e3b6ede4c90b42340490cc7c9b2fa76f6d8b9bc0e7707f2a999d" 2025-12-04T09:22:08.6009197Z }, 2025-12-04T09:22:08.6009327Z { 2025-12-04T09:22:08.6009540Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6009815Z "size": 826, 2025-12-04T09:22:08.6010092Z "digest": "sha256:bc6cea873c9d43aea85a4ce777bd2501bb2f6536d3da762e87efd6753154a087" 2025-12-04T09:22:08.6010420Z }, 2025-12-04T09:22:08.6010543Z { 2025-12-04T09:22:08.6010758Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6011046Z "size": 724, 2025-12-04T09:22:08.6011335Z "digest": "sha256:d2df3c6e9a4296b9974862a939f99537d63e6abf11d3ecd840a51c94e63b5963" 2025-12-04T09:22:08.6011653Z }, 2025-12-04T09:22:08.6011784Z { 2025-12-04T09:22:08.6011997Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6012271Z "size": 149, 2025-12-04T09:22:08.6012549Z "digest": "sha256:aff4f7c127573f6beec3f4f2958cd385ac57eae837a98cf8e941da9a1c0bb609" 2025-12-04T09:22:08.6012877Z }, 2025-12-04T09:22:08.6012999Z { 2025-12-04T09:22:08.6013212Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6013493Z "size": 136, 2025-12-04T09:22:08.6013770Z "digest": "sha256:789ad9f01b04ce6db60f519ea666288d82670b4a1b6e91952b698b60f91f541b" 2025-12-04T09:22:08.6014091Z }, 2025-12-04T09:22:08.6014218Z { 2025-12-04T09:22:08.6014432Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6014711Z "size": 140, 2025-12-04T09:22:08.6014988Z "digest": "sha256:5518eeaaad5cec2e8542a13aed21d03b8f745df6cae4b21536c32475144eddcb" 2025-12-04T09:22:08.6015313Z }, 2025-12-04T09:22:08.6015439Z { 2025-12-04T09:22:08.6015660Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6015942Z "size": 32, 2025-12-04T09:22:08.6016226Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:08.6016552Z }, 2025-12-04T09:22:08.6016695Z { 2025-12-04T09:22:08.6016910Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6017193Z "size": 223, 2025-12-04T09:22:08.6017474Z "digest": "sha256:19bc2afaa008723d0e437df4dfae409a8f4c7c1f0dae4e78792c23f07de22a4d" 2025-12-04T09:22:08.6017792Z }, 2025-12-04T09:22:08.6017919Z { 2025-12-04T09:22:08.6018146Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6018451Z "size": 255, 2025-12-04T09:22:08.6018755Z "digest": "sha256:8f2d412983fcfab8df61cd0ca935e5dc7289d0c1f0d51cc64ddfb937d7831d1f" 2025-12-04T09:22:08.6019145Z }, 2025-12-04T09:22:08.6019277Z { 2025-12-04T09:22:08.6019503Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6019794Z "size": 32, 2025-12-04T09:22:08.6020079Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:08.6020399Z }, 2025-12-04T09:22:08.6020532Z { 2025-12-04T09:22:08.6020751Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6021029Z "size": 106, 2025-12-04T09:22:08.6021305Z "digest": "sha256:4105c5fe8cd8c654c38abb3c963566cd72df5545dad858ebb233953002629c98" 2025-12-04T09:22:08.6021625Z }, 2025-12-04T09:22:08.6021747Z { 2025-12-04T09:22:08.6021976Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6022261Z "size": 312474099, 2025-12-04T09:22:08.6022546Z "digest": "sha256:3e9cb2b1435dbf85af9564e2318525dd7c94187507d8d83b317eb333803d97dc" 2025-12-04T09:22:08.6035120Z }, 2025-12-04T09:22:08.6035276Z { 2025-12-04T09:22:08.6035668Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6035993Z "size": 3212520030, 2025-12-04T09:22:08.6036313Z "digest": "sha256:c6a8abfd84c7ab5e089ef5edb261743494223f177b1801042787bff673dc8eb3" 2025-12-04T09:22:08.6036644Z }, 2025-12-04T09:22:08.6036780Z { 2025-12-04T09:22:08.6037017Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6037308Z "size": 129, 2025-12-04T09:22:08.6037601Z "digest": "sha256:3c675c8f3264b8d81e5b6ce459e8cb0b0fb108a48dd79758407b2baa5fdf2874" 2025-12-04T09:22:08.6037930Z }, 2025-12-04T09:22:08.6038058Z { 2025-12-04T09:22:08.6038286Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6038578Z "size": 880, 2025-12-04T09:22:08.6038868Z "digest": "sha256:6710b1677220c35b8deb1405cda972f2885e49f1981c8cbf51e4f80dc5e08c0d" 2025-12-04T09:22:08.6039194Z }, 2025-12-04T09:22:08.6039335Z { 2025-12-04T09:22:08.6039571Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6039858Z "size": 724, 2025-12-04T09:22:08.6040147Z "digest": "sha256:d2df3c6e9a4296b9974862a939f99537d63e6abf11d3ecd840a51c94e63b5963" 2025-12-04T09:22:08.6040474Z }, 2025-12-04T09:22:08.6040598Z { 2025-12-04T09:22:08.6040827Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6041114Z "size": 139, 2025-12-04T09:22:08.6041391Z "digest": "sha256:4e207476a41b4ee839c65bbe4e8b436a532b9af4b8ec3901c76f60b5d4f458b7" 2025-12-04T09:22:08.6041715Z }, 2025-12-04T09:22:08.6041846Z { 2025-12-04T09:22:08.6042060Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6042343Z "size": 32, 2025-12-04T09:22:08.6042622Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:08.6042945Z }, 2025-12-04T09:22:08.6043066Z { 2025-12-04T09:22:08.6043282Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6043559Z "size": 159, 2025-12-04T09:22:08.6043836Z "digest": "sha256:6ff8104d1ed70b2d5a23a6cfaca2fef11271e236ca17807eb62f1c8729414cfb" 2025-12-04T09:22:08.6044162Z }, 2025-12-04T09:22:08.6044288Z { 2025-12-04T09:22:08.6044505Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6044785Z "size": 1011, 2025-12-04T09:22:08.6045060Z "digest": "sha256:ff02d54c9496636aa3cd738ec840e20f3e193e2d7d9f05b5683f5635c90bfc60" 2025-12-04T09:22:08.6045369Z }, 2025-12-04T09:22:08.6045497Z { 2025-12-04T09:22:08.6045712Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6046003Z "size": 724, 2025-12-04T09:22:08.6046269Z "digest": "sha256:d2df3c6e9a4296b9974862a939f99537d63e6abf11d3ecd840a51c94e63b5963" 2025-12-04T09:22:08.6046585Z }, 2025-12-04T09:22:08.6046714Z { 2025-12-04T09:22:08.6046927Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6047203Z "size": 134, 2025-12-04T09:22:08.6047480Z "digest": "sha256:0fa03eca92d7c8c85249e83d241ac08a0a458045d49fcd0aa79d915c21453c21" 2025-12-04T09:22:08.6047800Z }, 2025-12-04T09:22:08.6047927Z { 2025-12-04T09:22:08.6048141Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6048415Z "size": 32, 2025-12-04T09:22:08.6048691Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:08.6049010Z }, 2025-12-04T09:22:08.6049133Z { 2025-12-04T09:22:08.6049349Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6049628Z "size": 158, 2025-12-04T09:22:08.6049908Z "digest": "sha256:a0d9f0995a5ce28dfcd4239fe7c8f95fc2005d55bdb8f62ad5fbf10da77c5b01" 2025-12-04T09:22:08.6050229Z }, 2025-12-04T09:22:08.6050357Z { 2025-12-04T09:22:08.6050578Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6050957Z "size": 602, 2025-12-04T09:22:08.6051237Z "digest": "sha256:a5f869eb36f8a5c259a63bdbc27f280e50ea8040cadfc239c44589bb6627a7ea" 2025-12-04T09:22:08.6051644Z }, 2025-12-04T09:22:08.6051773Z { 2025-12-04T09:22:08.6051989Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6052269Z "size": 724, 2025-12-04T09:22:08.6052536Z "digest": "sha256:d2df3c6e9a4296b9974862a939f99537d63e6abf11d3ecd840a51c94e63b5963" 2025-12-04T09:22:08.6052868Z }, 2025-12-04T09:22:08.6052997Z { 2025-12-04T09:22:08.6053208Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6053490Z "size": 155, 2025-12-04T09:22:08.6053759Z "digest": "sha256:9a06093e85549848cf3b329b304f1523e545354fbd82f51eaba797120927a042" 2025-12-04T09:22:08.6054073Z }, 2025-12-04T09:22:08.6054198Z { 2025-12-04T09:22:08.6054414Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6054705Z "size": 32, 2025-12-04T09:22:08.6054973Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:08.6055292Z }, 2025-12-04T09:22:08.6055422Z { 2025-12-04T09:22:08.6055630Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6055908Z "size": 188, 2025-12-04T09:22:08.6056176Z "digest": "sha256:8ebd7ec5382a782d56c9e7308e62699773716a7f860c3f6e40be18e4fe337143" 2025-12-04T09:22:08.6056482Z }, 2025-12-04T09:22:08.6056608Z { 2025-12-04T09:22:08.6056821Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6057102Z "size": 1371, 2025-12-04T09:22:08.6057380Z "digest": "sha256:332b6399906c6af732a60077f9ced277f81e80c802d7ea8fa5eee033a95f1103" 2025-12-04T09:22:08.6057692Z }, 2025-12-04T09:22:08.6057820Z { 2025-12-04T09:22:08.6058029Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6058304Z "size": 32, 2025-12-04T09:22:08.6058582Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:08.6058894Z }, 2025-12-04T09:22:08.6059022Z { 2025-12-04T09:22:08.6059337Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6059627Z "size": 136, 2025-12-04T09:22:08.6059905Z "digest": "sha256:102a81f67e13c532cc084067f0e38913d7bceb58837126c981421f3c71dea3dc" 2025-12-04T09:22:08.6060265Z }, 2025-12-04T09:22:08.6060389Z { 2025-12-04T09:22:08.6060606Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6060886Z "size": 528, 2025-12-04T09:22:08.6061163Z "digest": "sha256:bf7f6de84512969ac1d827bf07446accfee26f5ca359271816b9feab80a663a0" 2025-12-04T09:22:08.6061478Z }, 2025-12-04T09:22:08.6061607Z { 2025-12-04T09:22:08.6061827Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6062111Z "size": 32, 2025-12-04T09:22:08.6062390Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:08.6062716Z }, 2025-12-04T09:22:08.6062838Z { 2025-12-04T09:22:08.6063059Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6063351Z "size": 104, 2025-12-04T09:22:08.6063620Z "digest": "sha256:9d62555d6be43da75808492f58913d02f1a10abc41fdc4b15deb046e88cab600" 2025-12-04T09:22:08.6063940Z }, 2025-12-04T09:22:08.6064068Z { 2025-12-04T09:22:08.6064279Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6064565Z "size": 435, 2025-12-04T09:22:08.6064833Z "digest": "sha256:d706fba330956068098c973708298026ed6dda8bf1370d8410b1474069c4d36f" 2025-12-04T09:22:08.6065142Z }, 2025-12-04T09:22:08.6065266Z { 2025-12-04T09:22:08.6065482Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6065759Z "size": 32, 2025-12-04T09:22:08.6066029Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:08.6066458Z }, 2025-12-04T09:22:08.6066585Z { 2025-12-04T09:22:08.6066799Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6067156Z "size": 109, 2025-12-04T09:22:08.6067434Z "digest": "sha256:f365fe2c419a433e91afc422788d12a49f73a52395acf040d59c87e9b70ed60b" 2025-12-04T09:22:08.6067747Z }, 2025-12-04T09:22:08.6067873Z { 2025-12-04T09:22:08.6068089Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6068377Z "size": 1896, 2025-12-04T09:22:08.6068656Z "digest": "sha256:0a14b6f135356e3d4bc8dd390e6d311dc34b1d2684146bfcf367bdea92ae92da" 2025-12-04T09:22:08.6068974Z }, 2025-12-04T09:22:08.6069101Z { 2025-12-04T09:22:08.6069309Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6069593Z "size": 245593148, 2025-12-04T09:22:08.6069875Z "digest": "sha256:bb9b293d51e0a30f8902e5058b0f17d65859101397b54df712e1b27733454f48" 2025-12-04T09:22:08.6070186Z }, 2025-12-04T09:22:08.6070312Z { 2025-12-04T09:22:08.6070538Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6070810Z "size": 106, 2025-12-04T09:22:08.6071419Z "digest": "sha256:85127fd9e56d6a9bab52eee62832e901cec7f8b990b110d11ee6425d14becaad" 2025-12-04T09:22:08.6071754Z }, 2025-12-04T09:22:08.6071880Z { 2025-12-04T09:22:08.6072105Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6072391Z "size": 164, 2025-12-04T09:22:08.6072669Z "digest": "sha256:2ead80154665c0d09a7db73742d1f9b3b7778646d568f2db5e9bb9dd637f6710" 2025-12-04T09:22:08.6072982Z }, 2025-12-04T09:22:08.6073113Z { 2025-12-04T09:22:08.6073329Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6073608Z "size": 7944, 2025-12-04T09:22:08.6073884Z "digest": "sha256:14cec286c62ba9555d80a80c97234c4510648e03d569f17cb0139946a12ab420" 2025-12-04T09:22:08.6074197Z }, 2025-12-04T09:22:08.6074321Z { 2025-12-04T09:22:08.6074546Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6074848Z "size": 8069, 2025-12-04T09:22:08.6075132Z "digest": "sha256:651b4d5ecb71a11cfe9fc2df393a5834aa5ee969eedb20fdc0f52edcd7ca87cd" 2025-12-04T09:22:08.6075461Z }, 2025-12-04T09:22:08.6075591Z { 2025-12-04T09:22:08.6075804Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6076082Z "size": 305, 2025-12-04T09:22:08.6076353Z "digest": "sha256:4f0169281f2b70ff3a4ca4add2726b06273cc93c958c01815ca73830054c03a2" 2025-12-04T09:22:08.6076669Z }, 2025-12-04T09:22:08.6076791Z { 2025-12-04T09:22:08.6077002Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6077294Z "size": 13360689, 2025-12-04T09:22:08.6077584Z "digest": "sha256:c331724bcc8aa3a6c61966afa7a2e4a685d5a69987ee302cf5fa1cec3aaa3c5f" 2025-12-04T09:22:08.6077904Z }, 2025-12-04T09:22:08.6078026Z { 2025-12-04T09:22:08.6078233Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6078515Z "size": 108, 2025-12-04T09:22:08.6078797Z "digest": "sha256:bf53fea04ea20d1ac088fde8a50bea46609a005690d4691113da6cd628fc8e87" 2025-12-04T09:22:08.6079118Z }, 2025-12-04T09:22:08.6079243Z { 2025-12-04T09:22:08.6079456Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6079733Z "size": 54145665, 2025-12-04T09:22:08.6080007Z "digest": "sha256:f659d4ac3583112e771ab25c9deba599993754f8bef7142def49c380f74a9534" 2025-12-04T09:22:08.6080322Z }, 2025-12-04T09:22:08.6080449Z { 2025-12-04T09:22:08.6080655Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:22:08.6080931Z "size": 32, 2025-12-04T09:22:08.6081206Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:22:08.6081523Z } 2025-12-04T09:22:08.6081652Z ] 2025-12-04T09:22:08.6081779Z } 2025-12-04T09:22:08.6118504Z ##[group]Run set -eux 2025-12-04T09:22:08.6118862Z set -eux 2025-12-04T09:22:08.6119165Z # It's ok if this steps fails, it would then be an anonymous user like what we used to have 2025-12-04T09:22:08.6120117Z aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token | jq --raw-output '.SecretString' | jq -r .docker_hub_readonly_token | docker login --username pytorchbot --password-stdin || true 2025-12-04T09:22:08.6128954Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:08.6129234Z env: 2025-12-04T09:22:08.6129386Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:08.6129571Z ##[endgroup] 2025-12-04T09:22:08.6159588Z + aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token 2025-12-04T09:22:08.6160348Z + jq --raw-output .SecretString 2025-12-04T09:22:08.6161378Z + jq -r .docker_hub_readonly_token 2025-12-04T09:22:08.6162390Z + docker login --username pytorchbot --password-stdin 2025-12-04T09:22:09.1288440Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T09:22:09.1289064Z Configure a credential helper to remove this warning. See 2025-12-04T09:22:09.1289603Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T09:22:09.1289954Z 2025-12-04T09:22:09.1290104Z Login Succeeded 2025-12-04T09:22:09.1369915Z ##[group]Run tag=${ECR_DOCKER_IMAGE##*:} 2025-12-04T09:22:09.1370245Z tag=${ECR_DOCKER_IMAGE##*:} 2025-12-04T09:22:09.1370546Z echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}" 2025-12-04T09:22:09.1378958Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:09.1379346Z env: 2025-12-04T09:22:09.1379507Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:09.1380136Z ECR_DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:09.1380773Z ##[endgroup] 2025-12-04T09:22:09.1411074Z docker pull ghcr.io/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:09.1461437Z ##[group]Run pytorch/test-infra/.github/actions/pull-docker-image@main 2025-12-04T09:22:09.1461774Z with: 2025-12-04T09:22:09.1462359Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:09.1463056Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:09.1463340Z env: 2025-12-04T09:22:09.1463487Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:09.1463673Z ##[endgroup] 2025-12-04T09:22:09.1481575Z ##[group]Run set -x 2025-12-04T09:22:09.1481770Z set -x 2025-12-04T09:22:09.1481930Z set +e 2025-12-04T09:22:09.1482083Z  2025-12-04T09:22:09.1482239Z login() { 2025-12-04T09:22:09.1482587Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-12-04T09:22:09.1482974Z } 2025-12-04T09:22:09.1483124Z  2025-12-04T09:22:09.1483294Z retry () { 2025-12-04T09:22:09.1483496Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-12-04T09:22:09.1483724Z } 2025-12-04T09:22:09.1483869Z  2025-12-04T09:22:09.1484034Z retry login "${DOCKER_REGISTRY}" 2025-12-04T09:22:09.1484249Z  2025-12-04T09:22:09.1484609Z IMAGE_SIZE=$(docker manifest inspect "${DOCKER_IMAGE}" | jq '[.layers[].size, .config.size] | add / 1024 / 1024') 2025-12-04T09:22:09.1485078Z echo "Compressed size of image in MB: ${IMAGE_SIZE}" 2025-12-04T09:22:09.1485354Z  2025-12-04T09:22:09.1485505Z set -e 2025-12-04T09:22:09.1485750Z # ignore output since only exit code is used for conditional 2025-12-04T09:22:09.1486103Z # only pull docker image if it's not available locally 2025-12-04T09:22:09.1486503Z if ! docker inspect --type=image "${DOCKER_IMAGE}" >/dev/null 2>/dev/null; then 2025-12-04T09:22:09.1487053Z  retry docker pull "${DOCKER_IMAGE}" 2025-12-04T09:22:09.1487280Z fi 2025-12-04T09:22:09.1494298Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:22:09.1494577Z env: 2025-12-04T09:22:09.1494730Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:22:09.1495340Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:09.1496046Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:09.1496329Z ##[endgroup] 2025-12-04T09:22:09.1523202Z + set +e 2025-12-04T09:22:09.1523656Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:09.1524279Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:09.1526763Z + aws ecr get-login-password --region us-east-1 2025-12-04T09:22:09.1527431Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:22:09.6018462Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T09:22:09.6019485Z Configure a credential helper to remove this warning. See 2025-12-04T09:22:09.6020257Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T09:22:09.6020777Z 2025-12-04T09:22:09.6022361Z Login Succeeded 2025-12-04T09:22:09.6046119Z ++ docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:09.6047058Z ++ jq '[.layers[].size, .config.size] | add / 1024 / 1024' 2025-12-04T09:22:09.7791922Z + IMAGE_SIZE=15313.830066680908 2025-12-04T09:22:09.7792303Z + echo 'Compressed size of image in MB: 15313.830066680908' 2025-12-04T09:22:09.7792687Z Compressed size of image in MB: 15313.830066680908 2025-12-04T09:22:09.7792985Z + set -e 2025-12-04T09:22:09.7794075Z + docker inspect --type=image 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:09.7914061Z + retry docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:09.7915385Z + docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:22:10.0448090Z pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a: Pulling from pytorch/ci-image 2025-12-04T09:22:10.0449101Z 63e5bc7682b8: Pulling fs layer 2025-12-04T09:22:10.0449535Z 67c41d48b158: Pulling fs layer 2025-12-04T09:22:10.0449961Z de0c40fd6dde: Pulling fs layer 2025-12-04T09:22:10.0450451Z 64b33a2ca109: Pulling fs layer 2025-12-04T09:22:10.0450945Z cc25ea4b715b: Pulling fs layer 2025-12-04T09:22:10.0451395Z 490bde020ad4: Pulling fs layer 2025-12-04T09:22:10.0451826Z 7790282ab43d: Pulling fs layer 2025-12-04T09:22:10.0452241Z 0c05eb2fa7ac: Pulling fs layer 2025-12-04T09:22:10.0452592Z af4393706f6e: Pulling fs layer 2025-12-04T09:22:10.0452879Z 5629cd1d11b1: Pulling fs layer 2025-12-04T09:22:10.0453303Z 91d2c6dd679e: Pulling fs layer 2025-12-04T09:22:10.0453653Z d2df3c6e9a42: Pulling fs layer 2025-12-04T09:22:10.0453902Z a45940460693: Pulling fs layer 2025-12-04T09:22:10.0454370Z 709dbb1dbcab: Pulling fs layer 2025-12-04T09:22:10.0454806Z 4f4fb700ef54: Pulling fs layer 2025-12-04T09:22:10.0455236Z 64b33a2ca109: Waiting 2025-12-04T09:22:10.0455602Z 7790282ab43d: Waiting 2025-12-04T09:22:10.0455996Z 1e5e1c867d5e: Pulling fs layer 2025-12-04T09:22:10.0456357Z cc25ea4b715b: Waiting 2025-12-04T09:22:10.0456623Z 490bde020ad4: Waiting 2025-12-04T09:22:10.0456905Z 0c05eb2fa7ac: Waiting 2025-12-04T09:22:10.0457207Z 708b34a8f89e: Pulling fs layer 2025-12-04T09:22:10.0457892Z 7fffbfcfa7b1: Pulling fs layer 2025-12-04T09:22:10.0458141Z d2df3c6e9a42: Waiting 2025-12-04T09:22:10.0458367Z 06d4791ec6fd: Pulling fs layer 2025-12-04T09:22:10.0458677Z 91d2c6dd679e: Waiting 2025-12-04T09:22:10.0458836Z 708b34a8f89e: Waiting 2025-12-04T09:22:10.0458996Z 4f4fb700ef54: Waiting 2025-12-04T09:22:10.0459260Z 5629cd1d11b1: Waiting 2025-12-04T09:22:10.0459518Z 3ba143c6c643: Pulling fs layer 2025-12-04T09:22:10.0459744Z c55c8138d4dc: Pulling fs layer 2025-12-04T09:22:10.0459931Z 709dbb1dbcab: Waiting 2025-12-04T09:22:10.0460103Z 7fffbfcfa7b1: Waiting 2025-12-04T09:22:10.0460369Z 105744735602: Pulling fs layer 2025-12-04T09:22:10.0460560Z a6ad19e5160a: Pulling fs layer 2025-12-04T09:22:10.0460745Z 7ef54e5262c5: Pulling fs layer 2025-12-04T09:22:10.0460918Z 1e5e1c867d5e: Waiting 2025-12-04T09:22:10.0461082Z 293e877559f8: Pulling fs layer 2025-12-04T09:22:10.0461264Z 336ce7e0d0b4: Pulling fs layer 2025-12-04T09:22:10.0461446Z c428e36fc1f0: Pulling fs layer 2025-12-04T09:22:10.0461645Z abaaf10a9939: Pulling fs layer 2025-12-04T09:22:10.0461831Z c55c8138d4dc: Waiting 2025-12-04T09:22:10.0461990Z 3ba143c6c643: Waiting 2025-12-04T09:22:10.0462155Z ce1510c422ec: Pulling fs layer 2025-12-04T09:22:10.0462331Z 105744735602: Waiting 2025-12-04T09:22:10.0462488Z da37c5f3fb22: Pulling fs layer 2025-12-04T09:22:10.0462664Z a6ad19e5160a: Waiting 2025-12-04T09:22:10.0462828Z e9a896e1b061: Pulling fs layer 2025-12-04T09:22:10.0462999Z c428e36fc1f0: Waiting 2025-12-04T09:22:10.0463162Z 336ce7e0d0b4: Waiting 2025-12-04T09:22:10.0463319Z 7ef54e5262c5: Waiting 2025-12-04T09:22:10.0463479Z 7aaa1f2fb91e: Pulling fs layer 2025-12-04T09:22:10.0463661Z ce1510c422ec: Waiting 2025-12-04T09:22:10.0463905Z 73403aa2879a: Pulling fs layer 2025-12-04T09:22:10.0464286Z c63beb86e833: Pulling fs layer 2025-12-04T09:22:10.0464607Z 7aaa1f2fb91e: Waiting 2025-12-04T09:22:10.0464881Z 73403aa2879a: Waiting 2025-12-04T09:22:10.0465159Z 5b35e7308c16: Pulling fs layer 2025-12-04T09:22:10.0465473Z bc6cea873c9d: Pulling fs layer 2025-12-04T09:22:10.0465793Z c63beb86e833: Waiting 2025-12-04T09:22:10.0466348Z da37c5f3fb22: Waiting 2025-12-04T09:22:10.0466632Z aff4f7c12757: Pulling fs layer 2025-12-04T09:22:10.0466825Z 5b35e7308c16: Waiting 2025-12-04T09:22:10.0466993Z bc6cea873c9d: Waiting 2025-12-04T09:22:10.0467163Z 789ad9f01b04: Pulling fs layer 2025-12-04T09:22:10.0467347Z 5518eeaaad5c: Pulling fs layer 2025-12-04T09:22:10.0467527Z aff4f7c12757: Waiting 2025-12-04T09:22:10.0467679Z 789ad9f01b04: Waiting 2025-12-04T09:22:10.0467828Z 5518eeaaad5c: Waiting 2025-12-04T09:22:10.0467992Z 19bc2afaa008: Pulling fs layer 2025-12-04T09:22:10.0468175Z abaaf10a9939: Waiting 2025-12-04T09:22:10.0468333Z 8f2d412983fc: Pulling fs layer 2025-12-04T09:22:10.0468516Z 4105c5fe8cd8: Pulling fs layer 2025-12-04T09:22:10.0468692Z 8f2d412983fc: Waiting 2025-12-04T09:22:10.0468931Z 3e9cb2b1435d: Pulling fs layer 2025-12-04T09:22:10.0469268Z 06d4791ec6fd: Waiting 2025-12-04T09:22:10.0469515Z 4105c5fe8cd8: Waiting 2025-12-04T09:22:10.0469685Z 293e877559f8: Waiting 2025-12-04T09:22:10.0469920Z 19bc2afaa008: Waiting 2025-12-04T09:22:10.0470237Z c6a8abfd84c7: Pulling fs layer 2025-12-04T09:22:10.0470569Z 3c675c8f3264: Pulling fs layer 2025-12-04T09:22:10.0470874Z 6710b1677220: Pulling fs layer 2025-12-04T09:22:10.0471352Z a45940460693: Waiting 2025-12-04T09:22:10.0471652Z 4e207476a41b: Pulling fs layer 2025-12-04T09:22:10.0471864Z 3e9cb2b1435d: Waiting 2025-12-04T09:22:10.0472034Z 6ff8104d1ed7: Pulling fs layer 2025-12-04T09:22:10.0472222Z c6a8abfd84c7: Waiting 2025-12-04T09:22:10.0472449Z 3c675c8f3264: Waiting 2025-12-04T09:22:10.0472744Z ff02d54c9496: Pulling fs layer 2025-12-04T09:22:10.0473068Z 4e207476a41b: Waiting 2025-12-04T09:22:10.0473353Z 6710b1677220: Waiting 2025-12-04T09:22:10.0473640Z 6ff8104d1ed7: Waiting 2025-12-04T09:22:10.0473939Z 0fa03eca92d7: Pulling fs layer 2025-12-04T09:22:10.0474240Z ff02d54c9496: Waiting 2025-12-04T09:22:10.0474542Z a0d9f0995a5c: Pulling fs layer 2025-12-04T09:22:10.0474826Z a5f869eb36f8: Pulling fs layer 2025-12-04T09:22:10.0475184Z 0fa03eca92d7: Waiting 2025-12-04T09:22:10.0475344Z a0d9f0995a5c: Waiting 2025-12-04T09:22:10.0475603Z a5f869eb36f8: Waiting 2025-12-04T09:22:10.0475814Z 9a06093e8554: Pulling fs layer 2025-12-04T09:22:10.0476138Z 8ebd7ec5382a: Pulling fs layer 2025-12-04T09:22:10.0476325Z 332b6399906c: Pulling fs layer 2025-12-04T09:22:10.0476525Z 102a81f67e13: Pulling fs layer 2025-12-04T09:22:10.0476779Z 8ebd7ec5382a: Waiting 2025-12-04T09:22:10.0476930Z 332b6399906c: Waiting 2025-12-04T09:22:10.0477085Z 102a81f67e13: Waiting 2025-12-04T09:22:10.0477254Z bf7f6de84512: Pulling fs layer 2025-12-04T09:22:10.0477435Z 9a06093e8554: Waiting 2025-12-04T09:22:10.0477603Z 9d62555d6be4: Pulling fs layer 2025-12-04T09:22:10.0477791Z d706fba33095: Pulling fs layer 2025-12-04T09:22:10.0477968Z bf7f6de84512: Waiting 2025-12-04T09:22:10.0478118Z 9d62555d6be4: Waiting 2025-12-04T09:22:10.0478280Z f365fe2c419a: Pulling fs layer 2025-12-04T09:22:10.0478459Z d706fba33095: Waiting 2025-12-04T09:22:10.0478613Z 0a14b6f13535: Pulling fs layer 2025-12-04T09:22:10.0478808Z f365fe2c419a: Waiting 2025-12-04T09:22:10.0478978Z bb9b293d51e0: Pulling fs layer 2025-12-04T09:22:10.0479150Z 0a14b6f13535: Waiting 2025-12-04T09:22:10.0479312Z 85127fd9e56d: Pulling fs layer 2025-12-04T09:22:10.0479498Z 2ead80154665: Pulling fs layer 2025-12-04T09:22:10.0479676Z 14cec286c62b: Pulling fs layer 2025-12-04T09:22:10.0479863Z 651b4d5ecb71: Pulling fs layer 2025-12-04T09:22:10.0480049Z 4f0169281f2b: Pulling fs layer 2025-12-04T09:22:10.0480223Z c331724bcc8a: Pulling fs layer 2025-12-04T09:22:10.0480461Z bf53fea04ea2: Pulling fs layer 2025-12-04T09:22:10.0480860Z 14cec286c62b: Waiting 2025-12-04T09:22:10.0481112Z bb9b293d51e0: Waiting 2025-12-04T09:22:10.0481484Z f659d4ac3583: Pulling fs layer 2025-12-04T09:22:10.0481858Z 651b4d5ecb71: Waiting 2025-12-04T09:22:10.0482240Z 4f0169281f2b: Waiting 2025-12-04T09:22:10.0494300Z 85127fd9e56d: Waiting 2025-12-04T09:22:10.0494556Z f659d4ac3583: Waiting 2025-12-04T09:22:10.0494736Z 2ead80154665: Waiting 2025-12-04T09:22:10.0494921Z bf53fea04ea2: Waiting 2025-12-04T09:22:10.0495100Z c331724bcc8a: Waiting 2025-12-04T09:22:10.1319387Z 67c41d48b158: Verifying Checksum 2025-12-04T09:22:10.1319644Z 67c41d48b158: Download complete 2025-12-04T09:22:10.2088144Z 64b33a2ca109: Verifying Checksum 2025-12-04T09:22:10.2088618Z 64b33a2ca109: Download complete 2025-12-04T09:22:10.2812360Z cc25ea4b715b: Verifying Checksum 2025-12-04T09:22:10.2812683Z cc25ea4b715b: Download complete 2025-12-04T09:22:10.3704814Z 490bde020ad4: Verifying Checksum 2025-12-04T09:22:10.3705229Z 490bde020ad4: Download complete 2025-12-04T09:22:10.4096267Z 63e5bc7682b8: Download complete 2025-12-04T09:22:10.4476033Z 7790282ab43d: Download complete 2025-12-04T09:22:10.4838563Z 0c05eb2fa7ac: Verifying Checksum 2025-12-04T09:22:10.4838855Z 0c05eb2fa7ac: Download complete 2025-12-04T09:22:10.5602900Z 5629cd1d11b1: Download complete 2025-12-04T09:22:10.6296904Z 91d2c6dd679e: Verifying Checksum 2025-12-04T09:22:10.6297214Z 91d2c6dd679e: Download complete 2025-12-04T09:22:10.7077069Z d2df3c6e9a42: Download complete 2025-12-04T09:22:10.7601430Z a45940460693: Verifying Checksum 2025-12-04T09:22:10.7601829Z a45940460693: Download complete 2025-12-04T09:22:11.2953021Z 63e5bc7682b8: Pull complete 2025-12-04T09:22:11.3169568Z 67c41d48b158: Pull complete 2025-12-04T09:22:11.5976765Z af4393706f6e: Verifying Checksum 2025-12-04T09:22:11.5977151Z af4393706f6e: Download complete 2025-12-04T09:22:11.6051893Z 4f4fb700ef54: Verifying Checksum 2025-12-04T09:22:11.6052160Z 4f4fb700ef54: Download complete 2025-12-04T09:22:11.6869383Z 1e5e1c867d5e: Verifying Checksum 2025-12-04T09:22:11.6869661Z 1e5e1c867d5e: Download complete 2025-12-04T09:22:11.7736735Z 708b34a8f89e: Verifying Checksum 2025-12-04T09:22:11.7737011Z 708b34a8f89e: Download complete 2025-12-04T09:22:11.8446755Z 7fffbfcfa7b1: Verifying Checksum 2025-12-04T09:22:11.8447180Z 7fffbfcfa7b1: Download complete 2025-12-04T09:22:11.9398220Z 06d4791ec6fd: Verifying Checksum 2025-12-04T09:22:11.9398511Z 06d4791ec6fd: Download complete 2025-12-04T09:22:12.0285630Z 3ba143c6c643: Verifying Checksum 2025-12-04T09:22:12.0285918Z 3ba143c6c643: Download complete 2025-12-04T09:22:12.1098730Z c55c8138d4dc: Download complete 2025-12-04T09:22:12.1790941Z 105744735602: Download complete 2025-12-04T09:22:12.2392615Z a6ad19e5160a: Verifying Checksum 2025-12-04T09:22:12.2392875Z a6ad19e5160a: Download complete 2025-12-04T09:22:13.2228516Z de0c40fd6dde: Verifying Checksum 2025-12-04T09:22:13.2229024Z de0c40fd6dde: Download complete 2025-12-04T09:22:13.3065967Z 293e877559f8: Download complete 2025-12-04T09:22:13.7038808Z 336ce7e0d0b4: Verifying Checksum 2025-12-04T09:22:13.7039228Z 336ce7e0d0b4: Download complete 2025-12-04T09:22:13.7749357Z c428e36fc1f0: Verifying Checksum 2025-12-04T09:22:13.7749650Z c428e36fc1f0: Download complete 2025-12-04T09:22:13.8625130Z abaaf10a9939: Download complete 2025-12-04T09:22:18.5137116Z ce1510c422ec: Verifying Checksum 2025-12-04T09:22:18.5137582Z ce1510c422ec: Download complete 2025-12-04T09:22:18.5876996Z da37c5f3fb22: Download complete 2025-12-04T09:22:18.6698281Z e9a896e1b061: Verifying Checksum 2025-12-04T09:22:18.6698773Z e9a896e1b061: Download complete 2025-12-04T09:22:18.7563520Z 7aaa1f2fb91e: Download complete 2025-12-04T09:22:18.8322044Z 73403aa2879a: Verifying Checksum 2025-12-04T09:22:18.8322760Z 73403aa2879a: Download complete 2025-12-04T09:22:19.1290470Z c63beb86e833: Verifying Checksum 2025-12-04T09:22:19.1291027Z c63beb86e833: Download complete 2025-12-04T09:22:19.2185283Z 5b35e7308c16: Verifying Checksum 2025-12-04T09:22:19.2185735Z 5b35e7308c16: Download complete 2025-12-04T09:22:19.2875909Z bc6cea873c9d: Verifying Checksum 2025-12-04T09:22:19.2876196Z bc6cea873c9d: Download complete 2025-12-04T09:22:19.3522307Z aff4f7c12757: Verifying Checksum 2025-12-04T09:22:19.3522714Z aff4f7c12757: Download complete 2025-12-04T09:22:19.4353287Z 789ad9f01b04: Download complete 2025-12-04T09:22:19.5147432Z 5518eeaaad5c: Download complete 2025-12-04T09:22:19.5937212Z 19bc2afaa008: Verifying Checksum 2025-12-04T09:22:19.5937695Z 19bc2afaa008: Download complete 2025-12-04T09:22:19.6620452Z 8f2d412983fc: Verifying Checksum 2025-12-04T09:22:19.6621045Z 8f2d412983fc: Download complete 2025-12-04T09:22:19.7382282Z 4105c5fe8cd8: Verifying Checksum 2025-12-04T09:22:19.7382844Z 4105c5fe8cd8: Download complete 2025-12-04T09:22:20.4556039Z de0c40fd6dde: Pull complete 2025-12-04T09:22:20.4821765Z 64b33a2ca109: Pull complete 2025-12-04T09:22:20.5101207Z cc25ea4b715b: Pull complete 2025-12-04T09:22:20.5365048Z 490bde020ad4: Pull complete 2025-12-04T09:22:20.5612497Z 7790282ab43d: Pull complete 2025-12-04T09:22:20.5835623Z 0c05eb2fa7ac: Pull complete 2025-12-04T09:22:22.5171362Z af4393706f6e: Pull complete 2025-12-04T09:22:22.7281336Z 5629cd1d11b1: Pull complete 2025-12-04T09:22:22.8904400Z 91d2c6dd679e: Pull complete 2025-12-04T09:22:22.9901261Z d2df3c6e9a42: Pull complete 2025-12-04T09:22:23.0638153Z a45940460693: Pull complete 2025-12-04T09:22:23.7942093Z 3e9cb2b1435d: Verifying Checksum 2025-12-04T09:22:23.7943128Z 3e9cb2b1435d: Download complete 2025-12-04T09:22:52.4330299Z 709dbb1dbcab: Verifying Checksum 2025-12-04T09:22:52.4330799Z 709dbb1dbcab: Download complete 2025-12-04T09:22:52.5134531Z 3c675c8f3264: Download complete 2025-12-04T09:22:52.6135237Z 6710b1677220: Verifying Checksum 2025-12-04T09:22:52.6135626Z 6710b1677220: Download complete 2025-12-04T09:22:52.7028021Z 4e207476a41b: Verifying Checksum 2025-12-04T09:22:52.7028394Z 4e207476a41b: Download complete 2025-12-04T09:22:52.7996773Z 6ff8104d1ed7: Verifying Checksum 2025-12-04T09:22:52.7997069Z 6ff8104d1ed7: Download complete 2025-12-04T09:22:52.8766191Z ff02d54c9496: Download complete 2025-12-04T09:22:52.9469802Z 0fa03eca92d7: Verifying Checksum 2025-12-04T09:22:52.9470247Z 0fa03eca92d7: Download complete 2025-12-04T09:22:53.0603084Z a0d9f0995a5c: Download complete 2025-12-04T09:22:53.1343376Z a5f869eb36f8: Verifying Checksum 2025-12-04T09:22:53.1343729Z a5f869eb36f8: Download complete 2025-12-04T09:22:53.2113101Z 9a06093e8554: Verifying Checksum 2025-12-04T09:22:53.2114084Z 9a06093e8554: Download complete 2025-12-04T09:22:53.2965373Z 8ebd7ec5382a: Verifying Checksum 2025-12-04T09:22:53.2965939Z 8ebd7ec5382a: Download complete 2025-12-04T09:22:53.3787156Z 332b6399906c: Verifying Checksum 2025-12-04T09:22:53.3787574Z 332b6399906c: Download complete 2025-12-04T09:22:53.4912863Z 102a81f67e13: Verifying Checksum 2025-12-04T09:22:53.4913309Z 102a81f67e13: Download complete 2025-12-04T09:22:53.5708218Z bf7f6de84512: Verifying Checksum 2025-12-04T09:22:53.5708549Z bf7f6de84512: Download complete 2025-12-04T09:22:53.6495262Z 9d62555d6be4: Download complete 2025-12-04T09:22:53.7195098Z d706fba33095: Verifying Checksum 2025-12-04T09:22:53.7195568Z d706fba33095: Download complete 2025-12-04T09:22:53.7994914Z f365fe2c419a: Verifying Checksum 2025-12-04T09:22:53.7995282Z f365fe2c419a: Download complete 2025-12-04T09:22:53.9163974Z 0a14b6f13535: Verifying Checksum 2025-12-04T09:22:56.8368683Z bb9b293d51e0: Verifying Checksum 2025-12-04T09:22:56.8369140Z bb9b293d51e0: Download complete 2025-12-04T09:22:56.9030898Z 85127fd9e56d: Verifying Checksum 2025-12-04T09:22:56.9031479Z 85127fd9e56d: Download complete 2025-12-04T09:22:56.9841856Z 2ead80154665: Verifying Checksum 2025-12-04T09:22:56.9842354Z 2ead80154665: Download complete 2025-12-04T09:22:57.0508315Z 14cec286c62b: Download complete 2025-12-04T09:22:57.1367100Z 651b4d5ecb71: Download complete 2025-12-04T09:22:57.2334500Z 4f0169281f2b: Verifying Checksum 2025-12-04T09:22:57.2334912Z 4f0169281f2b: Download complete 2025-12-04T09:22:57.4194932Z c331724bcc8a: Verifying Checksum 2025-12-04T09:22:57.4195428Z c331724bcc8a: Download complete 2025-12-04T09:22:57.5138018Z bf53fea04ea2: Verifying Checksum 2025-12-04T09:22:57.5138329Z bf53fea04ea2: Download complete 2025-12-04T09:22:58.1126885Z f659d4ac3583: Verifying Checksum 2025-12-04T09:22:58.1127192Z f659d4ac3583: Download complete 2025-12-04T09:23:06.4230447Z c6a8abfd84c7: Verifying Checksum 2025-12-04T09:23:06.4230824Z c6a8abfd84c7: Download complete 2025-12-04T09:23:54.4029248Z 7ef54e5262c5: Verifying Checksum 2025-12-04T09:23:54.4029646Z 7ef54e5262c5: Download complete 2025-12-04T09:24:10.0174589Z 709dbb1dbcab: Pull complete 2025-12-04T09:24:10.2413970Z 4f4fb700ef54: Pull complete 2025-12-04T09:24:10.4408024Z 1e5e1c867d5e: Pull complete 2025-12-04T09:24:10.6025907Z 708b34a8f89e: Pull complete 2025-12-04T09:24:10.7453356Z 7fffbfcfa7b1: Pull complete 2025-12-04T09:24:10.9368694Z 06d4791ec6fd: Pull complete 2025-12-04T09:24:11.0711077Z 3ba143c6c643: Pull complete 2025-12-04T09:24:11.2966070Z c55c8138d4dc: Pull complete 2025-12-04T09:24:11.5033195Z 105744735602: Pull complete 2025-12-04T09:24:11.7147341Z a6ad19e5160a: Pull complete 2025-12-04T09:25:41.8603540Z 7ef54e5262c5: Pull complete 2025-12-04T09:25:42.0778509Z 293e877559f8: Pull complete 2025-12-04T09:25:42.6726679Z 336ce7e0d0b4: Pull complete 2025-12-04T09:25:42.8940598Z c428e36fc1f0: Pull complete 2025-12-04T09:25:42.9200626Z abaaf10a9939: Pull complete 2025-12-04T09:25:49.5551043Z ce1510c422ec: Pull complete 2025-12-04T09:25:49.7183138Z da37c5f3fb22: Pull complete 2025-12-04T09:25:49.8072621Z e9a896e1b061: Pull complete 2025-12-04T09:25:50.0803852Z 7aaa1f2fb91e: Pull complete 2025-12-04T09:25:50.2880669Z 73403aa2879a: Pull complete 2025-12-04T09:25:50.6591290Z c63beb86e833: Pull complete 2025-12-04T09:25:50.8585716Z 5b35e7308c16: Pull complete 2025-12-04T09:25:50.9805988Z bc6cea873c9d: Pull complete 2025-12-04T09:25:51.2463110Z aff4f7c12757: Pull complete 2025-12-04T09:25:51.3963162Z 789ad9f01b04: Pull complete 2025-12-04T09:25:51.5674874Z 5518eeaaad5c: Pull complete 2025-12-04T09:25:52.0115319Z 19bc2afaa008: Pull complete 2025-12-04T09:25:52.2202209Z 8f2d412983fc: Pull complete 2025-12-04T09:25:52.6659619Z 4105c5fe8cd8: Pull complete 2025-12-04T09:25:53.7458391Z 3e9cb2b1435d: Pull complete 2025-12-04T09:26:38.9066018Z c6a8abfd84c7: Pull complete 2025-12-04T09:26:39.1237573Z 3c675c8f3264: Pull complete 2025-12-04T09:26:39.3321235Z 6710b1677220: Pull complete 2025-12-04T09:26:39.6459957Z 4e207476a41b: Pull complete 2025-12-04T09:26:39.9968746Z 6ff8104d1ed7: Pull complete 2025-12-04T09:26:40.2046220Z ff02d54c9496: Pull complete 2025-12-04T09:26:40.5871892Z 0fa03eca92d7: Pull complete 2025-12-04T09:26:40.9555377Z a0d9f0995a5c: Pull complete 2025-12-04T09:26:41.1759558Z a5f869eb36f8: Pull complete 2025-12-04T09:26:41.5532672Z 9a06093e8554: Pull complete 2025-12-04T09:26:41.9459297Z 8ebd7ec5382a: Pull complete 2025-12-04T09:26:42.1646143Z 332b6399906c: Pull complete 2025-12-04T09:26:42.5732603Z 102a81f67e13: Pull complete 2025-12-04T09:26:42.7917169Z bf7f6de84512: Pull complete 2025-12-04T09:26:43.1919023Z 9d62555d6be4: Pull complete 2025-12-04T09:26:43.4076409Z d706fba33095: Pull complete 2025-12-04T09:26:43.7858292Z f365fe2c419a: Pull complete 2025-12-04T09:26:44.0006586Z 0a14b6f13535: Pull complete 2025-12-04T09:26:49.7124095Z bb9b293d51e0: Pull complete 2025-12-04T09:26:49.9366199Z 85127fd9e56d: Pull complete 2025-12-04T09:26:50.1230553Z 2ead80154665: Pull complete 2025-12-04T09:26:50.3144914Z 14cec286c62b: Pull complete 2025-12-04T09:26:50.5122358Z 651b4d5ecb71: Pull complete 2025-12-04T09:26:50.7196889Z 4f0169281f2b: Pull complete 2025-12-04T09:26:51.9985186Z c331724bcc8a: Pull complete 2025-12-04T09:26:52.1236983Z bf53fea04ea2: Pull complete 2025-12-04T09:26:53.3460414Z f659d4ac3583: Pull complete 2025-12-04T09:26:53.6064337Z Digest: sha256:26aca0736a4ce89ce91ccedb1f5d90ce14b0dd6252f01598064e66377063a214 2025-12-04T09:26:53.6315806Z Status: Downloaded newer image for 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:26:53.6382784Z 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:26:53.6443023Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:26:53.6443786Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:26:53.6453929Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:26:53.6454209Z env: 2025-12-04T09:26:53.6454373Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:26:53.6454561Z ##[endgroup] 2025-12-04T09:26:53.6608540Z ##[group]Run pytorch/test-infra/.github/actions/setup-nvidia@main 2025-12-04T09:26:53.6608952Z with: 2025-12-04T09:26:53.6609154Z driver-version: 580.82.07 2025-12-04T09:26:53.6609376Z env: 2025-12-04T09:26:53.6609559Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:26:53.6609785Z ##[endgroup] 2025-12-04T09:26:53.6628311Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:26:53.6628985Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:26:53.6636719Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:26:53.6637006Z env: 2025-12-04T09:26:53.6637165Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:26:53.6637347Z ##[endgroup] 2025-12-04T09:26:53.6701924Z ##[group]Run set -euo pipefail 2025-12-04T09:26:53.6702181Z set -euo pipefail 2025-12-04T09:26:53.6702400Z  2025-12-04T09:26:53.6702550Z has_gpu=false 2025-12-04T09:26:53.6702729Z devices="" 2025-12-04T09:26:53.6702900Z  2025-12-04T09:26:53.6703102Z if command -v nvidia-smi >/dev/null 2>&1; then 2025-12-04T09:26:53.6703436Z  if nvidia-smi -L >/tmp/nvidia_devices 2>/dev/null; then 2025-12-04T09:26:53.6703715Z  has_gpu=true 2025-12-04T09:26:53.6703927Z  devices=$(cat /tmp/nvidia_devices) 2025-12-04T09:26:53.6704145Z  fi 2025-12-04T09:26:53.6704297Z fi 2025-12-04T09:26:53.6704444Z  2025-12-04T09:26:53.6704601Z if [ "$has_gpu" = false ]; then 2025-12-04T09:26:53.6705088Z  if ls /dev/nvidia* >/tmp/nvidia_devices 2>/dev/null; then 2025-12-04T09:26:53.6705360Z  has_gpu=true 2025-12-04T09:26:53.6705580Z  devices=$(cat /tmp/nvidia_devices) 2025-12-04T09:26:53.6705797Z  fi 2025-12-04T09:26:53.6705943Z fi 2025-12-04T09:26:53.6706094Z  2025-12-04T09:26:53.6706307Z if [ "$has_gpu" = false ] && command -v lspci >/dev/null 2>&1; then 2025-12-04T09:26:53.6706668Z  if lspci | grep -i 'nvidia' >/tmp/nvidia_devices 2>/dev/null; then 2025-12-04T09:26:53.6706958Z  has_gpu=true 2025-12-04T09:26:53.6707158Z  devices=$(cat /tmp/nvidia_devices) 2025-12-04T09:26:53.6707380Z  fi 2025-12-04T09:26:53.6707527Z fi 2025-12-04T09:26:53.6707667Z  2025-12-04T09:26:53.6707885Z printf 'HAS_NVIDIA=%s\n' "$has_gpu" >> "$GITHUB_OUTPUT" 2025-12-04T09:26:53.6708265Z printf 'DETECTED_DEVICES<> "$GITHUB_OUTPUT" 2025-12-04T09:26:53.6715322Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:26:53.6715591Z env: 2025-12-04T09:26:53.6715748Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:26:53.6715935Z ##[endgroup] 2025-12-04T09:26:55.3390849Z ##[group]Run if [ "${HAS_NVIDIA}" = "true" ]; then 2025-12-04T09:26:55.3391231Z if [ "${HAS_NVIDIA}" = "true" ]; then 2025-12-04T09:26:55.3391560Z  echo "HAS_NVIDIA_GPU=true" >> "${GITHUB_ENV}" 2025-12-04T09:26:55.3392024Z  echo "GPU_FLAG=--gpus all -e NVIDIA_DRIVER_CAPABILITIES=all" >> "${GITHUB_ENV}" 2025-12-04T09:26:55.3392366Z else 2025-12-04T09:26:55.3392570Z  echo "HAS_NVIDIA_GPU=false" >> "${GITHUB_ENV}" 2025-12-04T09:26:55.3392811Z fi 2025-12-04T09:26:55.3402091Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:26:55.3402379Z env: 2025-12-04T09:26:55.3402638Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:26:55.3402830Z HAS_NVIDIA: true 2025-12-04T09:26:55.3403015Z ##[endgroup] 2025-12-04T09:26:55.3490667Z ##[group]Run nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482 2025-12-04T09:26:55.3490974Z with: 2025-12-04T09:26:55.3491126Z timeout_minutes: 10 2025-12-04T09:26:55.3491306Z max_attempts: 3 2025-12-04T09:26:55.3510548Z command: # Is it disgusting to have a full shell script here in this github action? Sure # But is it the best way to make it so that this action relies on nothing else? Absolutely set -eou pipefail DISTRIBUTION=$(. /etc/os-release;echo $ID$VERSION_ID) DRIVER_FN="NVIDIA-Linux-x86_64-${DRIVER_VERSION}.run" install_nvidia_docker2_amzn2() { ( set -x # Needed for yum-config-manager sudo yum install -y yum-utils if [[ "${DISTRIBUTION}" == "amzn2023" ]] ; then YUM_REPO_URL="https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo" else # Amazon Linux 2 YUM_REPO_URL="https://nvidia.github.io/nvidia-docker/${DISTRIBUTION}/nvidia-docker.repo" fi sudo yum-config-manager --add-repo "${YUM_REPO_URL}" sudo yum install -y \ nvidia-container-toolkit-1.17.8 \ libnvidia-container-tools-1.17.8 \ libnvidia-container1-1.17.8 \ nvidia-container-toolkit-base-1.17.8 sudo systemctl restart docker ) } install_nvidia_docker2_ubuntu20() { ( set -x # Install nvidia-driver package if not installed status="$(dpkg-query -W --showformat='${db:Status-Status}' nvidia-docker2 2>&1)" if [ ! $? = 0 ] || [ ! "$status" = installed ]; then sudo apt-get install -y nvidia-container-toolkit-1.17.8 sudo systemctl restart docker fi ) } pre_install_nvidia_driver_amzn2() { ( # Purge any nvidia driver installed from RHEL repo sudo yum remove -y nvidia-driver-latest-dkms ) } install_nvidia_driver_common() { ( # Try to gather more information about the runner and its existing NVIDIA driver if any echo "Before installing NVIDIA driver" lspci lsmod modinfo nvidia || true HAS_NVIDIA_DRIVER=0 # Check if NVIDIA driver has already been installed if [ -x "$(command -v nvidia-smi)" ]; then set +e # The driver exists, check its version next. Also check only the first GPU if there are more than one of them # so that the same driver version is not print over multiple lines INSTALLED_DRIVER_VERSION=$(nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0) NVIDIA_SMI_STATUS=$? if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then echo "Failed to get NVIDIA driver version ($INSTALLED_DRIVER_VERSION). Continuing" elif [ "$INSTALLED_DRIVER_VERSION" != "$DRIVER_VERSION" ]; then echo "NVIDIA driver ($INSTALLED_DRIVER_VERSION) has been installed, but we expect to have $DRIVER_VERSION instead. Continuing" # Turn off persistent mode so that the installation script can unload the kernel module sudo killall nvidia-persistenced || true else HAS_NVIDIA_DRIVER=1 echo "NVIDIA driver ($INSTALLED_DRIVER_VERSION) has already been installed. Skipping NVIDIA driver installation" fi set -e fi if [ "$HAS_NVIDIA_DRIVER" -eq 0 ]; then # CAUTION: this may need to be updated in future if [ "${DISTRIBUTION}" != ubuntu20.04 ]; then sudo yum groupinstall -y "Development Tools" # ensure our kernel install is the same as our underlying kernel, # groupinstall "Development Tools" has a habit of mismatching kernel headers sudo yum install -y "kernel-devel-uname-r == $(uname -r)" sudo modprobe backlight fi sudo curl -fsL -o /tmp/nvidia_driver "https://s3.amazonaws.com/ossci-linux/nvidia_driver/$DRIVER_FN" set +e sudo /bin/bash /tmp/nvidia_driver -s --no-drm NVIDIA_INSTALLATION_STATUS=$? RESET_GPU=0 if [ "$NVIDIA_INSTALLATION_STATUS" -ne 0 ]; then sudo cat /var/log/nvidia-installer.log # Fail to install NVIDIA driver, try to reset the GPU RESET_GPU=1 elif [ -x "$(command -v nvidia-smi)" ]; then # Check again if nvidia-smi works even if the driver installation completes successfully INSTALLED_DRIVER_VERSION=$(nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0) NVIDIA_SMI_STATUS=$? if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then RESET_GPU=1 fi fi if [ "$RESET_GPU" -eq 1 ]; then NVIDIA_DEVICES=$(lspci -D | grep -i NVIDIA | cut -d' ' -f1) # The GPU can get stuck in a failure state if somehow the test crashs the GPU microcode. When this # happens, we'll try to reset all NVIDIA devices https://github.com/pytorch/pytorch/issues/88388 for PCI_ID in $NVIDIA_DEVICES; do DEVICE_ENABLED=$(cat /sys/bus/pci/devices/$PCI_ID/enable) echo "Reseting $PCI_ID (enabled state: $DEVICE_ENABLED)" # This requires sudo permission of course echo "1" | sudo tee /sys/bus/pci/devices/$PCI_ID/reset sleep 1 done fi sudo rm -fv /tmp/nvidia_driver set -e fi ) } post_install_nvidia_driver_common() { ( sudo modprobe nvidia || true echo "After installing NVIDIA driver" lspci lsmod modinfo nvidia || true ( set +e nvidia-smi # NB: Annoyingly, nvidia-smi command returns successfully with return code 0 even in # the case where the driver has already crashed as it still can get the driver version # and some basic information like the bus ID. However, the rest of the information # would be missing (ERR!), for example: # # +-----------------------------------------------------------------------------+ # | NVIDIA-SMI 525.89.02 Driver Version: 525.89.02 CUDA Version: 12.0 | # |-------------------------------+----------------------+----------------------+ # | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | # | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | # | | | MIG M. | # |===============================+======================+======================| # | 0 ERR! Off | 00000000:00:1E.0 Off | ERR! | # |ERR! ERR! ERR! ERR! / ERR! | 4184MiB / 23028MiB | ERR! Default | # | | | ERR! | # +-------------------------------+----------------------+----------------------+ # # +-----------------------------------------------------------------------------+ # | Processes: | # | GPU GI CI PID Type Process name GPU Memory | # | ID ID Usage | # |=============================================================================| # +-----------------------------------------------------------------------------+ # # This should be reported as a failure instead as it will guarantee to fail when # Docker tries to run with --gpus all # # So, the correct check here is to query one of the missing piece of info like # GPU name, so that the command can fail accordingly nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 NVIDIA_SMI_STATUS=$? # Allowable exit statuses for nvidia-smi, see: https://github.com/NVIDIA/gpu-operator/issues/285 if [ "$NVIDIA_SMI_STATUS" -eq 0 ] || [ "$NVIDIA_SMI_STATUS" -eq 14 ]; then echo "INFO: Ignoring allowed status ${NVIDIA_SMI_STATUS}" else echo "ERROR: nvidia-smi exited with unresolved status ${NVIDIA_SMI_STATUS}" exit ${NVIDIA_SMI_STATUS} fi set -e ) ) } install_nvidia_driver_amzn2() { ( set -x pre_install_nvidia_driver_amzn2 install_nvidia_driver_common post_install_nvidia_driver_common ) } install_nvidia_driver_ubuntu20() { ( set -x install_nvidia_driver_common post_install_nvidia_driver_common ) } echo "== Installing nvidia driver ${DRIVER_FN} ==" case "${DISTRIBUTION}" in amzn*) install_nvidia_driver_amzn2 ;; ubuntu20.04) install_nvidia_driver_ubuntu20 ;; *) echo "ERROR: Unknown distribution ${DISTRIBUTION}" exit 1 ;; esac # Install container toolkit based on distribution echo "== Installing nvidia container toolkit for ${DISTRIBUTION} ==" case "${DISTRIBUTION}" in amzn*) install_nvidia_docker2_amzn2 ;; ubuntu20.04) install_nvidia_docker2_ubuntu20 ;; *) echo "ERROR: Unknown distribution ${DISTRIBUTION}" exit 1 ;; esac # Fix https://github.com/NVIDIA/nvidia-docker/issues/1648 on runners with # more than one GPUs. This just needs to be run once. The command fails # on subsequent runs and complains that the mode is already on, but that's # ok sudo nvidia-persistenced || true # This should show persistence mode ON nvidia-smi # check if the container-toolkit is correctly installed and CUDA is available inside a container docker run --rm -t --gpus=all public.ecr.aws/docker/library/python:3.13 nvidia-smi 2025-12-04T09:26:55.3529869Z retry_wait_seconds: 10 2025-12-04T09:26:55.3530069Z polling_interval_seconds: 1 2025-12-04T09:26:55.3530268Z warning_on_retry: true 2025-12-04T09:26:55.3530472Z continue_on_error: false 2025-12-04T09:26:55.3530660Z env: 2025-12-04T09:26:55.3530808Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:26:55.3530995Z HAS_NVIDIA_GPU: true 2025-12-04T09:26:55.3531224Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:26:55.3531488Z DRIVER_VERSION: 580.82.07 2025-12-04T09:26:55.3531684Z ##[endgroup] 2025-12-04T09:26:55.4224462Z == Installing nvidia driver NVIDIA-Linux-x86_64-580.82.07.run == 2025-12-04T09:26:55.4225928Z + pre_install_nvidia_driver_amzn2 2025-12-04T09:26:55.4227601Z + sudo yum remove -y nvidia-driver-latest-dkms 2025-12-04T09:26:56.0367219Z No match for argument: nvidia-driver-latest-dkms 2025-12-04T09:26:56.0367954Z No packages marked for removal. 2025-12-04T09:26:56.0424721Z Dependencies resolved. 2025-12-04T09:26:56.0434552Z Nothing to do. 2025-12-04T09:26:56.0435650Z Complete! 2025-12-04T09:26:56.1511103Z + install_nvidia_driver_common 2025-12-04T09:26:56.1513855Z + echo 'Before installing NVIDIA driver' 2025-12-04T09:26:56.1515553Z Before installing NVIDIA driver 2025-12-04T09:26:56.1517952Z + lspci 2025-12-04T09:26:56.2419398Z 00:00.0 Host bridge: Intel Corporation 440FX - 82441FX PMC [Natoma] 2025-12-04T09:26:56.2420175Z 00:01.0 ISA bridge: Intel Corporation 82371SB PIIX3 ISA [Natoma/Triton II] 2025-12-04T09:26:56.2420759Z 00:01.3 Non-VGA unclassified device: Intel Corporation 82371AB/EB/MB PIIX4 ACPI (rev 08) 2025-12-04T09:26:56.2421266Z 00:03.0 VGA compatible controller: Amazon.com, Inc. Device 1111 2025-12-04T09:26:56.2421886Z 00:04.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe EBS Controller 2025-12-04T09:26:56.2422681Z 01:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2422994Z 02:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2423296Z 03:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2423584Z 03:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2423849Z 03:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2424384Z 03:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2424650Z 03:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2424911Z 03:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2425159Z 03:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2425403Z 03:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2425643Z 03:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2425887Z 03:01.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2426129Z 03:01.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2426372Z 03:01.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2426618Z 03:01.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2426870Z 03:01.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2427128Z 03:01.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2427367Z 03:01.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2427614Z 03:02.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2427861Z 03:02.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2428098Z 03:02.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2428339Z 03:02.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2428579Z 03:02.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2428816Z 03:02.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2429058Z 03:02.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2429300Z 03:02.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2429541Z 03:03.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2429775Z 03:03.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2430166Z 03:03.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2430429Z 03:03.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2430678Z 03:03.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2430927Z 03:03.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2431168Z 03:03.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2431402Z 03:03.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2431646Z 24:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2431892Z 25:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2432135Z 26:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2432385Z 26:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2432641Z 26:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2432887Z 26:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2433123Z 26:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2433369Z 26:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2433611Z 26:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2433847Z 26:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2434097Z 26:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2434419Z 27:00.0 Ethernet controller: Amazon.com, Inc. Elastic Network Adapter (ENA) 2025-12-04T09:26:56.2434761Z 30:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2435005Z 31:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2435258Z 32:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2435578Z 33:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-12-04T09:26:56.2435905Z 34:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:56.2436193Z 35:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-12-04T09:26:56.2436461Z + lsmod 2025-12-04T09:26:56.2473441Z Module Size Used by 2025-12-04T09:26:56.2473764Z nvidia_uvm 1925120 0 2025-12-04T09:26:56.2474031Z nvidia 14286848 1 nvidia_uvm 2025-12-04T09:26:56.2474306Z drm 602112 1 nvidia 2025-12-04T09:26:56.2474810Z drm_panel_orientation_quirks 32768 1 drm 2025-12-04T09:26:56.2475598Z backlight 24576 1 drm 2025-12-04T09:26:56.2475928Z i2c_core 110592 2 nvidia,drm 2025-12-04T09:26:56.2476216Z xt_conntrack 16384 1 2025-12-04T09:26:56.2476462Z nft_chain_nat 16384 3 2025-12-04T09:26:56.2476696Z xt_MASQUERADE 20480 1 2025-12-04T09:26:56.2476977Z nf_nat 57344 2 nft_chain_nat,xt_MASQUERADE 2025-12-04T09:26:56.2477298Z nf_conntrack_netlink 57344 0 2025-12-04T09:26:56.2477679Z nf_conntrack 184320 4 xt_conntrack,nf_nat,nf_conntrack_netlink,xt_MASQUERADE 2025-12-04T09:26:56.2478107Z nf_defrag_ipv6 24576 1 nf_conntrack 2025-12-04T09:26:56.2478390Z nf_defrag_ipv4 16384 1 nf_conntrack 2025-12-04T09:26:56.2478664Z xfrm_user 57344 1 2025-12-04T09:26:56.2478914Z xfrm_algo 16384 1 xfrm_user 2025-12-04T09:26:56.2479190Z xt_addrtype 16384 2 2025-12-04T09:26:56.2479429Z nft_compat 20480 4 2025-12-04T09:26:56.2479716Z nf_tables 311296 57 nft_compat,nft_chain_nat 2025-12-04T09:26:56.2480109Z nfnetlink 20480 4 nft_compat,nf_conntrack_netlink,nf_tables 2025-12-04T09:26:56.2480453Z br_netfilter 36864 0 2025-12-04T09:26:56.2480710Z bridge 323584 1 br_netfilter 2025-12-04T09:26:56.2480983Z stp 16384 1 bridge 2025-12-04T09:26:56.2481238Z llc 16384 2 bridge,stp 2025-12-04T09:26:56.2481496Z overlay 167936 0 2025-12-04T09:26:56.2481722Z tls 139264 0 2025-12-04T09:26:56.2481952Z nls_ascii 16384 1 2025-12-04T09:26:56.2482171Z nls_cp437 20480 1 2025-12-04T09:26:56.2482389Z vfat 24576 1 2025-12-04T09:26:56.2482734Z fat 86016 1 vfat 2025-12-04T09:26:56.2482945Z sunrpc 700416 1 2025-12-04T09:26:56.2483133Z ena 184320 0 2025-12-04T09:26:56.2483313Z i8042 45056 0 2025-12-04T09:26:56.2483495Z serio 28672 3 i8042 2025-12-04T09:26:56.2483703Z ghash_clmulni_intel 16384 0 2025-12-04T09:26:56.2483896Z button 24576 0 2025-12-04T09:26:56.2484073Z sch_fq_codel 20480 9 2025-12-04T09:26:56.2484252Z dm_mod 188416 0 2025-12-04T09:26:56.2484427Z fuse 184320 1 2025-12-04T09:26:56.2484596Z loop 36864 0 2025-12-04T09:26:56.2484786Z configfs 57344 1 2025-12-04T09:26:56.2484973Z dmi_sysfs 20480 0 2025-12-04T09:26:56.2485147Z crc32_pclmul 16384 0 2025-12-04T09:26:56.2485332Z crc32c_intel 24576 0 2025-12-04T09:26:56.2485519Z efivarfs 24576 1 2025-12-04T09:26:56.2485701Z + modinfo nvidia 2025-12-04T09:26:56.2492015Z filename: /lib/modules/6.1.150-174.273.amzn2023.x86_64/kernel/drivers/video/nvidia.ko 2025-12-04T09:26:56.2492403Z import_ns: DMA_BUF 2025-12-04T09:26:56.2492605Z alias: char-major-195-* 2025-12-04T09:26:56.2493009Z version: 580.82.07 2025-12-04T09:26:56.2493425Z supported: external 2025-12-04T09:26:56.2493831Z license: Dual MIT/GPL 2025-12-04T09:26:56.2494310Z firmware: nvidia/580.82.07/gsp_tu10x.bin 2025-12-04T09:26:56.2494880Z firmware: nvidia/580.82.07/gsp_ga10x.bin 2025-12-04T09:26:56.2495316Z srcversion: BA7240A71DCF7DC6FE88C1D 2025-12-04T09:26:56.2495644Z alias: of:N*T*Cnvidia,tegra264-displayC* 2025-12-04T09:26:56.2495980Z alias: of:N*T*Cnvidia,tegra264-display 2025-12-04T09:26:56.2496302Z alias: of:N*T*Cnvidia,tegra234-displayC* 2025-12-04T09:26:56.2496775Z alias: of:N*T*Cnvidia,tegra234-display 2025-12-04T09:26:56.2497351Z alias: pci:v000010DEd*sv*sd*bc06sc80i00* 2025-12-04T09:26:56.2497913Z alias: pci:v000010DEd*sv*sd*bc03sc02i00* 2025-12-04T09:26:56.2498222Z alias: pci:v000010DEd*sv*sd*bc03sc00i00* 2025-12-04T09:26:56.2498505Z depends: i2c-core,drm 2025-12-04T09:26:56.2498739Z retpoline: Y 2025-12-04T09:26:56.2499221Z name: nvidia 2025-12-04T09:26:56.2499672Z vermagic: 6.1.150-174.273.amzn2023.x86_64 SMP preempt mod_unload modversions 2025-12-04T09:26:56.2500513Z parm: NvSwitchRegDwords:NvSwitch regkey (charp) 2025-12-04T09:26:56.2501301Z parm: NvSwitchBlacklist:NvSwitchBlacklist=uuid[,uuid...] (charp) 2025-12-04T09:26:56.2501995Z parm: NVreg_ResmanDebugLevel:int 2025-12-04T09:26:56.2502305Z parm: NVreg_RmLogonRC:int 2025-12-04T09:26:56.2502598Z parm: NVreg_ModifyDeviceFiles:int 2025-12-04T09:26:56.2502891Z parm: NVreg_DeviceFileUID:int 2025-12-04T09:26:56.2503272Z parm: NVreg_DeviceFileGID:int 2025-12-04T09:26:56.2503688Z parm: NVreg_DeviceFileMode:int 2025-12-04T09:26:56.2504173Z parm: NVreg_InitializeSystemMemoryAllocations:int 2025-12-04T09:26:56.2504471Z parm: NVreg_UsePageAttributeTable:int 2025-12-04T09:26:56.2504713Z parm: NVreg_EnablePCIeGen3:int 2025-12-04T09:26:56.2504941Z parm: NVreg_EnableMSI:int 2025-12-04T09:26:56.2505174Z parm: NVreg_EnableStreamMemOPs:int 2025-12-04T09:26:56.2505435Z parm: NVreg_RestrictProfilingToAdminUsers:int 2025-12-04T09:26:56.2505725Z parm: NVreg_PreserveVideoMemoryAllocations:int 2025-12-04T09:26:56.2506005Z parm: NVreg_EnableS0ixPowerManagement:int 2025-12-04T09:26:56.2506305Z parm: NVreg_S0ixPowerManagementVideoMemoryThreshold:int 2025-12-04T09:26:56.2506603Z parm: NVreg_DynamicPowerManagement:int 2025-12-04T09:26:56.2506909Z parm: NVreg_DynamicPowerManagementVideoMemoryThreshold:int 2025-12-04T09:26:56.2507208Z parm: NVreg_EnableGpuFirmware:int 2025-12-04T09:26:56.2507451Z parm: NVreg_EnableGpuFirmwareLogs:int 2025-12-04T09:26:56.2507843Z parm: NVreg_OpenRmEnableUnsupportedGpus:int 2025-12-04T09:26:56.2508117Z parm: NVreg_EnableUserNUMAManagement:int 2025-12-04T09:26:56.2508359Z parm: NVreg_MemoryPoolSize:int 2025-12-04T09:26:56.2508599Z parm: NVreg_KMallocHeapMaxSize:int 2025-12-04T09:26:56.2508839Z parm: NVreg_VMallocHeapMaxSize:int 2025-12-04T09:26:56.2509074Z parm: NVreg_IgnoreMMIOCheck:int 2025-12-04T09:26:56.2509305Z parm: NVreg_NvLinkDisable:int 2025-12-04T09:26:56.2509568Z parm: NVreg_EnablePCIERelaxedOrderingMode:int 2025-12-04T09:26:56.2509844Z parm: NVreg_RegisterPCIDriver:int 2025-12-04T09:26:56.2510113Z parm: NVreg_RegisterPlatformDeviceDriver:int 2025-12-04T09:26:56.2510379Z parm: NVreg_EnableResizableBar:int 2025-12-04T09:26:56.2510630Z parm: NVreg_EnableDbgBreakpoint:int 2025-12-04T09:26:56.2510879Z parm: NVreg_EnableNonblockingOpen:int 2025-12-04T09:26:56.2511151Z parm: NVreg_CoherentGPUMemoryMode:charp 2025-12-04T09:26:56.2511406Z parm: NVreg_RegistryDwords:charp 2025-12-04T09:26:56.2511661Z parm: NVreg_RegistryDwordsPerDevice:charp 2025-12-04T09:26:56.2511899Z parm: NVreg_RmMsg:charp 2025-12-04T09:26:56.2512130Z parm: NVreg_GpuBlacklist:charp 2025-12-04T09:26:56.2512371Z parm: NVreg_TemporaryFilePath:charp 2025-12-04T09:26:56.2512601Z parm: NVreg_ExcludedGpus:charp 2025-12-04T09:26:56.2512834Z parm: NVreg_DmaRemapPeerMmio:int 2025-12-04T09:26:56.2513067Z parm: NVreg_RmNvlinkBandwidth:charp 2025-12-04T09:26:56.2513318Z parm: NVreg_RmNvlinkBandwidthLinkCount:int 2025-12-04T09:26:56.2513572Z parm: NVreg_ImexChannelCount:int 2025-12-04T09:26:56.2513805Z parm: NVreg_CreateImexChannel0:int 2025-12-04T09:26:56.2514051Z parm: NVreg_GrdmaPciTopoCheckOverride:int 2025-12-04T09:26:56.2514298Z parm: rm_firmware_active:charp 2025-12-04T09:26:56.2514528Z + HAS_NVIDIA_DRIVER=0 2025-12-04T09:26:56.2514712Z ++ command -v nvidia-smi 2025-12-04T09:26:56.2514902Z + '[' -x /usr/bin/nvidia-smi ']' 2025-12-04T09:26:56.2515095Z + set +e 2025-12-04T09:26:56.2515424Z ++ nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0 2025-12-04T09:26:57.8882199Z + INSTALLED_DRIVER_VERSION=580.82.07 2025-12-04T09:26:57.8882561Z + NVIDIA_SMI_STATUS=0 2025-12-04T09:26:57.8883478Z + '[' 0 -ne 0 ']' 2025-12-04T09:26:57.8883693Z + '[' 580.82.07 '!=' 580.82.07 ']' 2025-12-04T09:26:57.8883945Z + HAS_NVIDIA_DRIVER=1 2025-12-04T09:26:57.8884387Z + echo 'NVIDIA driver (580.82.07) has already been installed. Skipping NVIDIA driver installation' 2025-12-04T09:26:57.8884741Z + set -e 2025-12-04T09:26:57.8884894Z + '[' 1 -eq 0 ']' 2025-12-04T09:26:57.8885197Z NVIDIA driver (580.82.07) has already been installed. Skipping NVIDIA driver installation 2025-12-04T09:26:57.8885559Z + post_install_nvidia_driver_common 2025-12-04T09:26:57.8888679Z + sudo modprobe nvidia 2025-12-04T09:26:58.0159713Z + echo 'After installing NVIDIA driver' 2025-12-04T09:26:58.0160086Z + lspci 2025-12-04T09:26:58.0160314Z After installing NVIDIA driver 2025-12-04T09:26:58.0363223Z 00:00.0 Host bridge: Intel Corporation 440FX - 82441FX PMC [Natoma] 2025-12-04T09:26:58.0364094Z 00:01.0 ISA bridge: Intel Corporation 82371SB PIIX3 ISA [Natoma/Triton II] 2025-12-04T09:26:58.0364772Z 00:01.3 Non-VGA unclassified device: Intel Corporation 82371AB/EB/MB PIIX4 ACPI (rev 08) 2025-12-04T09:26:58.0365279Z 00:03.0 VGA compatible controller: Amazon.com, Inc. Device 1111 2025-12-04T09:26:58.0365733Z 00:04.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe EBS Controller 2025-12-04T09:26:58.0366151Z 01:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0366480Z 02:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0366803Z 03:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0367107Z 03:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0368058Z 03:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0368582Z 03:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0368975Z 03:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0369290Z 03:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0369609Z 03:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0369919Z 03:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0370220Z 03:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0370514Z 03:01.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0370807Z 03:01.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0371455Z 03:01.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0371760Z 03:01.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0372047Z 03:01.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0372339Z 03:01.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0372646Z 03:01.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0372934Z 03:02.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0373229Z 03:02.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0373529Z 03:02.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0373837Z 03:02.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0374146Z 03:02.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0374404Z 03:02.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0374639Z 03:02.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0374881Z 03:02.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0375120Z 03:03.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0375357Z 03:03.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0375589Z 03:03.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0375825Z 03:03.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0376088Z 03:03.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0376320Z 03:03.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0376559Z 03:03.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0376800Z 03:03.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0377233Z 24:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0377480Z 25:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0377718Z 26:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0377974Z 26:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0378209Z 26:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0378449Z 26:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0378689Z 26:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0378922Z 26:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0379272Z 26:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0379524Z 26:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0379765Z 26:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0380080Z 27:00.0 Ethernet controller: Amazon.com, Inc. Elastic Network Adapter (ENA) 2025-12-04T09:26:58.0380427Z 30:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0380676Z 31:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0380922Z 32:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0381256Z 33:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-12-04T09:26:58.0381584Z 34:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:26:58.0381861Z 35:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-12-04T09:26:58.0382117Z + lsmod 2025-12-04T09:26:58.0404712Z Module Size Used by 2025-12-04T09:26:58.0405187Z nvidia_uvm 1925120 0 2025-12-04T09:26:58.0405599Z nvidia 14286848 1 nvidia_uvm 2025-12-04T09:26:58.0406050Z drm 602112 1 nvidia 2025-12-04T09:26:58.0406813Z drm_panel_orientation_quirks 32768 1 drm 2025-12-04T09:26:58.0407295Z backlight 24576 1 drm 2025-12-04T09:26:58.0407734Z i2c_core 110592 2 nvidia,drm 2025-12-04T09:26:58.0408047Z xt_conntrack 16384 1 2025-12-04T09:26:58.0408291Z nft_chain_nat 16384 3 2025-12-04T09:26:58.0408513Z xt_MASQUERADE 20480 1 2025-12-04T09:26:58.0408785Z nf_nat 57344 2 nft_chain_nat,xt_MASQUERADE 2025-12-04T09:26:58.0409096Z nf_conntrack_netlink 57344 0 2025-12-04T09:26:58.0409456Z nf_conntrack 184320 4 xt_conntrack,nf_nat,nf_conntrack_netlink,xt_MASQUERADE 2025-12-04T09:26:58.0409859Z nf_defrag_ipv6 24576 1 nf_conntrack 2025-12-04T09:26:58.0410143Z nf_defrag_ipv4 16384 1 nf_conntrack 2025-12-04T09:26:58.0410407Z xfrm_user 57344 1 2025-12-04T09:26:58.0410655Z xfrm_algo 16384 1 xfrm_user 2025-12-04T09:26:58.0410935Z xt_addrtype 16384 2 2025-12-04T09:26:58.0411175Z nft_compat 20480 4 2025-12-04T09:26:58.0411447Z nf_tables 311296 57 nft_compat,nft_chain_nat 2025-12-04T09:26:58.0411834Z nfnetlink 20480 4 nft_compat,nf_conntrack_netlink,nf_tables 2025-12-04T09:26:58.0412179Z br_netfilter 36864 0 2025-12-04T09:26:58.0412430Z bridge 323584 1 br_netfilter 2025-12-04T09:26:58.0412704Z stp 16384 1 bridge 2025-12-04T09:26:58.0412962Z llc 16384 2 bridge,stp 2025-12-04T09:26:58.0413214Z overlay 167936 0 2025-12-04T09:26:58.0413432Z tls 139264 0 2025-12-04T09:26:58.0413651Z nls_ascii 16384 1 2025-12-04T09:26:58.0413864Z nls_cp437 20480 1 2025-12-04T09:26:58.0414082Z vfat 24576 1 2025-12-04T09:26:58.0414294Z fat 86016 1 vfat 2025-12-04T09:26:58.0414489Z sunrpc 700416 1 2025-12-04T09:26:58.0414665Z ena 184320 0 2025-12-04T09:26:58.0414845Z i8042 45056 0 2025-12-04T09:26:58.0415027Z serio 28672 3 i8042 2025-12-04T09:26:58.0415226Z ghash_clmulni_intel 16384 0 2025-12-04T09:26:58.0415417Z button 24576 0 2025-12-04T09:26:58.0415603Z sch_fq_codel 20480 9 2025-12-04T09:26:58.0415903Z dm_mod 188416 0 2025-12-04T09:26:58.0416093Z fuse 184320 1 2025-12-04T09:26:58.0416271Z loop 36864 0 2025-12-04T09:26:58.0416443Z configfs 57344 1 2025-12-04T09:26:58.0416623Z dmi_sysfs 20480 0 2025-12-04T09:26:58.0416809Z crc32_pclmul 16384 0 2025-12-04T09:26:58.0416982Z crc32c_intel 24576 0 2025-12-04T09:26:58.0417169Z efivarfs 24576 1 2025-12-04T09:26:58.0417357Z + modinfo nvidia 2025-12-04T09:26:58.0422564Z filename: /lib/modules/6.1.150-174.273.amzn2023.x86_64/kernel/drivers/video/nvidia.ko 2025-12-04T09:26:58.0423126Z import_ns: DMA_BUF 2025-12-04T09:26:58.0423443Z alias: char-major-195-* 2025-12-04T09:26:58.0423669Z version: 580.82.07 2025-12-04T09:26:58.0423847Z supported: external 2025-12-04T09:26:58.0424031Z license: Dual MIT/GPL 2025-12-04T09:26:58.0424246Z firmware: nvidia/580.82.07/gsp_tu10x.bin 2025-12-04T09:26:58.0424494Z firmware: nvidia/580.82.07/gsp_ga10x.bin 2025-12-04T09:26:58.0424731Z srcversion: BA7240A71DCF7DC6FE88C1D 2025-12-04T09:26:58.0424975Z alias: of:N*T*Cnvidia,tegra264-displayC* 2025-12-04T09:26:58.0425238Z alias: of:N*T*Cnvidia,tegra264-display 2025-12-04T09:26:58.0425491Z alias: of:N*T*Cnvidia,tegra234-displayC* 2025-12-04T09:26:58.0425746Z alias: of:N*T*Cnvidia,tegra234-display 2025-12-04T09:26:58.0425994Z alias: pci:v000010DEd*sv*sd*bc06sc80i00* 2025-12-04T09:26:58.0426248Z alias: pci:v000010DEd*sv*sd*bc03sc02i00* 2025-12-04T09:26:58.0426486Z alias: pci:v000010DEd*sv*sd*bc03sc00i00* 2025-12-04T09:26:58.0426717Z depends: i2c-core,drm 2025-12-04T09:26:58.0427200Z retpoline: Y 2025-12-04T09:26:58.0427460Z name: nvidia 2025-12-04T09:26:58.0427915Z vermagic: 6.1.150-174.273.amzn2023.x86_64 SMP preempt mod_unload modversions 2025-12-04T09:26:58.0428480Z parm: NvSwitchRegDwords:NvSwitch regkey (charp) 2025-12-04T09:26:58.0428966Z parm: NvSwitchBlacklist:NvSwitchBlacklist=uuid[,uuid...] (charp) 2025-12-04T09:26:58.0429427Z parm: NVreg_ResmanDebugLevel:int 2025-12-04T09:26:58.0429761Z parm: NVreg_RmLogonRC:int 2025-12-04T09:26:58.0430112Z parm: NVreg_ModifyDeviceFiles:int 2025-12-04T09:26:58.0430506Z parm: NVreg_DeviceFileUID:int 2025-12-04T09:26:58.0430754Z parm: NVreg_DeviceFileGID:int 2025-12-04T09:26:58.0430976Z parm: NVreg_DeviceFileMode:int 2025-12-04T09:26:58.0431237Z parm: NVreg_InitializeSystemMemoryAllocations:int 2025-12-04T09:26:58.0431524Z parm: NVreg_UsePageAttributeTable:int 2025-12-04T09:26:58.0431773Z parm: NVreg_EnablePCIeGen3:int 2025-12-04T09:26:58.0431985Z parm: NVreg_EnableMSI:int 2025-12-04T09:26:58.0432209Z parm: NVreg_EnableStreamMemOPs:int 2025-12-04T09:26:58.0432477Z parm: NVreg_RestrictProfilingToAdminUsers:int 2025-12-04T09:26:58.0432909Z parm: NVreg_PreserveVideoMemoryAllocations:int 2025-12-04T09:26:58.0433195Z parm: NVreg_EnableS0ixPowerManagement:int 2025-12-04T09:26:58.0433500Z parm: NVreg_S0ixPowerManagementVideoMemoryThreshold:int 2025-12-04T09:26:58.0433801Z parm: NVreg_DynamicPowerManagement:int 2025-12-04T09:26:58.0434106Z parm: NVreg_DynamicPowerManagementVideoMemoryThreshold:int 2025-12-04T09:26:58.0434409Z parm: NVreg_EnableGpuFirmware:int 2025-12-04T09:26:58.0434661Z parm: NVreg_EnableGpuFirmwareLogs:int 2025-12-04T09:26:58.0434928Z parm: NVreg_OpenRmEnableUnsupportedGpus:int 2025-12-04T09:26:58.0435207Z parm: NVreg_EnableUserNUMAManagement:int 2025-12-04T09:26:58.0435462Z parm: NVreg_MemoryPoolSize:int 2025-12-04T09:26:58.0435694Z parm: NVreg_KMallocHeapMaxSize:int 2025-12-04T09:26:58.0435950Z parm: NVreg_VMallocHeapMaxSize:int 2025-12-04T09:26:58.0436193Z parm: NVreg_IgnoreMMIOCheck:int 2025-12-04T09:26:58.0436549Z parm: NVreg_NvLinkDisable:int 2025-12-04T09:26:58.0436805Z parm: NVreg_EnablePCIERelaxedOrderingMode:int 2025-12-04T09:26:58.0437071Z parm: NVreg_RegisterPCIDriver:int 2025-12-04T09:26:58.0437339Z parm: NVreg_RegisterPlatformDeviceDriver:int 2025-12-04T09:26:58.0437597Z parm: NVreg_EnableResizableBar:int 2025-12-04T09:26:58.0437837Z parm: NVreg_EnableDbgBreakpoint:int 2025-12-04T09:26:58.0438084Z parm: NVreg_EnableNonblockingOpen:int 2025-12-04T09:26:58.0438332Z parm: NVreg_CoherentGPUMemoryMode:charp 2025-12-04T09:26:58.0438637Z parm: NVreg_RegistryDwords:charp 2025-12-04T09:26:58.0439052Z parm: NVreg_RegistryDwordsPerDevice:charp 2025-12-04T09:26:58.0439413Z parm: NVreg_RmMsg:charp 2025-12-04T09:26:58.0439728Z parm: NVreg_GpuBlacklist:charp 2025-12-04T09:26:58.0440078Z parm: NVreg_TemporaryFilePath:charp 2025-12-04T09:26:58.0440428Z parm: NVreg_ExcludedGpus:charp 2025-12-04T09:26:58.0440794Z parm: NVreg_DmaRemapPeerMmio:int 2025-12-04T09:26:58.0441208Z parm: NVreg_RmNvlinkBandwidth:charp 2025-12-04T09:26:58.0441673Z parm: NVreg_RmNvlinkBandwidthLinkCount:int 2025-12-04T09:26:58.0441947Z parm: NVreg_ImexChannelCount:int 2025-12-04T09:26:58.0442183Z parm: NVreg_CreateImexChannel0:int 2025-12-04T09:26:58.0442432Z parm: NVreg_GrdmaPciTopoCheckOverride:int 2025-12-04T09:26:58.0442673Z parm: rm_firmware_active:charp 2025-12-04T09:26:58.0442879Z + set +e 2025-12-04T09:26:58.0443020Z + nvidia-smi 2025-12-04T09:26:59.4946533Z Thu Dec 4 09:26:59 2025 2025-12-04T09:26:59.4947003Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:26:59.4947847Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-12-04T09:26:59.4948323Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:26:59.4948795Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T09:26:59.4949308Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-12-04T09:26:59.4949716Z | | | MIG M. | 2025-12-04T09:26:59.4950016Z |=========================================+========================+======================| 2025-12-04T09:26:59.5015088Z | 0 NVIDIA L4 Off | 00000000:35:00.0 Off | 0 | 2025-12-04T09:26:59.5015522Z | N/A 39C P0 29W / 72W | 0MiB / 23034MiB | 4% Default | 2025-12-04T09:26:59.5015896Z | | | N/A | 2025-12-04T09:26:59.5016254Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:26:59.5016537Z 2025-12-04T09:26:59.5016709Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:26:59.5017094Z | Processes: | 2025-12-04T09:26:59.5017494Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T09:26:59.5017861Z | ID ID Usage | 2025-12-04T09:26:59.5018164Z |=========================================================================================| 2025-12-04T09:26:59.5020322Z | No running processes found | 2025-12-04T09:26:59.5020809Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:26:59.8232285Z + nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 2025-12-04T09:27:01.2741739Z NVIDIA L4 2025-12-04T09:27:01.4571554Z + NVIDIA_SMI_STATUS=0 2025-12-04T09:27:01.4572857Z + '[' 0 -eq 0 ']' 2025-12-04T09:27:01.4573256Z + echo 'INFO: Ignoring allowed status 0' 2025-12-04T09:27:01.4573666Z + set -e 2025-12-04T09:27:01.4573874Z INFO: Ignoring allowed status 0 2025-12-04T09:27:01.4580651Z == Installing nvidia container toolkit for amzn2023 == 2025-12-04T09:27:01.4584020Z + sudo yum install -y yum-utils 2025-12-04T09:27:01.8893461Z Last metadata expiration check: 0:26:48 ago on Thu Dec 4 09:00:13 2025. 2025-12-04T09:27:01.9111641Z Package dnf-utils-4.3.0-13.amzn2023.0.5.noarch is already installed. 2025-12-04T09:27:01.9517667Z Dependencies resolved. 2025-12-04T09:27:01.9755668Z Nothing to do. 2025-12-04T09:27:01.9756137Z Complete! 2025-12-04T09:27:02.0451579Z + [[ amzn2023 == \a\m\z\n\2\0\2\3 ]] 2025-12-04T09:27:02.0452557Z + YUM_REPO_URL=https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-12-04T09:27:02.0453674Z + sudo yum-config-manager --add-repo https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-12-04T09:27:02.3542466Z Adding repo from: https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-12-04T09:27:02.3964538Z + sudo yum install -y nvidia-container-toolkit-1.17.8 libnvidia-container-tools-1.17.8 libnvidia-container1-1.17.8 nvidia-container-toolkit-base-1.17.8 2025-12-04T09:27:02.9433572Z nvidia-container-toolkit 24 kB/s | 833 B 00:00 2025-12-04T09:27:03.0096427Z Dependencies resolved. 2025-12-04T09:27:03.0325769Z ================================================================================ 2025-12-04T09:27:03.0326339Z Package Arch Version Repository Size 2025-12-04T09:27:03.0327138Z ================================================================================ 2025-12-04T09:27:03.0327472Z Downgrading: 2025-12-04T09:27:03.0327959Z libnvidia-container-tools x86_64 1.17.8-1 nvidia-container-toolkit 40 k 2025-12-04T09:27:03.0328755Z libnvidia-container1 x86_64 1.17.8-1 nvidia-container-toolkit 1.0 M 2025-12-04T09:27:03.0329491Z nvidia-container-toolkit x86_64 1.17.8-1 nvidia-container-toolkit 1.2 M 2025-12-04T09:27:03.0330303Z nvidia-container-toolkit-base x86_64 1.17.8-1 nvidia-container-toolkit 5.8 M 2025-12-04T09:27:03.0330786Z 2025-12-04T09:27:03.0330902Z Transaction Summary 2025-12-04T09:27:03.0331223Z ================================================================================ 2025-12-04T09:27:03.0331634Z Downgrade 4 Packages 2025-12-04T09:27:03.0331847Z 2025-12-04T09:27:03.0331986Z Total download size: 8.0 M 2025-12-04T09:27:03.0332501Z Downloading Packages: 2025-12-04T09:27:03.1429893Z (1/4): libnvidia-container-tools-1.17.8-1.x86_6 373 kB/s | 40 kB 00:00 2025-12-04T09:27:03.1758270Z (2/4): libnvidia-container1-1.17.8-1.x86_64.rpm 6.9 MB/s | 1.0 MB 00:00 2025-12-04T09:27:03.2295388Z (3/4): nvidia-container-toolkit-1.17.8-1.x86_64 6.4 MB/s | 1.2 MB 00:00 2025-12-04T09:27:03.3637423Z (4/4): nvidia-container-toolkit-base-1.17.8-1.x 26 MB/s | 5.8 MB 00:00 2025-12-04T09:27:03.3646803Z -------------------------------------------------------------------------------- 2025-12-04T09:27:03.3649622Z Total 24 MB/s | 8.0 MB 00:00 2025-12-04T09:27:03.3652423Z Running transaction check 2025-12-04T09:27:03.3769796Z Transaction check succeeded. 2025-12-04T09:27:03.3770253Z Running transaction test 2025-12-04T09:27:03.4201761Z Transaction test succeeded. 2025-12-04T09:27:03.4204620Z Running transaction 2025-12-04T09:27:04.0217309Z Preparing : 1/1 2025-12-04T09:27:04.1764440Z Downgrading : nvidia-container-toolkit-base-1.17.8-1.x86_64 1/8 2025-12-04T09:27:04.2037161Z Downgrading : libnvidia-container1-1.17.8-1.x86_64 2/8 2025-12-04T09:27:04.2750486Z Running scriptlet: libnvidia-container1-1.17.8-1.x86_64 2/8 2025-12-04T09:27:04.3897907Z Downgrading : libnvidia-container-tools-1.17.8-1.x86_64 3/8 2025-12-04T09:27:04.4140967Z Downgrading : nvidia-container-toolkit-1.17.8-1.x86_64 4/8 2025-12-04T09:27:04.4625503Z Running scriptlet: nvidia-container-toolkit-1.17.8-1.x86_64 4/8 2025-12-04T09:27:04.4688091Z Running scriptlet: nvidia-container-toolkit-1.18.1-1.x86_64 5/8 2025-12-04T09:27:04.4689432Z Cleanup : nvidia-container-toolkit-1.18.1-1.x86_64 5/8 2025-12-04T09:27:04.5066238Z Running scriptlet: nvidia-container-toolkit-1.18.1-1.x86_64 5/8 2025-12-04T09:27:04.5120087Z Running scriptlet: libnvidia-container-tools-1.18.1-1.x86_64 6/8 2025-12-04T09:27:04.5121274Z Cleanup : libnvidia-container-tools-1.18.1-1.x86_64 6/8 2025-12-04T09:27:04.5334948Z Running scriptlet: libnvidia-container-tools-1.18.1-1.x86_64 6/8 2025-12-04T09:27:04.5395983Z Running scriptlet: libnvidia-container1-1.18.1-1.x86_64 7/8 2025-12-04T09:27:04.5396958Z Cleanup : libnvidia-container1-1.18.1-1.x86_64 7/8 2025-12-04T09:27:04.5775266Z Running scriptlet: libnvidia-container1-1.18.1-1.x86_64 7/8 2025-12-04T09:27:04.5833774Z Running scriptlet: nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8 2025-12-04T09:27:04.5834920Z Cleanup : nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8 2025-12-04T09:27:04.6287154Z Running scriptlet: nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8 2025-12-04T09:27:04.6738631Z Running scriptlet: nvidia-container-toolkit-1.17.8-1.x86_64 8/8 2025-12-04T09:27:50.3227632Z Running scriptlet: nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8 2025-12-04T09:27:50.3230040Z Verifying : libnvidia-container-tools-1.17.8-1.x86_64 1/8 2025-12-04T09:27:50.3230874Z Verifying : libnvidia-container-tools-1.18.1-1.x86_64 2/8 2025-12-04T09:27:50.3231388Z Verifying : libnvidia-container1-1.17.8-1.x86_64 3/8 2025-12-04T09:27:50.3231787Z Verifying : libnvidia-container1-1.18.1-1.x86_64 4/8 2025-12-04T09:27:50.3232168Z Verifying : nvidia-container-toolkit-1.17.8-1.x86_64 5/8 2025-12-04T09:27:50.3232554Z Verifying : nvidia-container-toolkit-1.18.1-1.x86_64 6/8 2025-12-04T09:27:50.3232945Z Verifying : nvidia-container-toolkit-base-1.17.8-1.x86_64 7/8 2025-12-04T09:27:50.4623925Z Verifying : nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8================================================================================ 2025-12-04T09:27:50.4624600Z WARNING: 2025-12-04T09:27:50.4624842Z A newer release of "Amazon Linux" is available. 2025-12-04T09:27:50.4625072Z 2025-12-04T09:27:50.4625159Z Available Versions: 2025-12-04T09:27:50.4625309Z 2025-12-04T09:27:50.4625404Z Version 2023.9.20250929: 2025-12-04T09:27:50.4625701Z Run the following command to upgrade to 2023.9.20250929: 2025-12-04T09:27:50.4625940Z 2025-12-04T09:27:50.4626058Z dnf upgrade --releasever=2023.9.20250929 2025-12-04T09:27:50.4626266Z 2025-12-04T09:27:50.4626346Z Release notes: 2025-12-04T09:27:50.4626745Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20250929.html 2025-12-04T09:27:50.4627106Z 2025-12-04T09:27:50.4627195Z Version 2023.9.20251014: 2025-12-04T09:27:50.4627476Z Run the following command to upgrade to 2023.9.20251014: 2025-12-04T09:27:50.4627717Z 2025-12-04T09:27:50.4627821Z dnf upgrade --releasever=2023.9.20251014 2025-12-04T09:27:50.4628019Z 2025-12-04T09:27:50.4628102Z Release notes: 2025-12-04T09:27:50.4628455Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251014.html 2025-12-04T09:27:50.4628801Z 2025-12-04T09:27:50.4628879Z Version 2023.9.20251020: 2025-12-04T09:27:50.4629482Z Run the following command to upgrade to 2023.9.20251020: 2025-12-04T09:27:50.4629730Z 2025-12-04T09:27:50.4629840Z dnf upgrade --releasever=2023.9.20251020 2025-12-04T09:27:50.4630027Z 2025-12-04T09:27:50.4630102Z Release notes: 2025-12-04T09:27:50.4630467Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251020.html 2025-12-04T09:27:50.4630803Z 2025-12-04T09:27:50.4630903Z Version 2023.9.20251027: 2025-12-04T09:27:50.4631182Z Run the following command to upgrade to 2023.9.20251027: 2025-12-04T09:27:50.4631401Z 2025-12-04T09:27:50.4631479Z dnf upgrade --releasever=2023.9.20251027 2025-12-04T09:27:50.4631631Z 2025-12-04T09:27:50.4631692Z Release notes: 2025-12-04T09:27:50.4631984Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251027.html 2025-12-04T09:27:50.4632251Z 2025-12-04T09:27:50.4632311Z Version 2023.9.20251105: 2025-12-04T09:27:50.4632530Z Run the following command to upgrade to 2023.9.20251105: 2025-12-04T09:27:50.4632718Z 2025-12-04T09:27:50.4632805Z dnf upgrade --releasever=2023.9.20251105 2025-12-04T09:27:50.4632955Z 2025-12-04T09:27:50.4633019Z Release notes: 2025-12-04T09:27:50.4633300Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251105.html 2025-12-04T09:27:50.4633571Z 2025-12-04T09:27:50.4633633Z Version 2023.9.20251110: 2025-12-04T09:27:50.4633849Z Run the following command to upgrade to 2023.9.20251110: 2025-12-04T09:27:50.4634040Z 2025-12-04T09:27:50.4634121Z dnf upgrade --releasever=2023.9.20251110 2025-12-04T09:27:50.4634274Z 2025-12-04T09:27:50.4634333Z Release notes: 2025-12-04T09:27:50.4634615Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251110.html 2025-12-04T09:27:50.4635049Z 2025-12-04T09:27:50.4635118Z Version 2023.9.20251117: 2025-12-04T09:27:50.4635328Z Run the following command to upgrade to 2023.9.20251117: 2025-12-04T09:27:50.4635513Z 2025-12-04T09:27:50.4635596Z dnf upgrade --releasever=2023.9.20251117 2025-12-04T09:27:50.4635750Z 2025-12-04T09:27:50.4635810Z Release notes: 2025-12-04T09:27:50.4636091Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251117.html 2025-12-04T09:27:50.4636364Z 2025-12-04T09:27:50.4636447Z ================================================================================ 2025-12-04T09:27:50.5088000Z 2025-12-04T09:27:50.5088184Z 2025-12-04T09:27:50.5088310Z Downgraded: 2025-12-04T09:27:50.5088671Z libnvidia-container-tools-1.17.8-1.x86_64 2025-12-04T09:27:50.5089186Z libnvidia-container1-1.17.8-1.x86_64 2025-12-04T09:27:50.5089699Z nvidia-container-toolkit-1.17.8-1.x86_64 2025-12-04T09:27:50.5090227Z nvidia-container-toolkit-base-1.17.8-1.x86_64 2025-12-04T09:27:50.5090539Z 2025-12-04T09:27:50.5090616Z Complete! 2025-12-04T09:27:50.5555120Z + sudo systemctl restart docker 2025-12-04T09:28:02.6179588Z Thu Dec 4 09:28:02 2025 2025-12-04T09:28:02.6180374Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:28:02.6181322Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-12-04T09:28:02.6182217Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:28:02.6183127Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T09:28:02.6184095Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-12-04T09:28:02.6184843Z | | | MIG M. | 2025-12-04T09:28:02.6185433Z |=========================================+========================+======================| 2025-12-04T09:28:02.6254830Z | 0 NVIDIA L4 On | 00000000:35:00.0 Off | 0 | 2025-12-04T09:28:02.6255607Z | N/A 39C P0 29W / 72W | 0MiB / 23034MiB | 4% Default | 2025-12-04T09:28:02.6256011Z | | | N/A | 2025-12-04T09:28:02.6256380Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:28:02.6256651Z 2025-12-04T09:28:02.6256815Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:28:02.6257207Z | Processes: | 2025-12-04T09:28:02.6257650Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T09:28:02.6258053Z | ID ID Usage | 2025-12-04T09:28:02.6258394Z |=========================================================================================| 2025-12-04T09:28:02.6259977Z | No running processes found | 2025-12-04T09:28:02.6260499Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:28:02.7862799Z Unable to find image 'public.ecr.aws/docker/library/python:3.13' locally 2025-12-04T09:28:03.0199983Z 3.13: Pulling from docker/library/python 2025-12-04T09:28:03.1064356Z 53c88f1dfeb7: Pulling fs layer 2025-12-04T09:28:03.1064700Z eae668646f44: Pulling fs layer 2025-12-04T09:28:03.1064976Z ff2e6e687b6c: Pulling fs layer 2025-12-04T09:28:03.1065261Z 7c40a3faff76: Pulling fs layer 2025-12-04T09:28:03.1065512Z 967a3b1c8fef: Pulling fs layer 2025-12-04T09:28:03.1066207Z a64e1a44f22a: Pulling fs layer 2025-12-04T09:28:03.1066454Z 52655f8a5bcc: Pulling fs layer 2025-12-04T09:28:03.1066685Z 7c40a3faff76: Waiting 2025-12-04T09:28:03.1066903Z 967a3b1c8fef: Waiting 2025-12-04T09:28:03.1067109Z 52655f8a5bcc: Waiting 2025-12-04T09:28:03.1067315Z a64e1a44f22a: Waiting 2025-12-04T09:28:03.2202082Z eae668646f44: Verifying Checksum 2025-12-04T09:28:03.2202386Z eae668646f44: Download complete 2025-12-04T09:28:03.3019889Z ff2e6e687b6c: Verifying Checksum 2025-12-04T09:28:03.3020229Z ff2e6e687b6c: Download complete 2025-12-04T09:28:03.3390971Z 53c88f1dfeb7: Verifying Checksum 2025-12-04T09:28:03.3391263Z 53c88f1dfeb7: Download complete 2025-12-04T09:28:03.3773475Z 967a3b1c8fef: Verifying Checksum 2025-12-04T09:28:03.3774025Z 967a3b1c8fef: Download complete 2025-12-04T09:28:03.4135531Z 52655f8a5bcc: Verifying Checksum 2025-12-04T09:28:03.4135830Z 52655f8a5bcc: Download complete 2025-12-04T09:28:03.4563309Z a64e1a44f22a: Verifying Checksum 2025-12-04T09:28:03.8582930Z a64e1a44f22a: Download complete 2025-12-04T09:28:03.8583379Z 7c40a3faff76: Verifying Checksum 2025-12-04T09:28:03.8598384Z 7c40a3faff76: Download complete 2025-12-04T09:28:04.6522704Z 53c88f1dfeb7: Pull complete 2025-12-04T09:28:05.1874591Z eae668646f44: Pull complete 2025-12-04T09:28:07.0149471Z ff2e6e687b6c: Pull complete 2025-12-04T09:28:12.2263864Z 7c40a3faff76: Pull complete 2025-12-04T09:28:12.6141501Z 967a3b1c8fef: Pull complete 2025-12-04T09:28:13.2447806Z a64e1a44f22a: Pull complete 2025-12-04T09:28:13.2681352Z 52655f8a5bcc: Pull complete 2025-12-04T09:28:13.2815889Z Digest: sha256:3f986299a7b8b44b0d8cf9bda2b22361ce5c3058ef5d7cb17fb7452506680ab0 2025-12-04T09:28:13.2858721Z Status: Downloaded newer image for public.ecr.aws/docker/library/python:3.13 2025-12-04T09:28:20.4404424Z Thu Dec 4 09:28:20 2025 2025-12-04T09:28:20.4405054Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:28:20.4405828Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-12-04T09:28:20.4406573Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:28:20.4407363Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T09:28:20.4408470Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-12-04T09:28:20.4408855Z | | | MIG M. | 2025-12-04T09:28:20.4409115Z |=========================================+========================+======================| 2025-12-04T09:28:20.4520332Z | 0 NVIDIA L4 On | 00000000:35:00.0 Off | 0 | 2025-12-04T09:28:20.4520816Z | N/A 38C P8 12W / 72W | 0MiB / 23034MiB | 0% Default | 2025-12-04T09:28:20.4521191Z | | | N/A | 2025-12-04T09:28:20.4521610Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:28:20.4523645Z 2025-12-04T09:28:20.4523856Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:28:20.4524281Z | Processes: | 2025-12-04T09:28:20.4524682Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T09:28:20.4525079Z | ID ID Usage | 2025-12-04T09:28:20.4525387Z |=========================================================================================| 2025-12-04T09:28:20.4528789Z | No running processes found | 2025-12-04T09:28:20.4529157Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:28:21.5093417Z Command completed after 1 attempt(s). 2025-12-04T09:28:21.5180991Z Prepare all required actions 2025-12-04T09:28:21.5205601Z ##[group]Run ./.github/actions/get-workflow-job-id 2025-12-04T09:28:21.5205848Z with: 2025-12-04T09:28:21.5206399Z github-token: *** 2025-12-04T09:28:21.5206577Z env: 2025-12-04T09:28:21.5206724Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:28:21.5206919Z HAS_NVIDIA_GPU: true 2025-12-04T09:28:21.5207148Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:28:21.5207405Z ##[endgroup] 2025-12-04T09:28:21.5221106Z ##[group]Run set -eux 2025-12-04T09:28:21.5221293Z set -eux 2025-12-04T09:28:21.5221613Z python3 .github/scripts/get_workflow_job_id.py "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-12-04T09:28:21.5234708Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:28:21.5235005Z env: 2025-12-04T09:28:21.5235170Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:28:21.5235360Z HAS_NVIDIA_GPU: true 2025-12-04T09:28:21.5235634Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:28:21.5236038Z GITHUB_TOKEN: *** 2025-12-04T09:28:21.5236204Z ##[endgroup] 2025-12-04T09:28:21.5272025Z + python3 .github/scripts/get_workflow_job_id.py 19922031852 i-0a8d72b6d567bdb22 2025-12-04T09:28:22.4246153Z Setting output job-id=57118543727 2025-12-04T09:28:22.4246777Z Setting output job-name=ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu, unstable) 2025-12-04T09:28:22.4359272Z ##[group]Run python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-12-04T09:28:22.4359891Z python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-12-04T09:28:22.4360586Z python3 -m tools.stats.monitor --log-interval "$MONITOR_LOG_INTERVAL" --data-collect-interval "$MONITOR_DATA_COLLECT_INTERVAL" > usage_log.txt 2>&1 & 2025-12-04T09:28:22.4361199Z echo "monitor-script-pid=${!}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:28:22.4370095Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:28:22.4370380Z env: 2025-12-04T09:28:22.4370537Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:28:22.4370724Z HAS_NVIDIA_GPU: true 2025-12-04T09:28:22.4371555Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:28:22.4371880Z JOB_ID: 57118543727 2025-12-04T09:28:22.4372285Z JOB_NAME: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu, unstable) 2025-12-04T09:28:22.4372708Z WORKFLOW_NAME: vllm-test 2025-12-04T09:28:22.4372902Z WORKFLOW_RUN_ID: 19922031852 2025-12-04T09:28:22.4373100Z MONITOR_LOG_INTERVAL: 5 2025-12-04T09:28:22.4373298Z MONITOR_DATA_COLLECT_INTERVAL: 1 2025-12-04T09:28:22.4373503Z ##[endgroup] 2025-12-04T09:28:22.7168280Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T09:28:23.0529341Z Collecting psutil==5.9.8 2025-12-04T09:28:23.0684357Z Downloading psutil-5.9.8-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (288 kB) 2025-12-04T09:28:23.1338551Z Collecting dataclasses_json==0.6.7 2025-12-04T09:28:23.1372803Z Downloading dataclasses_json-0.6.7-py3-none-any.whl (28 kB) 2025-12-04T09:28:23.1645004Z Collecting nvidia-ml-py==11.525.84 2025-12-04T09:28:23.1676206Z Downloading nvidia_ml_py-11.525.84-py3-none-any.whl (34 kB) 2025-12-04T09:28:23.2748249Z Collecting marshmallow<4.0.0,>=3.18.0 2025-12-04T09:28:23.2782254Z Downloading marshmallow-3.26.1-py3-none-any.whl (50 kB) 2025-12-04T09:28:23.3012730Z Collecting typing-inspect<1,>=0.4.0 2025-12-04T09:28:23.3044726Z Downloading typing_inspect-0.9.0-py3-none-any.whl (8.8 kB) 2025-12-04T09:28:23.3545867Z Collecting packaging>=17.0 2025-12-04T09:28:23.3574283Z Downloading packaging-25.0-py3-none-any.whl (66 kB) 2025-12-04T09:28:23.3811129Z Collecting mypy-extensions>=0.3.0 2025-12-04T09:28:23.3843632Z Downloading mypy_extensions-1.1.0-py3-none-any.whl (5.0 kB) 2025-12-04T09:28:23.4296989Z Collecting typing-extensions>=3.7.4 2025-12-04T09:28:23.4332010Z Downloading typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2025-12-04T09:28:23.5219375Z Installing collected packages: typing-extensions, packaging, mypy-extensions, typing-inspect, marshmallow, psutil, nvidia-ml-py, dataclasses-json 2025-12-04T09:28:23.7731704Z Successfully installed dataclasses-json-0.6.7 marshmallow-3.26.1 mypy-extensions-1.1.0 nvidia-ml-py-11.525.84 packaging-25.0 psutil-5.9.8 typing-extensions-4.15.0 typing-inspect-0.9.0 2025-12-04T09:28:23.9304690Z Prepare all required actions 2025-12-04T09:28:23.9305045Z Getting action download info 2025-12-04T09:28:24.0875703Z Download action repository 'seemethere/download-artifact-s3@v4' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-12-04T09:28:24.4245347Z Download action repository 'actions/download-artifact@v4' (SHA:d3f86a106a0bac45b974a628896c90dbdf5c8093) 2025-12-04T09:28:24.8004703Z ##[group]Run ./.github/actions/download-build-artifacts 2025-12-04T09:28:24.8005129Z with: 2025-12-04T09:28:24.8005322Z name: linux-jammy-cuda12.8-py3.12-gcc11 2025-12-04T09:28:24.8005585Z s3-bucket: gha-artifacts 2025-12-04T09:28:24.8005764Z env: 2025-12-04T09:28:24.8005916Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:28:24.8006102Z HAS_NVIDIA_GPU: true 2025-12-04T09:28:24.8006321Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:28:24.8006582Z ##[endgroup] 2025-12-04T09:28:24.8032141Z ##[group]Run seemethere/download-artifact-s3@v4 2025-12-04T09:28:24.8032394Z with: 2025-12-04T09:28:24.8032607Z name: linux-jammy-cuda12.8-py3.12-gcc11 2025-12-04T09:28:24.8032842Z s3-bucket: gha-artifacts 2025-12-04T09:28:24.8033037Z region: us-east-1 2025-12-04T09:28:24.8033201Z env: 2025-12-04T09:28:24.8033344Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:28:24.8033526Z HAS_NVIDIA_GPU: true 2025-12-04T09:28:24.8033750Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:28:24.8034014Z ##[endgroup] 2025-12-04T09:28:25.2186656Z (node:61342) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-12-04T09:28:25.2187160Z 2025-12-04T09:28:25.2187344Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-12-04T09:28:25.2187840Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-12-04T09:28:25.2188361Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-12-04T09:28:25.5184090Z Found 1 objects with prefix pytorch/pytorch/19922031852/linux-jammy-cuda12.8-py3.12-gcc11/ 2025-12-04T09:28:25.5184853Z Starting download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-12-04T09:28:46.8667428Z Finished download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-12-04T09:28:46.8673251Z Artifact download has finished successfully 2025-12-04T09:28:46.8948045Z ##[group]Run unzip -o artifacts.zip 2025-12-04T09:28:46.8948297Z unzip -o artifacts.zip 2025-12-04T09:28:46.8956204Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:28:46.8956515Z env: 2025-12-04T09:28:46.8956672Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:28:46.8956867Z HAS_NVIDIA_GPU: true 2025-12-04T09:28:46.8957093Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:28:46.8957350Z ##[endgroup] 2025-12-04T09:28:46.9503237Z Archive: artifacts.zip 2025-12-04T09:28:46.9654891Z creating: dist/ 2025-12-04T09:28:49.5092440Z inflating: dist/torch-2.10.0a0+gitffd9b0f-cp312-cp312-linux_x86_64.whl 2025-12-04T09:28:49.5094410Z creating: dist/vision/ 2025-12-04T09:28:49.5223826Z inflating: dist/vision/torchvision-0.25.0a0+617079d-cp312-cp312-linux_x86_64.whl 2025-12-04T09:28:49.5224304Z creating: dist/audio/ 2025-12-04T09:28:49.5296483Z inflating: dist/audio/torchaudio-2.10.0a0+e90a398-cp312-cp312-linux_x86_64.whl 2025-12-04T09:28:49.5412557Z inflating: dist/.ninja_log 2025-12-04T09:28:49.5412866Z creating: dist/external/ 2025-12-04T09:28:49.5413157Z creating: dist/external/vllm/ 2025-12-04T09:28:49.5413591Z creating: dist/external/vllm/wheels/ 2025-12-04T09:28:49.5414625Z inflating: dist/external/vllm/wheels/build_summary.txt 2025-12-04T09:28:49.5414998Z creating: dist/external/vllm/wheels/vllm/ 2025-12-04T09:28:52.1941374Z inflating: dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev389+ge51928192.d20251204-cp38-abi3-linux_x86_64.whl 2025-12-04T09:28:52.1942050Z creating: dist/external/vllm/wheels/xformers/ 2025-12-04T09:28:53.8069287Z inflating: dist/external/vllm/wheels/xformers/xformers-0.0.34+3f91ad6f.d20251204-cp39-abi3-linux_x86_64.whl 2025-12-04T09:28:53.8069966Z creating: build/custom_test_artifacts/ 2025-12-04T09:28:53.8070326Z creating: build/custom_test_artifacts/custom-op-build/ 2025-12-04T09:28:53.8070764Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/ 2025-12-04T09:28:53.8071596Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:28:53.8078102Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:28:53.8078737Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/ 2025-12-04T09:28:53.8079331Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:28:53.8079960Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:28:53.8080567Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:28:53.8082860Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:28:53.8084218Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:28:53.8085145Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:28:53.8085772Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:28:53.8086390Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:28:53.8088641Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:28:53.8090188Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:28:53.8091147Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:28:53.8092911Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:28:53.8094905Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:28:53.8095476Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-12-04T09:28:53.8095981Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-12-04T09:28:53.8147785Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-12-04T09:28:53.8200538Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-12-04T09:28:53.8201459Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-12-04T09:28:53.8257535Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-12-04T09:28:53.8258485Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-12-04T09:28:53.8259622Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-12-04T09:28:53.8260551Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-12-04T09:28:53.8261744Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-12-04T09:28:53.8262648Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-12-04T09:28:53.8263713Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-12-04T09:28:53.8264609Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-12-04T09:28:53.8265486Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-12-04T09:28:53.8266334Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-12-04T09:28:53.8267450Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-12-04T09:28:53.8268348Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-12-04T09:28:53.8269015Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-12-04T09:28:53.8270110Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-12-04T09:28:53.8272925Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-12-04T09:28:53.8337305Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-12-04T09:28:53.8338105Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-12-04T09:28:53.8403050Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-12-04T09:28:53.8403765Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:28:53.8404309Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeTmp/ 2025-12-04T09:28:53.8404894Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/cmake.check_cache 2025-12-04T09:28:53.8405478Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/ 2025-12-04T09:28:53.8406133Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.ts 2025-12-04T09:28:53.8406873Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.make 2025-12-04T09:28:53.8407577Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/depend.make 2025-12-04T09:28:53.8408241Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/link.txt 2025-12-04T09:28:53.8408773Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/cmake_clean.cmake 2025-12-04T09:28:53.8409890Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/build.make 2025-12-04T09:28:53.8410689Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/DependInfo.cmake 2025-12-04T09:28:53.8411489Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/flags.make 2025-12-04T09:28:53.8412443Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/progress.make 2025-12-04T09:28:53.8431270Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o.d 2025-12-04T09:28:53.8606747Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o 2025-12-04T09:28:53.8607415Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/ 2025-12-04T09:28:53.8608134Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.ts 2025-12-04T09:28:53.8608963Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.make 2025-12-04T09:28:53.8609550Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/depend.make 2025-12-04T09:28:53.8610101Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/link.txt 2025-12-04T09:28:53.8610908Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/cmake_clean.cmake 2025-12-04T09:28:53.8611889Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/build.make 2025-12-04T09:28:53.8612715Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/DependInfo.cmake 2025-12-04T09:28:53.8614088Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/flags.make 2025-12-04T09:28:53.8615118Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/progress.make 2025-12-04T09:28:53.8633935Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o.d 2025-12-04T09:28:53.8706145Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o 2025-12-04T09:28:53.8706966Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:28:53.8707661Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:28:53.8708407Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/progress.marks 2025-12-04T09:28:53.8709355Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile2 2025-12-04T09:28:53.8711037Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile.cmake 2025-12-04T09:28:53.8711592Z inflating: build/custom_test_artifacts/custom-op-build/detect_cuda_version.cc 2025-12-04T09:28:53.8714733Z inflating: build/custom_test_artifacts/custom-op-build/CMakeCache.txt 2025-12-04T09:28:53.8715687Z inflating: build/custom_test_artifacts/custom-op-build/Makefile 2025-12-04T09:28:53.8716530Z inflating: build/custom_test_artifacts/custom-op-build/cmake_install.cmake 2025-12-04T09:28:53.8867572Z inflating: build/custom_test_artifacts/custom-op-build/libcustom_ops.so 2025-12-04T09:28:53.8917167Z inflating: build/custom_test_artifacts/custom-op-build/test_custom_ops 2025-12-04T09:28:53.8917636Z creating: build/custom_test_artifacts/jit-hook-build/ 2025-12-04T09:28:53.8918068Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/ 2025-12-04T09:28:53.8931329Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:28:53.8942965Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:28:53.8990954Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/ 2025-12-04T09:28:53.8992683Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:28:53.8993854Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:28:53.8995023Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:28:53.9000068Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:28:53.9001487Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:28:53.9002422Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:28:53.9003318Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:28:53.9083730Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:28:53.9089352Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:28:53.9091035Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:28:53.9091947Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:28:53.9094150Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:28:53.9095436Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:28:53.9096101Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-12-04T09:28:53.9096747Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-12-04T09:28:53.9148567Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-12-04T09:28:53.9201378Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-12-04T09:28:53.9202532Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-12-04T09:28:53.9258789Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-12-04T09:28:53.9259935Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-12-04T09:28:53.9261026Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-12-04T09:28:53.9262057Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-12-04T09:28:53.9263070Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-12-04T09:28:53.9264144Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-12-04T09:28:53.9265147Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-12-04T09:28:53.9266064Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-12-04T09:28:53.9267111Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-12-04T09:28:53.9268043Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-12-04T09:28:53.9268808Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-12-04T09:28:53.9269572Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-12-04T09:28:53.9270242Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-12-04T09:28:53.9271519Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-12-04T09:28:53.9273892Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-12-04T09:28:53.9338366Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-12-04T09:28:53.9339502Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-12-04T09:28:53.9404388Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-12-04T09:28:53.9405185Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:28:53.9406201Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeTmp/ 2025-12-04T09:28:53.9406851Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/cmake.check_cache 2025-12-04T09:28:53.9407486Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/ 2025-12-04T09:28:53.9408504Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.ts 2025-12-04T09:28:53.9409224Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.make 2025-12-04T09:28:53.9409960Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/depend.make 2025-12-04T09:28:53.9410560Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/link.txt 2025-12-04T09:28:53.9411199Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/cmake_clean.cmake 2025-12-04T09:28:53.9411899Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/build.make 2025-12-04T09:28:53.9412550Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/DependInfo.cmake 2025-12-04T09:28:53.9413182Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/flags.make 2025-12-04T09:28:53.9414059Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/progress.make 2025-12-04T09:28:53.9432891Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o.d 2025-12-04T09:28:53.9488650Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o 2025-12-04T09:28:53.9489678Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:28:53.9490429Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:28:53.9491184Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/progress.marks 2025-12-04T09:28:53.9491920Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile2 2025-12-04T09:28:53.9493443Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile.cmake 2025-12-04T09:28:53.9494099Z inflating: build/custom_test_artifacts/jit-hook-build/detect_cuda_version.cc 2025-12-04T09:28:53.9496807Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeCache.txt 2025-12-04T09:28:53.9497581Z inflating: build/custom_test_artifacts/jit-hook-build/Makefile 2025-12-04T09:28:53.9498466Z inflating: build/custom_test_artifacts/jit-hook-build/cmake_install.cmake 2025-12-04T09:28:53.9532737Z inflating: build/custom_test_artifacts/jit-hook-build/test_jit_hooks 2025-12-04T09:28:53.9533299Z creating: build/custom_test_artifacts/custom-backend-build/ 2025-12-04T09:28:53.9533971Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/ 2025-12-04T09:28:53.9534686Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:28:53.9541256Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:28:53.9542023Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/ 2025-12-04T09:28:53.9542761Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:28:53.9543548Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:28:53.9544355Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:28:53.9545734Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:28:53.9547221Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:28:53.9548317Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:28:53.9548994Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:28:53.9549589Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:28:53.9551907Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:28:53.9553084Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:28:53.9554215Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:28:53.9556093Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:28:53.9557835Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:28:53.9558496Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-12-04T09:28:53.9559222Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-12-04T09:28:53.9611325Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-12-04T09:28:53.9664042Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-12-04T09:28:53.9665156Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-12-04T09:28:53.9721069Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-12-04T09:28:53.9722294Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-12-04T09:28:53.9723418Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-12-04T09:28:53.9724514Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-12-04T09:28:53.9725713Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-12-04T09:28:53.9726724Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-12-04T09:28:53.9727776Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-12-04T09:28:53.9728853Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-12-04T09:28:53.9729680Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-12-04T09:28:53.9730513Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-12-04T09:28:53.9731254Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-12-04T09:28:53.9731995Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-12-04T09:28:53.9732824Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-12-04T09:28:53.9733593Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-12-04T09:28:53.9736116Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-12-04T09:28:53.9800912Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-12-04T09:28:53.9801805Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-12-04T09:28:53.9866525Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-12-04T09:28:53.9867488Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:28:53.9868168Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeTmp/ 2025-12-04T09:28:53.9868882Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/cmake.check_cache 2025-12-04T09:28:53.9869698Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/ 2025-12-04T09:28:53.9870562Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.ts 2025-12-04T09:28:53.9871649Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.make 2025-12-04T09:28:53.9872641Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/depend.make 2025-12-04T09:28:53.9873478Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/link.txt 2025-12-04T09:28:53.9874317Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/cmake_clean.cmake 2025-12-04T09:28:53.9875285Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/build.make 2025-12-04T09:28:53.9876125Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/DependInfo.cmake 2025-12-04T09:28:53.9877010Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/flags.make 2025-12-04T09:28:53.9877897Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/progress.make 2025-12-04T09:28:53.9880286Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o.d 2025-12-04T09:28:53.9985484Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o 2025-12-04T09:28:53.9986424Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/ 2025-12-04T09:28:53.9987355Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.ts 2025-12-04T09:28:53.9988408Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.make 2025-12-04T09:28:53.9989328Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/depend.make 2025-12-04T09:28:53.9990162Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/link.txt 2025-12-04T09:28:53.9991139Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/cmake_clean.cmake 2025-12-04T09:28:53.9992046Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/build.make 2025-12-04T09:28:53.9992928Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/DependInfo.cmake 2025-12-04T09:28:53.9993886Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/flags.make 2025-12-04T09:28:53.9994777Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/progress.make 2025-12-04T09:28:54.0012076Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o.d 2025-12-04T09:28:54.0060940Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o 2025-12-04T09:28:54.0062165Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:28:54.0062947Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:28:54.0063994Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/progress.marks 2025-12-04T09:28:54.0064732Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile2 2025-12-04T09:28:54.0065974Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile.cmake 2025-12-04T09:28:54.0066653Z inflating: build/custom_test_artifacts/custom-backend-build/detect_cuda_version.cc 2025-12-04T09:28:54.0069111Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeCache.txt 2025-12-04T09:28:54.0070141Z inflating: build/custom_test_artifacts/custom-backend-build/Makefile 2025-12-04T09:28:54.0071134Z inflating: build/custom_test_artifacts/custom-backend-build/cmake_install.cmake 2025-12-04T09:28:54.0160339Z inflating: build/custom_test_artifacts/custom-backend-build/libcustom_backend.so 2025-12-04T09:28:54.0195469Z inflating: build/custom_test_artifacts/custom-backend-build/test_custom_backend 2025-12-04T09:28:54.0196091Z creating: build/lib/ 2025-12-04T09:28:54.0267697Z inflating: build/lib/libprotobuf-lite.a 2025-12-04T09:28:54.0652720Z inflating: build/lib/libprotobuf.a 2025-12-04T09:28:54.0661134Z inflating: build/lib/libpthreadpool.a 2025-12-04T09:28:54.0668491Z inflating: build/lib/libcpuinfo.a 2025-12-04T09:28:54.0675733Z inflating: build/lib/libcpuinfo_internals.a 2025-12-04T09:28:54.0676720Z inflating: build/lib/libclog.a 2025-12-04T09:28:54.1108500Z inflating: build/lib/libprotoc.a 2025-12-04T09:28:54.1110902Z inflating: build/lib/libnnpack_reference_layers.a 2025-12-04T09:28:54.1127581Z inflating: build/lib/libpytorch_qnnpack.a 2025-12-04T09:28:54.1289342Z inflating: build/lib/libmicrokernels-prod.a 2025-12-04T09:28:54.1305036Z inflating: build/lib/libnnpack.a 2025-12-04T09:28:54.2063131Z inflating: build/lib/libmicrokernels-all.a 2025-12-04T09:28:54.2124079Z inflating: build/lib/libgtest.a 2025-12-04T09:28:54.2139289Z inflating: build/lib/libgmock.a 2025-12-04T09:28:54.2140117Z inflating: build/lib/libgmock_main.a 2025-12-04T09:28:54.2141155Z inflating: build/lib/libgtest_main.a 2025-12-04T09:28:54.2207313Z inflating: build/lib/libbenchmark.a 2025-12-04T09:28:54.2208219Z inflating: build/lib/libbenchmark_main.a 2025-12-04T09:28:54.2287035Z inflating: build/lib/libXNNPACK.a 2025-12-04T09:28:54.2294242Z inflating: build/lib/libittnotify.a 2025-12-04T09:28:54.2295109Z inflating: build/lib/libjitprofiling.a 2025-12-04T09:28:54.2352301Z inflating: build/lib/libasmjit.a 2025-12-04T09:28:54.3378067Z inflating: build/lib/libfbgemm.a 2025-12-04T09:28:54.3404836Z inflating: build/lib/libtensorpipe_uv.a 2025-12-04T09:28:54.3879730Z inflating: build/lib/libtensorpipe.a 2025-12-04T09:28:54.4092737Z inflating: build/lib/libtensorpipe_cuda.a 2025-12-04T09:28:54.4209748Z inflating: build/lib/libgloo.a 2025-12-04T09:28:54.4250914Z inflating: build/lib/libonnx_proto.a 2025-12-04T09:28:54.4873041Z inflating: build/lib/libonnx.a 2025-12-04T09:28:54.5256417Z inflating: build/lib/libgloo_cuda.a 2025-12-04T09:28:54.5273777Z inflating: build/lib/libfmt.a 2025-12-04T09:28:55.4123603Z inflating: build/lib/libdnnl.a 2025-12-04T09:28:55.4538799Z inflating: build/lib/libkineto.a 2025-12-04T09:28:55.4540217Z inflating: build/lib/libtorch_global_deps.so 2025-12-04T09:28:55.4640913Z inflating: build/lib/libc10.so 2025-12-04T09:28:55.4642844Z inflating: build/lib/libcaffe2_nvrtc.so 2025-12-04T09:28:55.4685791Z inflating: build/lib/libc10_cuda.so 2025-12-04T09:28:58.1473053Z inflating: build/lib/libtorch_cpu.so 2025-12-04T09:28:58.2160187Z inflating: build/lib/libtorch_nvshmem.so 2025-12-04T09:29:02.2758479Z inflating: build/lib/libtorch_cuda.so 2025-12-04T09:29:02.2762469Z inflating: build/lib/libshm.so 2025-12-04T09:29:02.2763984Z inflating: build/lib/libtorch.so 2025-12-04T09:29:02.2766571Z inflating: build/lib/libc10d_cuda_test.so 2025-12-04T09:29:02.2811680Z inflating: build/lib/libtorch_cuda_linalg.so 2025-12-04T09:29:02.2835074Z inflating: build/lib/libaoti_custom_ops.so 2025-12-04T09:29:02.2851464Z inflating: build/lib/libjitbackend_test.so 2025-12-04T09:29:02.2913736Z inflating: build/lib/libtorchbind_test.so 2025-12-04T09:29:02.2934791Z inflating: build/lib/libbackend_with_compiler.so 2025-12-04T09:29:02.5014749Z inflating: build/lib/libtorch_python.so 2025-12-04T09:29:02.5046195Z inflating: build/lib/libnnapi_backend.so 2025-12-04T09:29:02.5046506Z creating: build/bin/ 2025-12-04T09:29:02.5442280Z inflating: build/bin/protoc-3.13.0.0 2025-12-04T09:29:02.5837857Z inflating: build/bin/protoc 2025-12-04T09:29:02.5889946Z inflating: build/bin/c10_AllocatorConfig_test 2025-12-04T09:29:02.5940387Z inflating: build/bin/c10_DeviceGuard_test 2025-12-04T09:29:02.5989123Z inflating: build/bin/c10_CompileTimeFunctionPointer_test 2025-12-04T09:29:02.6038981Z inflating: build/bin/c10_Device_test 2025-12-04T09:29:02.6096689Z inflating: build/bin/c10_DispatchKeySet_test 2025-12-04T09:29:02.6144307Z inflating: build/bin/c10_StreamGuard_test 2025-12-04T09:29:02.6196373Z inflating: build/bin/c10_Scalar_test 2025-12-04T09:29:02.6245448Z inflating: build/bin/c10_Half_test 2025-12-04T09:29:02.6299480Z inflating: build/bin/c10_SizesAndStrides_test 2025-12-04T09:29:02.6346903Z inflating: build/bin/c10_ConstexprCrc_test 2025-12-04T09:29:02.6397945Z inflating: build/bin/c10_Bitset_test 2025-12-04T09:29:02.6451689Z inflating: build/bin/c10_LeftRight_test 2025-12-04T09:29:02.6502706Z inflating: build/bin/c10_IntrusiveList_test 2025-12-04T09:29:02.6554851Z inflating: build/bin/c10_InlineDeviceGuard_test 2025-12-04T09:29:02.6609068Z inflating: build/bin/c10_SymInt_test 2025-12-04T09:29:02.6660512Z inflating: build/bin/c10_NetworkFlow_test 2025-12-04T09:29:02.6715665Z inflating: build/bin/c10_Enumerate_test 2025-12-04T09:29:02.6763903Z inflating: build/bin/c10_DeadlockDetection_test 2025-12-04T09:29:02.6817429Z inflating: build/bin/c10_ThreadLocal_test 2025-12-04T09:29:02.6867470Z inflating: build/bin/c10_TypeIndex_test 2025-12-04T09:29:02.6921236Z inflating: build/bin/c10_InlineStreamGuard_test 2025-12-04T09:29:02.6974824Z inflating: build/bin/c10_bfloat16_test 2025-12-04T09:29:02.7022993Z inflating: build/bin/c10_Semaphore_test 2025-12-04T09:29:02.7072959Z inflating: build/bin/c10_accumulate_test 2025-12-04T09:29:02.7121570Z inflating: build/bin/c10_Synchronized_test 2025-12-04T09:29:02.7169579Z inflating: build/bin/c10_error_test 2025-12-04T09:29:02.7218627Z inflating: build/bin/c10_bit_cast_test 2025-12-04T09:29:02.7270178Z inflating: build/bin/c10_lazy_test 2025-12-04T09:29:02.7336770Z inflating: build/bin/c10_cow_test 2025-12-04T09:29:02.7385117Z inflating: build/bin/c10_ArrayRef_test 2025-12-04T09:29:02.7434736Z inflating: build/bin/c10_ssize_test 2025-12-04T09:29:02.7489209Z inflating: build/bin/c10_complex_math_test 2025-12-04T09:29:02.7538480Z inflating: build/bin/c10_irange_test 2025-12-04T09:29:02.7597786Z inflating: build/bin/c10_ordered_preserving_dict_test 2025-12-04T09:29:02.7649007Z inflating: build/bin/c10_registry_test 2025-12-04T09:29:02.7698119Z inflating: build/bin/c10_flags_test 2025-12-04T09:29:02.7751116Z inflating: build/bin/c10_complex_test 2025-12-04T09:29:02.7800254Z inflating: build/bin/c10_generic_math_test 2025-12-04T09:29:02.7850761Z inflating: build/bin/c10_exception_test 2025-12-04T09:29:02.7905069Z inflating: build/bin/c10_string_util_test 2025-12-04T09:29:02.7953279Z inflating: build/bin/c10_nofatal_test 2025-12-04T09:29:02.8001895Z inflating: build/bin/c10_tempfile_test 2025-12-04T09:29:02.8072752Z inflating: build/bin/c10_optional_test 2025-12-04T09:29:02.8214399Z inflating: build/bin/c10_small_vector_test 2025-12-04T09:29:02.8359764Z inflating: build/bin/c10_intrusive_ptr_test 2025-12-04T09:29:02.8413723Z inflating: build/bin/c10_typeid_test 2025-12-04T09:29:02.8456112Z inflating: build/bin/c10_intrusive_ptr_benchmark 2025-12-04T09:29:02.8510899Z inflating: build/bin/c10_logging_test 2025-12-04T09:29:02.8558372Z inflating: build/bin/c10_string_view_test 2025-12-04T09:29:02.8606491Z inflating: build/bin/c10_cuda_CUDATest 2025-12-04T09:29:02.8654765Z inflating: build/bin/test_vec_half_AVX2 2025-12-04T09:29:02.9183841Z inflating: build/bin/vec_test_all_types_DEFAULT 2025-12-04T09:29:02.9733071Z inflating: build/bin/vec_test_all_types_AVX2 2025-12-04T09:29:02.9781451Z inflating: build/bin/test_vec_half_AVX512 2025-12-04T09:29:02.9872550Z inflating: build/bin/test_aoti_abi_check 2025-12-04T09:29:02.9920245Z inflating: build/bin/test_vec_half_DEFAULT 2025-12-04T09:29:03.0461599Z inflating: build/bin/vec_test_all_types_AVX512 2025-12-04T09:29:03.0513001Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_1_var_test 2025-12-04T09:29:03.0564157Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_same_block 2025-12-04T09:29:03.0615772Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_multiple_blocks 2025-12-04T09:29:03.0667114Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_catches_thread_and_block_and_device 2025-12-04T09:29:03.0717813Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_from_2_processes 2025-12-04T09:29:03.0768673Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_blocks_and_threads 2025-12-04T09:29:03.0820611Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_catches_stream 2025-12-04T09:29:03.0871946Z inflating: build/bin/FileStoreTest 2025-12-04T09:29:03.0926965Z inflating: build/bin/TCPStoreTest 2025-12-04T09:29:03.0978742Z inflating: build/bin/HashStoreTest 2025-12-04T09:29:03.1029833Z inflating: build/bin/BackoffTest 2025-12-04T09:29:03.1042318Z inflating: build/bin/ProcessGroupMPITest 2025-12-04T09:29:03.1045873Z inflating: build/bin/torch_shm_manager 2025-12-04T09:29:03.1048697Z inflating: build/bin/example_allreduce 2025-12-04T09:29:03.1109949Z inflating: build/bin/ProcessGroupNCCLTest 2025-12-04T09:29:03.1173632Z inflating: build/bin/ProcessGroupGlooTest 2025-12-04T09:29:03.1227828Z inflating: build/bin/ProcessGroupGlooAsyncTest 2025-12-04T09:29:03.1286442Z inflating: build/bin/ProcessGroupNCCLErrorsTest 2025-12-04T09:29:03.1338490Z inflating: build/bin/broadcast_test 2025-12-04T09:29:03.1399424Z inflating: build/bin/basic 2025-12-04T09:29:03.1455549Z inflating: build/bin/apply_utils_test 2025-12-04T09:29:03.1509055Z inflating: build/bin/half_test 2025-12-04T09:29:03.1564382Z inflating: build/bin/cpu_generator_test 2025-12-04T09:29:03.1633725Z inflating: build/bin/Dict_test 2025-12-04T09:29:03.1685159Z inflating: build/bin/mobile_memory_cleanup 2025-12-04T09:29:03.1771392Z inflating: build/bin/cpu_rng_test 2025-12-04T09:29:03.1826561Z inflating: build/bin/extension_backend_test 2025-12-04T09:29:03.1881258Z inflating: build/bin/NamedTensor_test 2025-12-04T09:29:03.1930148Z inflating: build/bin/cpu_allocator_test 2025-12-04T09:29:03.1984641Z inflating: build/bin/quantized_test 2025-12-04T09:29:03.2036002Z inflating: build/bin/memory_overlapping_test 2025-12-04T09:29:03.2084252Z inflating: build/bin/lazy_tensor_test 2025-12-04T09:29:03.2146147Z inflating: build/bin/MaybeOwned_test 2025-12-04T09:29:03.2195175Z inflating: build/bin/operator_name_test 2025-12-04T09:29:03.2244295Z inflating: build/bin/dlconvertor_test 2025-12-04T09:29:03.2295310Z inflating: build/bin/math_kernel_test 2025-12-04T09:29:03.2345145Z inflating: build/bin/StorageUtils_test 2025-12-04T09:29:03.2394509Z inflating: build/bin/reportMemoryUsage_test 2025-12-04T09:29:03.2445017Z inflating: build/bin/Dimname_test 2025-12-04T09:29:03.2493585Z inflating: build/bin/thread_init_test 2025-12-04T09:29:03.2584572Z inflating: build/bin/ivalue_test 2025-12-04T09:29:03.2640814Z inflating: build/bin/atest 2025-12-04T09:29:03.2692116Z inflating: build/bin/cpu_profiling_allocator_test 2025-12-04T09:29:03.2741880Z inflating: build/bin/stride_properties_test 2025-12-04T09:29:03.2794958Z inflating: build/bin/type_ptr_test 2025-12-04T09:29:03.2844294Z inflating: build/bin/xla_tensor_test 2025-12-04T09:29:03.2895095Z inflating: build/bin/undefined_tensor_test 2025-12-04T09:29:03.2959445Z inflating: build/bin/pow_test 2025-12-04T09:29:03.3008427Z inflating: build/bin/operators_test 2025-12-04T09:29:03.3119541Z inflating: build/bin/kernel_function_legacy_test 2025-12-04T09:29:03.3170595Z inflating: build/bin/memory_format_test 2025-12-04T09:29:03.3227448Z inflating: build/bin/type_test 2025-12-04T09:29:03.3275561Z inflating: build/bin/verify_api_visibility 2025-12-04T09:29:03.3329046Z inflating: build/bin/scalar_tensor_test 2025-12-04T09:29:03.3391871Z inflating: build/bin/KernelFunction_test 2025-12-04T09:29:03.3439232Z inflating: build/bin/op_allowlist_test 2025-12-04T09:29:03.3488931Z inflating: build/bin/wrapdim_test 2025-12-04T09:29:03.3545746Z inflating: build/bin/IListRef_test 2025-12-04T09:29:03.3596209Z inflating: build/bin/packedtensoraccessor_test 2025-12-04T09:29:03.3649259Z inflating: build/bin/backend_fallback_test 2025-12-04T09:29:03.3729136Z inflating: build/bin/cuda_distributions_test 2025-12-04T09:29:03.3846142Z inflating: build/bin/kernel_lambda_legacy_test 2025-12-04T09:29:03.3899957Z inflating: build/bin/native_test 2025-12-04T09:29:03.3968329Z inflating: build/bin/cuda_complex_math_test 2025-12-04T09:29:03.4031785Z inflating: build/bin/inline_container_test 2025-12-04T09:29:03.4106488Z inflating: build/bin/tensor_iterator_test 2025-12-04T09:29:03.4157861Z inflating: build/bin/cuda_allocator_test 2025-12-04T09:29:03.4247046Z inflating: build/bin/make_boxed_from_unboxed_functor_test 2025-12-04T09:29:03.4304387Z inflating: build/bin/cuda_complex_test 2025-12-04T09:29:03.4399316Z inflating: build/bin/kernel_lambda_test 2025-12-04T09:29:03.4448632Z inflating: build/bin/weakref_test 2025-12-04T09:29:03.4547200Z inflating: build/bin/List_test 2025-12-04T09:29:03.4829379Z inflating: build/bin/op_registration_test 2025-12-04T09:29:03.4896501Z inflating: build/bin/legacy_vmap_test 2025-12-04T09:29:03.4946418Z inflating: build/bin/cuda_dlconvertor_test 2025-12-04T09:29:03.5005041Z inflating: build/bin/cuda_atomic_ops_test 2025-12-04T09:29:03.5054246Z inflating: build/bin/CppSignature_test 2025-12-04T09:29:03.5109071Z inflating: build/bin/scalar_test 2025-12-04T09:29:03.5163042Z inflating: build/bin/cuda_caching_host_allocator_test 2025-12-04T09:29:03.5211665Z inflating: build/bin/cuda_half_test 2025-12-04T09:29:03.5259977Z inflating: build/bin/reduce_ops_test 2025-12-04T09:29:03.5310914Z inflating: build/bin/cuda_cublas_handle_pool_test 2025-12-04T09:29:03.5373396Z inflating: build/bin/cuda_cub_test 2025-12-04T09:29:03.5428085Z inflating: build/bin/cuda_generator_test 2025-12-04T09:29:03.5517167Z inflating: build/bin/kernel_function_test 2025-12-04T09:29:03.5564888Z inflating: build/bin/cuda_exchange_device_test 2025-12-04T09:29:03.5612900Z inflating: build/bin/cuda_device_test 2025-12-04T09:29:03.5662701Z inflating: build/bin/cuda_integer_divider_test 2025-12-04T09:29:03.5714228Z inflating: build/bin/cuda_event_test 2025-12-04T09:29:03.5762584Z inflating: build/bin/cuda_optional_test 2025-12-04T09:29:03.5813544Z inflating: build/bin/cuda_apply_test 2025-12-04T09:29:03.5865703Z inflating: build/bin/test_parallel 2025-12-04T09:29:03.5923763Z inflating: build/bin/kernel_stackbased_test 2025-12-04T09:29:03.5974658Z inflating: build/bin/cuda_reportMemoryUsage_test 2025-12-04T09:29:03.6032906Z inflating: build/bin/cuda_stream_test 2025-12-04T09:29:03.6084016Z inflating: build/bin/cuda_vectorized_test 2025-12-04T09:29:03.6134750Z inflating: build/bin/cuda_packedtensoraccessor_test 2025-12-04T09:29:03.6183383Z inflating: build/bin/cuda_allocatorTraceTracker_test 2025-12-04T09:29:03.6497691Z inflating: build/bin/test_lazy 2025-12-04T09:29:03.6550575Z inflating: build/bin/test_dist_autograd 2025-12-04T09:29:03.6615618Z inflating: build/bin/test_cpp_rpc 2025-12-04T09:29:03.7659572Z inflating: build/bin/test_api 2025-12-04T09:29:03.7708107Z inflating: build/bin/cuda_cudnn_test 2025-12-04T09:29:03.7710548Z inflating: build/bin/parallel_benchmark 2025-12-04T09:29:03.8687306Z inflating: build/bin/test_jit 2025-12-04T09:29:03.8687638Z creating: .additional_ci_files/ 2025-12-04T09:29:03.8743906Z inflating: .additional_ci_files/test-times.json 2025-12-04T09:29:03.8948843Z inflating: .additional_ci_files/test-class-times.json 2025-12-04T09:29:03.8978848Z ##[group]Run rm artifacts.zip 2025-12-04T09:29:03.8979075Z rm artifacts.zip 2025-12-04T09:29:03.8988882Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:29:03.8989168Z env: 2025-12-04T09:29:03.8989316Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:29:03.8989518Z HAS_NVIDIA_GPU: true 2025-12-04T09:29:03.8989746Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:29:03.8990003Z ##[endgroup] 2025-12-04T09:29:04.0870556Z ##[group]Run df -H 2025-12-04T09:29:04.0870738Z df -H 2025-12-04T09:29:04.0878009Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:29:04.0878284Z env: 2025-12-04T09:29:04.0878440Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:29:04.0878629Z HAS_NVIDIA_GPU: true 2025-12-04T09:29:04.0878895Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:29:04.0879165Z ##[endgroup] 2025-12-04T09:29:04.0928571Z Filesystem Size Used Avail Use% Mounted on 2025-12-04T09:29:04.0928950Z devtmpfs 4.2M 0 4.2M 0% /dev 2025-12-04T09:29:04.0929260Z tmpfs 33G 0 33G 0% /dev/shm 2025-12-04T09:29:04.0929545Z tmpfs 13G 779k 13G 1% /run 2025-12-04T09:29:04.0929824Z /dev/nvme0n1p1 161G 56G 106G 35% / 2025-12-04T09:29:04.0930134Z tmpfs 33G 17k 33G 1% /tmp 2025-12-04T09:29:04.0930434Z /dev/nvme0n1p128 11M 1.4M 9.2M 13% /boot/efi 2025-12-04T09:29:04.0930754Z tmpfs 6.5G 0 6.5G 0% /run/user/0 2025-12-04T09:29:04.0958996Z Prepare all required actions 2025-12-04T09:29:04.0959809Z Getting action download info 2025-12-04T09:29:04.2776274Z ##[group]Run ./.github/actions/download-td-artifacts 2025-12-04T09:29:04.2776545Z with: 2025-12-04T09:29:04.2776693Z env: 2025-12-04T09:29:04.2776845Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:29:04.2777056Z HAS_NVIDIA_GPU: true 2025-12-04T09:29:04.2777289Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:29:04.2777543Z ##[endgroup] 2025-12-04T09:29:04.2800653Z ##[group]Run seemethere/download-artifact-s3@v4 2025-12-04T09:29:04.2800895Z with: 2025-12-04T09:29:04.2801047Z name: td_results 2025-12-04T09:29:04.2801225Z s3-bucket: gha-artifacts 2025-12-04T09:29:04.2801416Z region: us-east-1 2025-12-04T09:29:04.2801565Z env: 2025-12-04T09:29:04.2801713Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:29:04.2801900Z HAS_NVIDIA_GPU: true 2025-12-04T09:29:04.2802122Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:29:04.2802418Z ##[endgroup] 2025-12-04T09:29:04.7132142Z (node:61375) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-12-04T09:29:04.7132618Z 2025-12-04T09:29:04.7132803Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-12-04T09:29:04.7133288Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-12-04T09:29:04.7134175Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-12-04T09:29:04.8028755Z Found 0 objects with prefix pytorch/pytorch/19922031852/td_results/ 2025-12-04T09:29:04.8034534Z Artifact download has finished successfully 2025-12-04T09:29:04.8276056Z ##[group]Run mkdir -p .additional_ci_files 2025-12-04T09:29:04.8276343Z mkdir -p .additional_ci_files 2025-12-04T09:29:04.8276658Z mv td_results.json .additional_ci_files/td_results.json || true 2025-12-04T09:29:04.8284672Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:29:04.8284997Z env: 2025-12-04T09:29:04.8285180Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:29:04.8285379Z HAS_NVIDIA_GPU: true 2025-12-04T09:29:04.8285600Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:29:04.8285855Z ##[endgroup] 2025-12-04T09:29:04.8348245Z mv: cannot stat 'td_results.json': No such file or directory 2025-12-04T09:29:04.8386883Z ##[group]Run .github/scripts/parse_ref.py 2025-12-04T09:29:04.8387209Z .github/scripts/parse_ref.py 2025-12-04T09:29:04.8394279Z shell: /usr/bin/bash -e {0} 2025-12-04T09:29:04.8394488Z env: 2025-12-04T09:29:04.8394644Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:29:04.8394832Z HAS_NVIDIA_GPU: true 2025-12-04T09:29:04.8395048Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:29:04.8395306Z ##[endgroup] 2025-12-04T09:29:04.8626723Z Setting output branch=main 2025-12-04T09:29:04.8725766Z Prepare all required actions 2025-12-04T09:29:04.8726100Z Getting action download info 2025-12-04T09:29:05.0089431Z ##[group]Run ./.github/actions/filter-test-configs 2025-12-04T09:29:05.0089675Z with: 2025-12-04T09:29:05.0089991Z github-token: *** 2025-12-04T09:29:05.0095690Z test-matrix: {"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_language_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu", "unstable": "unstable"}]} 2025-12-04T09:29:05.0101942Z job-name: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu, unstable) 2025-12-04T09:29:05.0102340Z env: 2025-12-04T09:29:05.0102492Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:29:05.0102682Z HAS_NVIDIA_GPU: true 2025-12-04T09:29:05.0102904Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:29:05.0103158Z ##[endgroup] 2025-12-04T09:29:05.0131347Z ##[group]Run nick-fields/retry@v3.0.0 2025-12-04T09:29:05.0131569Z with: 2025-12-04T09:29:05.0131715Z shell: bash 2025-12-04T09:29:05.0131874Z timeout_minutes: 10 2025-12-04T09:29:05.0132037Z max_attempts: 5 2025-12-04T09:29:05.0132205Z retry_wait_seconds: 30 2025-12-04T09:29:05.0132783Z command: set -eux # PyYAML 6.0 doesn't work with MacOS x86 anymore # This must run on Python-3.7 (AmazonLinux2) so can't use request=3.32.2 python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-12-04T09:29:05.0133394Z polling_interval_seconds: 1 2025-12-04T09:29:05.0133589Z warning_on_retry: true 2025-12-04T09:29:05.0133770Z continue_on_error: false 2025-12-04T09:29:05.0133948Z env: 2025-12-04T09:29:05.0134088Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:29:05.0134272Z HAS_NVIDIA_GPU: true 2025-12-04T09:29:05.0134491Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:29:05.0135068Z GITHUB_TOKEN: *** 2025-12-04T09:29:05.0135255Z ##[endgroup] 2025-12-04T09:29:05.1105752Z + python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-12-04T09:29:05.3248980Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T09:29:05.4294215Z Collecting requests==2.27.1 2025-12-04T09:29:05.4481181Z Downloading requests-2.27.1-py2.py3-none-any.whl (63 kB) 2025-12-04T09:29:05.6093519Z Collecting pyyaml==6.0.2 2025-12-04T09:29:05.6136826Z Downloading PyYAML-6.0.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (737 kB) 2025-12-04T09:29:05.6373131Z Requirement already satisfied: urllib3<1.27,>=1.21.1 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (1.25.10) 2025-12-04T09:29:05.6381367Z Requirement already satisfied: idna<4,>=2.5 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (2.10) 2025-12-04T09:29:05.9846639Z Collecting charset-normalizer~=2.0.0 2025-12-04T09:29:05.9890349Z Downloading charset_normalizer-2.0.12-py3-none-any.whl (39 kB) 2025-12-04T09:29:06.0378017Z Collecting certifi>=2017.4.17 2025-12-04T09:29:06.0421214Z Downloading certifi-2025.11.12-py3-none-any.whl (159 kB) 2025-12-04T09:29:06.1196183Z Installing collected packages: charset-normalizer, certifi, requests, pyyaml 2025-12-04T09:29:06.2300884Z Successfully installed certifi-2025.11.12 charset-normalizer-2.0.12 pyyaml-6.0.2 requests-2.27.1 2025-12-04T09:29:07.0877298Z Command completed after 1 attempt(s). 2025-12-04T09:29:07.0947681Z ##[group]Run set -x 2025-12-04T09:29:07.0947911Z set -x 2025-12-04T09:29:07.0948071Z  2025-12-04T09:29:07.0948341Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-12-04T09:29:07.0948683Z # in runner workspace 2025-12-04T09:29:07.0948957Z python3 "${GITHUB_ACTION_PATH}/../../scripts/parse_ref.py" 2025-12-04T09:29:07.0956807Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:29:07.0957076Z env: 2025-12-04T09:29:07.0957251Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:29:07.0957435Z HAS_NVIDIA_GPU: true 2025-12-04T09:29:07.0957648Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:29:07.0957896Z ##[endgroup] 2025-12-04T09:29:07.0987715Z + python3 /home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/filter-test-configs/../../scripts/parse_ref.py 2025-12-04T09:29:07.1157985Z Setting output branch=main 2025-12-04T09:29:07.1205504Z ##[group]Run echo "Workflow: ${GITHUB_WORKFLOW}" 2025-12-04T09:29:07.1205806Z echo "Workflow: ${GITHUB_WORKFLOW}" 2025-12-04T09:29:07.1206253Z echo "Job name: ${JOB_NAME}" 2025-12-04T09:29:07.1206452Z  2025-12-04T09:29:07.1206716Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-12-04T09:29:07.1207053Z # in runner workspace 2025-12-04T09:29:07.1207352Z python3 "${GITHUB_ACTION_PATH}/../../scripts/filter_test_configs.py" \ 2025-12-04T09:29:07.1207701Z  --workflow "${GITHUB_WORKFLOW}" \ 2025-12-04T09:29:07.1207985Z  --job-name "${JOB_NAME}" \ 2025-12-04T09:29:07.1213827Z  --test-matrix "{"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_language_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu", "unstable": "unstable"}]}" \ 2025-12-04T09:29:07.1219791Z  --selected-test-configs "" \ 2025-12-04T09:29:07.1220036Z  --pr-number "${PR_NUMBER}" \ 2025-12-04T09:29:07.1220256Z  --tag "${TAG}" \ 2025-12-04T09:29:07.1220463Z  --event-name "${EVENT_NAME}" \ 2025-12-04T09:29:07.1220685Z  --schedule "${SCHEDULE}" \ 2025-12-04T09:29:07.1220903Z  --branch "${HEAD_BRANCH}" 2025-12-04T09:29:07.1228097Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:29:07.1228376Z env: 2025-12-04T09:29:07.1228540Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:29:07.1228726Z HAS_NVIDIA_GPU: true 2025-12-04T09:29:07.1228943Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:29:07.1229390Z GITHUB_TOKEN: *** 2025-12-04T09:29:07.1229771Z JOB_NAME: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu, unstable) 2025-12-04T09:29:07.1230173Z PR_NUMBER: 2025-12-04T09:29:07.1230321Z TAG: 2025-12-04T09:29:07.1230468Z EVENT_NAME: schedule 2025-12-04T09:29:07.1230757Z SCHEDULE: 0 */8 * * * 2025-12-04T09:29:07.1230926Z HEAD_BRANCH: main 2025-12-04T09:29:07.1231089Z ##[endgroup] 2025-12-04T09:29:07.1256965Z Workflow: vllm-test 2025-12-04T09:29:07.1257467Z Job name: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu, unstable) 2025-12-04T09:29:07.2973885Z INFO:root:Issue https://github.com/pytorch/pytorch/issues/169298 created by huydhn has unstable all the test jobs for vllm-test / ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu, unstable) 2025-12-04T09:29:07.3028958Z Setting output keep-going=True 2025-12-04T09:29:07.3029276Z Setting output ci-verbose-test-logs=False 2025-12-04T09:29:07.3029588Z Setting output ci-test-showlocals=False 2025-12-04T09:29:07.3029889Z Setting output ci-no-test-timeout=False 2025-12-04T09:29:07.3030170Z Setting output ci-no-td=False 2025-12-04T09:29:07.3030449Z Setting output ci-td-distributed=False 2025-12-04T09:29:07.3030747Z Setting output is-unstable=True 2025-12-04T09:29:07.3031012Z Setting output reenabled-issues= 2025-12-04T09:29:07.3038452Z Setting output test-matrix={"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_language_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu", "unstable": "unstable"}]} 2025-12-04T09:29:07.3044705Z Setting output is-test-matrix-empty=False 2025-12-04T09:29:07.3147217Z ##[group]Run echo "Filtered matrix:" 2025-12-04T09:29:07.3147667Z echo "Filtered matrix:" 2025-12-04T09:29:07.3157980Z echo "{"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_language_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu", "unstable": "unstable"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu", "unstable": "unstable"}]}" 2025-12-04T09:29:07.3168817Z  2025-12-04T09:29:07.3169076Z echo 2025-12-04T09:29:07.3169410Z echo "Is the current job unstable? True" 2025-12-04T09:29:07.3169842Z  2025-12-04T09:29:07.3170094Z echo 2025-12-04T09:29:07.3170398Z echo "Is keep-going label set? True" 2025-12-04T09:29:07.3170801Z  2025-12-04T09:29:07.3171321Z echo 2025-12-04T09:29:07.3171503Z echo "Reenabled issues? " 2025-12-04T09:29:07.3179417Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:29:07.3179692Z env: 2025-12-04T09:29:07.3179845Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:29:07.3180054Z HAS_NVIDIA_GPU: true 2025-12-04T09:29:07.3180290Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:29:07.3180546Z ##[endgroup] 2025-12-04T09:29:07.3207754Z Filtered matrix: 2025-12-04T09:29:07.3215074Z {include: [{config: vllm_basic_correctness_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu, unstable: unstable}, {config: vllm_basic_models_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu, unstable: unstable}, {config: vllm_entrypoints_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu, unstable: unstable}, {config: vllm_regression_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu, unstable: unstable}, {config: vllm_multi_model_processor_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu, unstable: unstable}, {config: vllm_pytorch_compilation_unit_tests, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu, unstable: unstable}, {config: vllm_lora_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu, unstable: unstable}, {config: vllm_multi_model_test_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu, unstable: unstable}, {config: vllm_language_model_test_extended_generation_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu, unstable: unstable}, {config: vllm_distributed_test_2_gpu_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu, unstable: unstable}, {config: vllm_lora_test, shard: 0, num_shards: 4, runner: linux.g6.4xlarge.experimental.nvidia.gpu, unstable: unstable}, {config: vllm_lora_test, shard: 1, num_shards: 4, runner: linux.g6.4xlarge.experimental.nvidia.gpu, unstable: unstable}, {config: vllm_lora_test, shard: 2, num_shards: 4, runner: linux.g6.4xlarge.experimental.nvidia.gpu, unstable: unstable}, {config: vllm_lora_test, shard: 3, num_shards: 4, runner: linux.g6.4xlarge.experimental.nvidia.gpu, unstable: unstable}, {config: vllm_lora_tp_test_distributed, shard: 1, num_shards: 1, runner: linux.g6.12xlarge.nvidia.gpu, unstable: unstable}, {config: vllm_distributed_test_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.12xlarge.nvidia.gpu, unstable: unstable}]} 2025-12-04T09:29:07.3221160Z 2025-12-04T09:29:07.3221259Z Is the current job unstable? True 2025-12-04T09:29:07.3221401Z 2025-12-04T09:29:07.3221477Z Is keep-going label set? True 2025-12-04T09:29:07.3221614Z 2025-12-04T09:29:07.3221679Z Reenabled issues? 2025-12-04T09:29:07.3248081Z ##[group]Run echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-12-04T09:29:07.3248499Z echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-12-04T09:29:07.3255567Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:29:07.3255846Z env: 2025-12-04T09:29:07.3256004Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:29:07.3256196Z HAS_NVIDIA_GPU: true 2025-12-04T09:29:07.3256416Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:29:07.3256674Z JOB_TIMEOUT: 240 2025-12-04T09:29:07.3256842Z ##[endgroup] 2025-12-04T09:29:07.3305354Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:29:07.3305785Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:29:07.3306110Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:29:07.3313068Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:29:07.3313354Z env: 2025-12-04T09:29:07.3313513Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:29:07.3313701Z HAS_NVIDIA_GPU: true 2025-12-04T09:29:07.3313923Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:29:07.3314177Z ##[endgroup] 2025-12-04T09:29:07.3413513Z ##[group]Run set -x 2025-12-04T09:29:07.3413796Z set -x 2025-12-04T09:29:07.3413960Z  2025-12-04T09:29:07.3414147Z if [[ $TEST_CONFIG == 'multigpu' ]]; then 2025-12-04T09:29:07.3414424Z  TEST_COMMAND=.ci/pytorch/multigpu-test.sh 2025-12-04T09:29:07.3414705Z elif [[ $BUILD_ENVIRONMENT == *onnx* ]]; then 2025-12-04T09:29:07.3414962Z  TEST_COMMAND=.ci/onnx/test.sh 2025-12-04T09:29:07.3415181Z else 2025-12-04T09:29:07.3415362Z  TEST_COMMAND=.ci/pytorch/test.sh 2025-12-04T09:29:07.3415580Z fi 2025-12-04T09:29:07.3415715Z  2025-12-04T09:29:07.3415926Z # Leaving 1GB for the runner and other things 2025-12-04T09:29:07.3416617Z TOTAL_AVAILABLE_MEMORY_IN_GB=$(awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo) 2025-12-04T09:29:07.3417327Z # https://docs.docker.com/engine/containers/resource_constraints/#--memory-swap-details, the 3GB swap 2025-12-04T09:29:07.3417830Z # comes from https://github.com/pytorch/test-infra/pull/6058 2025-12-04T09:29:07.3418222Z TOTAL_MEMORY_WITH_SWAP=$(("${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}" + 3)) 2025-12-04T09:29:07.3418513Z  2025-12-04T09:29:07.3418704Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-12-04T09:29:07.3418953Z  SHM_OPTS= 2025-12-04T09:29:07.3419211Z  JENKINS_USER= 2025-12-04T09:29:07.3419648Z  # ensure that docker container cleanly exits in 12 hours 2025-12-04T09:29:07.3419991Z  # if for some reason cleanup action doesn't stop container 2025-12-04T09:29:07.3420268Z  # when job is cancelled 2025-12-04T09:29:07.3420487Z  DOCKER_SHELL_CMD="sleep 12h" 2025-12-04T09:29:07.3420719Z  USED_IMAGE="${DOCKER_IMAGE_S390X}" 2025-12-04T09:29:07.3420956Z else 2025-12-04T09:29:07.3421144Z  SHM_OPTS="--shm-size=${SHM_SIZE}" 2025-12-04T09:29:07.3421379Z  JENKINS_USER="--user jenkins" 2025-12-04T09:29:07.3421608Z  DOCKER_SHELL_CMD= 2025-12-04T09:29:07.3421813Z  USED_IMAGE="${DOCKER_IMAGE}" 2025-12-04T09:29:07.3422015Z fi 2025-12-04T09:29:07.3422158Z  2025-12-04T09:29:07.3422391Z # detached container should get cleaned up by teardown_ec2_linux 2025-12-04T09:29:07.3422765Z # TODO: Stop building test binaries as part of the build phase 2025-12-04T09:29:07.3423189Z # Used for GPU_FLAG, SHM_OPTS, JENKINS_USER and DOCKER_SHELL_CMD since that doesn't play nice 2025-12-04T09:29:07.3423563Z # shellcheck disable=SC2086,SC2090 2025-12-04T09:29:07.3423800Z container_name=$(docker run \ 2025-12-04T09:29:07.3424020Z  ${GPU_FLAG:-} \ 2025-12-04T09:29:07.3424235Z  ${SCCACHE_SERVER_PORT_DOCKER_FLAG:-} \ 2025-12-04T09:29:07.3424479Z  -e BUILD_ENVIRONMENT \ 2025-12-04T09:29:07.3424685Z  -e PR_NUMBER \ 2025-12-04T09:29:07.3424874Z  -e GITHUB_ACTIONS \ 2025-12-04T09:29:07.3425079Z  -e GITHUB_REPOSITORY \ 2025-12-04T09:29:07.3425286Z  -e GITHUB_WORKFLOW \ 2025-12-04T09:29:07.3425478Z  -e GITHUB_JOB \ 2025-12-04T09:29:07.3425679Z  -e GITHUB_RUN_ID \ 2025-12-04T09:29:07.3425874Z  -e GITHUB_RUN_NUMBER \ 2025-12-04T09:29:07.3426070Z  -e GITHUB_RUN_ATTEMPT \ 2025-12-04T09:29:07.3426276Z  -e JOB_ID \ 2025-12-04T09:29:07.3426458Z  -e JOB_NAME \ 2025-12-04T09:29:07.3426638Z  -e BASE_SHA \ 2025-12-04T09:29:07.3426812Z  -e BRANCH \ 2025-12-04T09:29:07.3426992Z  -e SHA1 \ 2025-12-04T09:29:07.3427166Z  -e AWS_DEFAULT_REGION \ 2025-12-04T09:29:07.3427364Z  -e IN_WHEEL_TEST \ 2025-12-04T09:29:07.3427555Z  -e SHARD_NUMBER \ 2025-12-04T09:29:07.3427746Z  -e TEST_CONFIG \ 2025-12-04T09:29:07.3427931Z  -e NUM_TEST_SHARDS \ 2025-12-04T09:29:07.3428261Z  -e REENABLED_ISSUES \ 2025-12-04T09:29:07.3428483Z  -e CONTINUE_THROUGH_ERROR \ 2025-12-04T09:29:07.3428713Z  -e VERBOSE_TEST_LOGS \ 2025-12-04T09:29:07.3428917Z  -e TEST_SHOWLOCALS \ 2025-12-04T09:29:07.3429113Z  -e NO_TEST_TIMEOUT \ 2025-12-04T09:29:07.3429303Z  -e NO_TD \ 2025-12-04T09:29:07.3429475Z  -e TD_DISTRIBUTED \ 2025-12-04T09:29:07.3429669Z  -e PR_LABELS \ 2025-12-04T09:29:07.3429882Z  -e MAX_JOBS="$(nproc --ignore=2)" \ 2025-12-04T09:29:07.3430109Z  -e SCCACHE_BUCKET \ 2025-12-04T09:29:07.3430302Z  -e SCCACHE_REGION \ 2025-12-04T09:29:07.3430502Z  -e XLA_CUDA \ 2025-12-04T09:29:07.3430698Z  -e XLA_CLANG_CACHE_S3_BUCKET_NAME \ 2025-12-04T09:29:07.3430950Z  -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK \ 2025-12-04T09:29:07.3431221Z  -e PYTORCH_TEST_RERUN_DISABLED_TESTS \ 2025-12-04T09:29:07.3431485Z  -e SKIP_SCCACHE_INITIALIZATION=1 \ 2025-12-04T09:29:07.3431721Z  -e HUGGING_FACE_HUB_TOKEN \ 2025-12-04T09:29:07.3431956Z  -e VLLM_TEST_HUGGING_FACE_TOKEN \ 2025-12-04T09:29:07.3432192Z  -e SCRIBE_GRAPHQL_ACCESS_TOKEN \ 2025-12-04T09:29:07.3432409Z  -e DASHBOARD_TAG \ 2025-12-04T09:29:07.3432610Z  -e ARTIFACTS_FILE_SUFFIX \ 2025-12-04T09:29:07.3432866Z  --memory="${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}g" \ 2025-12-04T09:29:07.3433248Z  --memory-swap="${TOTAL_MEMORY_WITH_SWAP}g" \ 2025-12-04T09:29:07.3433543Z  --env-file="/tmp/github_env_${GITHUB_RUN_ID}" \ 2025-12-04T09:29:07.3433821Z  --security-opt seccomp=unconfined \ 2025-12-04T09:29:07.3434062Z  --cap-add=SYS_PTRACE \ 2025-12-04T09:29:07.3434259Z  --ipc=host \ 2025-12-04T09:29:07.3434438Z  ${SHM_OPTS} \ 2025-12-04T09:29:07.3434611Z  --tty \ 2025-12-04T09:29:07.3434771Z  --detach \ 2025-12-04T09:29:07.3434960Z  --name="${container_name}" \ 2025-12-04T09:29:07.3435186Z  ${JENKINS_USER} \ 2025-12-04T09:29:07.3435423Z  -v "${GITHUB_WORKSPACE}:/var/lib/jenkins/workspace" \ 2025-12-04T09:29:07.3435715Z  -w /var/lib/jenkins/workspace \ 2025-12-04T09:29:07.3435935Z  "${USED_IMAGE}" \ 2025-12-04T09:29:07.3436118Z  ${DOCKER_SHELL_CMD} 2025-12-04T09:29:07.3436298Z ) 2025-12-04T09:29:07.3436533Z echo "DOCKER_CONTAINER_ID=${container_name}" >> "${GITHUB_ENV}" 2025-12-04T09:29:07.3436819Z  2025-12-04T09:29:07.3436996Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-12-04T09:29:07.3437407Z  docker exec -t "${container_name}" sh -c "python3 -m pip install -r .ci/docker/requirements-ci.txt" 2025-12-04T09:29:07.3437771Z fi 2025-12-04T09:29:07.3437915Z  2025-12-04T09:29:07.3438254Z docker exec -t "${container_name}" sh -c "python3 -m pip install $(echo dist/*.whl)[opt-einsum] && ${TEST_COMMAND}" 2025-12-04T09:29:07.3445270Z shell: /usr/bin/bash -e {0} 2025-12-04T09:29:07.3445470Z env: 2025-12-04T09:29:07.3445631Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:29:07.3445827Z HAS_NVIDIA_GPU: true 2025-12-04T09:29:07.3446056Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:29:07.3446364Z BUILD_ENVIRONMENT: linux-jammy-cuda12.8-py3.12-gcc11 2025-12-04T09:29:07.3446627Z PR_NUMBER: 2025-12-04T09:29:07.3446819Z GITHUB_REPOSITORY: pytorch/pytorch 2025-12-04T09:29:07.3447048Z GITHUB_WORKFLOW: vllm-test 2025-12-04T09:29:07.3447230Z GITHUB_JOB: test 2025-12-04T09:29:07.3447410Z GITHUB_RUN_ID: 19922031852 2025-12-04T09:29:07.3447601Z GITHUB_RUN_NUMBER: 4606 2025-12-04T09:29:07.3447780Z GITHUB_RUN_ATTEMPT: 1 2025-12-04T09:29:07.3447956Z JOB_ID: 57118543727 2025-12-04T09:29:07.3448330Z JOB_NAME: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu, unstable) 2025-12-04T09:29:07.3448720Z BRANCH: main 2025-12-04T09:29:07.3449021Z SHA1: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:29:07.3449304Z BASE_SHA: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:29:07.3449576Z TEST_CONFIG: vllm_basic_models_test 2025-12-04T09:29:07.3449779Z SHARD_NUMBER: 1 2025-12-04T09:29:07.3449944Z NUM_TEST_SHARDS: 1 2025-12-04T09:29:07.3450104Z EXTRA_FLAGS: 2025-12-04T09:29:07.3450260Z OP_BENCHMARK_TESTS: 2025-12-04T09:29:07.3450434Z REENABLED_ISSUES: 2025-12-04T09:29:07.3450610Z CONTINUE_THROUGH_ERROR: True 2025-12-04T09:29:07.3450799Z VERBOSE_TEST_LOGS: False 2025-12-04T09:29:07.3450981Z TEST_SHOWLOCALS: False 2025-12-04T09:29:07.3451166Z NO_TEST_TIMEOUT: False 2025-12-04T09:29:07.3451328Z NO_TD: False 2025-12-04T09:29:07.3451482Z TD_DISTRIBUTED: False 2025-12-04T09:29:07.3451701Z SCCACHE_BUCKET: ossci-compiler-cache-circleci-v2 2025-12-04T09:29:07.3451947Z SCCACHE_REGION: us-east-1 2025-12-04T09:29:07.3452145Z SHM_SIZE: 2g 2025-12-04T09:29:07.3452719Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:29:07.3453765Z DOCKER_IMAGE_S390X: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:29:07.3454389Z XLA_CUDA: 2025-12-04T09:29:07.3454642Z XLA_CLANG_CACHE_S3_BUCKET_NAME: ossci-compiler-clang-cache-circleci-xla 2025-12-04T09:29:07.3455066Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK: 0 2025-12-04T09:29:07.3455296Z PYTORCH_TEST_RERUN_DISABLED_TESTS: 0 2025-12-04T09:29:07.3455511Z DASHBOARD_TAG: 2025-12-04T09:29:07.3455851Z VLLM_TEST_HUGGING_FACE_TOKEN: *** 2025-12-04T09:29:07.3456148Z HUGGING_FACE_HUB_TOKEN: *** 2025-12-04T09:29:07.3456451Z SCRIBE_GRAPHQL_ACCESS_TOKEN: *** 2025-12-04T09:29:07.3456876Z ARTIFACTS_FILE_SUFFIX: test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_57118543727 2025-12-04T09:29:07.3457288Z ##[endgroup] 2025-12-04T09:29:07.3484139Z + [[ vllm_basic_models_test == \m\u\l\t\i\g\p\u ]] 2025-12-04T09:29:07.3484515Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *onnx* ]] 2025-12-04T09:29:07.3484839Z + TEST_COMMAND=.ci/pytorch/test.sh 2025-12-04T09:29:07.3487617Z ++ awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo 2025-12-04T09:29:07.3510178Z + TOTAL_AVAILABLE_MEMORY_IN_GB='59.453 ' 2025-12-04T09:29:07.3510474Z + TOTAL_MEMORY_WITH_SWAP=62 2025-12-04T09:29:07.3510795Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *\s\3\9\0\x* ]] 2025-12-04T09:29:07.3511115Z + SHM_OPTS=--shm-size=2g 2025-12-04T09:29:07.3511374Z + JENKINS_USER='--user jenkins' 2025-12-04T09:29:07.3511622Z + DOCKER_SHELL_CMD= 2025-12-04T09:29:07.3512355Z + USED_IMAGE=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:29:07.3519244Z +++ nproc --ignore=2 2025-12-04T09:29:07.3560127Z ++ docker run --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all -e BUILD_ENVIRONMENT -e PR_NUMBER -e GITHUB_ACTIONS -e GITHUB_REPOSITORY -e GITHUB_WORKFLOW -e GITHUB_JOB -e GITHUB_RUN_ID -e GITHUB_RUN_NUMBER -e GITHUB_RUN_ATTEMPT -e JOB_ID -e JOB_NAME -e BASE_SHA -e BRANCH -e SHA1 -e AWS_DEFAULT_REGION -e IN_WHEEL_TEST -e SHARD_NUMBER -e TEST_CONFIG -e NUM_TEST_SHARDS -e REENABLED_ISSUES -e CONTINUE_THROUGH_ERROR -e VERBOSE_TEST_LOGS -e TEST_SHOWLOCALS -e NO_TEST_TIMEOUT -e NO_TD -e TD_DISTRIBUTED -e PR_LABELS -e MAX_JOBS=14 -e SCCACHE_BUCKET -e SCCACHE_REGION -e XLA_CUDA -e XLA_CLANG_CACHE_S3_BUCKET_NAME -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK -e PYTORCH_TEST_RERUN_DISABLED_TESTS -e SKIP_SCCACHE_INITIALIZATION=1 -e HUGGING_FACE_HUB_TOKEN -e VLLM_TEST_HUGGING_FACE_TOKEN -e SCRIBE_GRAPHQL_ACCESS_TOKEN -e DASHBOARD_TAG -e ARTIFACTS_FILE_SUFFIX --memory=59g --memory-swap=62g --env-file=/tmp/github_env_19922031852 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --ipc=host --shm-size=2g --tty --detach --name= --user jenkins -v /home/ec2-user/actions-runner/_work/pytorch/pytorch:/var/lib/jenkins/workspace -w /var/lib/jenkins/workspace 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:29:28.5043867Z + container_name=dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T09:29:28.5045333Z + echo DOCKER_CONTAINER_ID=dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T09:29:28.5046005Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *\s\3\9\0\x* ]] 2025-12-04T09:29:28.5052114Z ++ echo dist/torch-2.10.0a0+gitffd9b0f-cp312-cp312-linux_x86_64.whl 2025-12-04T09:29:28.5055033Z + docker exec -t dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f sh -c 'python3 -m pip install dist/torch-2.10.0a0+gitffd9b0f-cp312-cp312-linux_x86_64.whl[opt-einsum] && .ci/pytorch/test.sh' 2025-12-04T09:29:28.9564733Z Processing ./dist/torch-2.10.0a0+gitffd9b0f-cp312-cp312-linux_x86_64.whl (from torch==2.10.0a0+gitffd9b0f) 2025-12-04T09:29:29.3537979Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.18.0) 2025-12-04T09:29:29.3539792Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (4.12.2) 2025-12-04T09:29:29.3543939Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (78.1.1) 2025-12-04T09:29:29.3546981Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (1.13.3) 2025-12-04T09:29:29.3549714Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (2.8.8) 2025-12-04T09:29:29.3552595Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.1.6) 2025-12-04T09:29:29.3555949Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (2025.10.0) 2025-12-04T09:29:29.3567160Z Requirement already satisfied: opt-einsum>=3.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.3.0) 2025-12-04T09:29:29.3660021Z Requirement already satisfied: numpy>=1.7 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from opt-einsum>=3.3->torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (1.26.2) 2025-12-04T09:29:29.3697602Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (1.3.0) 2025-12-04T09:29:29.3738491Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.0.3) 2025-12-04T09:29:29.5672682Z Installing collected packages: torch 2025-12-04T09:29:41.6472500Z Successfully installed torch-2.10.0a0+gitffd9b0f 2025-12-04T09:29:41.7076168Z + export TERM=vt100 2025-12-04T09:29:41.7076444Z + TERM=vt100 2025-12-04T09:29:41.7078605Z ++ dirname .ci/pytorch/test.sh 2025-12-04T09:29:41.7088178Z + source .ci/pytorch/common.sh 2025-12-04T09:29:41.7091863Z +++ dirname .ci/pytorch/common.sh 2025-12-04T09:29:41.7099624Z ++ source .ci/pytorch/common_utils.sh 2025-12-04T09:29:41.7101398Z +++ declare -f -t trap_add 2025-12-04T09:29:41.7105740Z ++ set -ex -o pipefail 2025-12-04T09:29:41.7106084Z ++ [[ linux-jammy-cuda12.8-py3.12-gcc11 == *rocm* ]] 2025-12-04T09:29:41.7106411Z ++ BUILD_TEST_LIBTORCH=0 2025-12-04T09:29:41.7110009Z ++ dirname .ci/pytorch/test.sh 2025-12-04T09:29:41.7118279Z + source .ci/pytorch/common-build.sh 2025-12-04T09:29:41.7119785Z ++ [[ linux-jammy-cuda12.8-py3.12-gcc11 != *win-* ]] 2025-12-04T09:29:41.7125270Z ++++ dirname .ci/pytorch/common-build.sh 2025-12-04T09:29:41.7133647Z +++ cd .ci/pytorch 2025-12-04T09:29:41.7133978Z +++ pwd -P 2025-12-04T09:29:41.7136500Z ++ script_dir=/var/lib/jenkins/workspace/.ci/pytorch 2025-12-04T09:29:41.7136939Z ++ [[ linux-jammy-cuda12.8-py3.12-gcc11 == *-pch* ]] 2025-12-04T09:29:41.7137225Z ++ which sccache 2025-12-04T09:29:41.7218154Z ++ [[ -z ossci-compiler-cache-circleci-v2 ]] 2025-12-04T09:29:41.7218491Z ++ sccache --stop-server 2025-12-04T09:29:41.7250749Z ++ true 2025-12-04T09:29:41.7250981Z ++ rm -f /var/lib/jenkins/sccache_error.log 2025-12-04T09:29:41.7263308Z ++ trap_add sccache_epilogue EXIT 2025-12-04T09:29:41.7263667Z ++ trap_add_cmd=sccache_epilogue 2025-12-04T09:29:41.7263946Z ++ shift 2025-12-04T09:29:41.7264112Z ++ for trap_add_name in "$@" 2025-12-04T09:29:41.7269184Z ++++ trap -p EXIT 2025-12-04T09:29:41.7272095Z +++ eval 'extract_trap_cmd ' 2025-12-04T09:29:41.7272318Z ++++ extract_trap_cmd 2025-12-04T09:29:41.7272504Z ++++ printf '%s\n' '' 2025-12-04T09:29:41.7272877Z +++ printf '%s\n' sccache_epilogue 2025-12-04T09:29:41.7274957Z ++ trap -- ' 2025-12-04T09:29:41.7275179Z sccache_epilogue' EXIT 2025-12-04T09:29:41.7275385Z ++ [[ -n 1 ]] 2025-12-04T09:29:41.7275665Z ++ echo 'Skipping sccache server initialization, setting environment variables' 2025-12-04T09:29:41.7276350Z Skipping sccache server initialization, setting environment variables 2025-12-04T09:29:41.7276684Z ++ export SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:29:41.7276889Z ++ SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:29:41.7277148Z ++ export SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:29:41.7277467Z ++ SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:29:41.7282708Z ++ export RUST_LOG=sccache::server=error 2025-12-04T09:29:41.7282983Z ++ RUST_LOG=sccache::server=error 2025-12-04T09:29:41.7283209Z ++ sccache --zero-stats 2025-12-04T09:29:42.1628600Z Statistics zeroed. 2025-12-04T09:29:42.1635069Z ++ which ccache 2025-12-04T09:29:42.1769165Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *rocm* ]] 2025-12-04T09:29:42.1769707Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *s390x* ]] 2025-12-04T09:29:42.1770065Z + [[ -d /var/lib/jenkins/workspace ]] 2025-12-04T09:29:42.1772898Z ++ stat -c %u /var/lib/jenkins/workspace 2025-12-04T09:29:42.1790031Z + WORKSPACE_ORIGINAL_OWNER_ID=1000 2025-12-04T09:29:42.1790482Z + trap_add cleanup_workspace EXIT 2025-12-04T09:29:42.1790765Z + trap_add_cmd=cleanup_workspace 2025-12-04T09:29:42.1791001Z + shift 2025-12-04T09:29:42.1791192Z + for trap_add_name in "$@" 2025-12-04T09:29:42.1797258Z +++ trap -p EXIT 2025-12-04T09:29:42.1800458Z ++ eval 'extract_trap_cmd trap -- '\'' 2025-12-04T09:29:42.1800905Z sccache_epilogue'\'' EXIT' 2025-12-04T09:29:42.1801222Z +++ extract_trap_cmd trap -- ' 2025-12-04T09:29:42.1801443Z sccache_epilogue' EXIT 2025-12-04T09:29:42.1801641Z +++ printf '%s\n' ' 2025-12-04T09:29:42.1801887Z sccache_epilogue' 2025-12-04T09:29:42.1802070Z ++ printf '%s\n' cleanup_workspace 2025-12-04T09:29:42.1803572Z + trap -- ' 2025-12-04T09:29:42.1803740Z sccache_epilogue 2025-12-04T09:29:42.1803924Z cleanup_workspace' EXIT 2025-12-04T09:29:42.1814761Z + sudo chown -R jenkins /var/lib/jenkins/workspace 2025-12-04T09:29:43.1560453Z + git config --global --add safe.directory /var/lib/jenkins/workspace 2025-12-04T09:29:43.1583266Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *cuda* ]] 2025-12-04T09:29:43.1587412Z ++ python -c 'import os;import numba.cuda; print(os.path.dirname(numba.cuda.__file__))' 2025-12-04T09:29:43.5973354Z + NUMBA_CUDA_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda 2025-12-04T09:29:43.5973974Z + '[' -n /opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda ']' 2025-12-04T09:29:43.5979800Z +++ realpath .ci/pytorch/test.sh 2025-12-04T09:29:43.5990989Z ++ dirname /var/lib/jenkins/workspace/.ci/pytorch/test.sh 2025-12-04T09:29:43.6168830Z + NUMBA_PATCH=/var/lib/jenkins/workspace/.ci/pytorch/numba-cuda-13.patch 2025-12-04T09:29:43.6169433Z + pushd /opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda 2025-12-04T09:29:43.6169997Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda ~/workspace 2025-12-04T09:29:43.6170428Z + patch -p4 2025-12-04T09:29:43.6186485Z patching file cudadrv/driver.py 2025-12-04T09:29:43.6190253Z + popd 2025-12-04T09:29:43.6190573Z ~/workspace 2025-12-04T09:29:43.6191021Z + echo 'Environment variables:' 2025-12-04T09:29:43.6191342Z Environment variables: 2025-12-04T09:29:43.6191558Z + env 2025-12-04T09:29:43.6202159Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:29:43.6202644Z CONTINUE_THROUGH_ERROR=True 2025-12-04T09:29:43.6203051Z BUILD_ENVIRONMENT=linux-jammy-cuda12.8-py3.12-gcc11 2025-12-04T09:29:43.6203817Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-12-04T09:29:43.6204084Z HOSTNAME=dc9d197ee609 2025-12-04T09:29:43.6204771Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_045a3d82-8263-42b1-9b16-ddf4dd82f36c 2025-12-04T09:29:43.6205537Z GITHUB_ACTION=__run_3 2025-12-04T09:29:43.6205855Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-12-04T09:29:43.6206223Z GITHUB_RUN_NUMBER=4606 2025-12-04T09:29:43.6206553Z TEST_CONFIG=vllm_basic_models_test 2025-12-04T09:29:43.6206782Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-12-04T09:29:43.6207008Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-12-04T09:29:43.6207491Z SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:29:43.6207815Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-12-04T09:29:43.6208042Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-12-04T09:29:43.6208266Z GITHUB_REF_TYPE=branch 2025-12-04T09:29:43.6208479Z BASE_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:29:43.6208859Z XLA_CUDA= 2025-12-04T09:29:43.6209017Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-12-04T09:29:43.6209310Z HUGGING_FACE_HUB_TOKEN=*** 2025-12-04T09:29:43.6209719Z *** 2025-12-04T09:29:43.6209875Z GITHUB_REPOSITORY_ID=65600975 2025-12-04T09:29:43.6210081Z GITHUB_ACTIONS=true 2025-12-04T09:29:43.6210260Z NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:29:43.6210498Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:29:43.6210805Z SHA1=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:29:43.6211065Z GITHUB_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:29:43.6211425Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/vllm.yml@refs/heads/main 2025-12-04T09:29:43.6211755Z UCC_HOME=/usr 2025-12-04T09:29:43.6211907Z VERBOSE_TEST_LOGS=False 2025-12-04T09:29:43.6212093Z GITHUB_REF=refs/heads/main 2025-12-04T09:29:43.6212275Z SHARD_NUMBER=1 2025-12-04T09:29:43.6212433Z GITHUB_REF_PROTECTED=true 2025-12-04T09:29:43.6212616Z HOME=/var/lib/jenkins 2025-12-04T09:29:43.6212814Z GITHUB_API_URL=https://api.github.com 2025-12-04T09:29:43.6213046Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-12-04T09:29:43.6213303Z UCX_COMMIT=7836b165abdbe468a2f607e7254011c07d788152 2025-12-04T09:29:43.6213538Z USE_SYSTEM_NCCL=1 2025-12-04T09:29:43.6213707Z NUM_TEST_SHARDS=1 2025-12-04T09:29:43.6213917Z UCX_HOME=/usr 2025-12-04T09:29:43.6214392Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_045a3d82-8263-42b1-9b16-ddf4dd82f36c 2025-12-04T09:29:43.6215033Z JOB_NAME=ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu, unstable) 2025-12-04T09:29:43.6215642Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_045a3d82-8263-42b1-9b16-ddf4dd82f36c 2025-12-04T09:29:43.6216217Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-12-04T09:29:43.6216569Z GITHUB_EVENT_NAME=schedule 2025-12-04T09:29:43.6216755Z DASHBOARD_TAG= 2025-12-04T09:29:43.6216914Z GITHUB_RUN_ID=19922031852 2025-12-04T09:29:43.6217097Z INSTALLED_OPENBLAS= 2025-12-04T09:29:43.6217526Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_045a3d82-8263-42b1-9b16-ddf4dd82f36c 2025-12-04T09:29:43.6218200Z GITHUB_ACTOR=pytorchmergebot 2025-12-04T09:29:43.6218414Z PR_NUMBER= 2025-12-04T09:29:43.6218572Z DESIRED_CUDA=12.8.1 2025-12-04T09:29:43.6218744Z GITHUB_RUN_ATTEMPT=1 2025-12-04T09:29:43.6218937Z ANACONDA_PYTHON_VERSION=3.12 2025-12-04T09:29:43.6219311Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-12-04T09:29:43.6219558Z TERM=vt100 2025-12-04T09:29:43.6219713Z INSTALLED_VISION=yes 2025-12-04T09:29:43.6219877Z BRANCH=main 2025-12-04T09:29:43.6220034Z SCCACHE_REGION=us-east-1 2025-12-04T09:29:43.6220226Z OPENSSL_ROOT_DIR=/opt/openssl 2025-12-04T09:29:43.6220422Z BUILD_AOT_INDUCTOR_TEST= 2025-12-04T09:29:43.6220595Z CUDA_PATH=/usr/local/cuda 2025-12-04T09:29:43.6220970Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-12-04T09:29:43.6221381Z GITHUB_SERVER_URL=https://github.com 2025-12-04T09:29:43.6221630Z UCC_COMMIT=430e241bf5d38cbc73fc7a6b89155397232e3f96 2025-12-04T09:29:43.6221863Z REENABLED_ISSUES= 2025-12-04T09:29:43.6222020Z DOCS= 2025-12-04T09:29:43.6222156Z SHLVL=1 2025-12-04T09:29:43.6222288Z MAX_JOBS=14 2025-12-04T09:29:43.6222443Z GITHUB_ACTOR_ID=97764156 2025-12-04T09:29:43.6222694Z GITHUB_WORKFLOW_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:29:43.6222961Z GITHUB_REF_NAME=main 2025-12-04T09:29:43.6223235Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-12-04T09:29:43.6223532Z GITHUB_JOB=test 2025-12-04T09:29:43.6223688Z NO_TEST_TIMEOUT=False 2025-12-04T09:29:43.6223973Z TD_DISTRIBUTED=False 2025-12-04T09:29:43.6224157Z GITHUB_REPOSITORY=pytorch/pytorch 2025-12-04T09:29:43.6224360Z GITHUB_RETENTION_DAYS=90 2025-12-04T09:29:43.6224541Z OPENSSL_DIR=/opt/openssl 2025-12-04T09:29:43.6224727Z GITHUB_ACTION_REPOSITORY= 2025-12-04T09:29:43.6225295Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:29:43.6225850Z GITHUB_BASE_REF= 2025-12-04T09:29:43.6226016Z INSTALLED_ACL= 2025-12-04T09:29:43.6226373Z ARTIFACTS_FILE_SUFFIX=test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_57118543727 2025-12-04T09:29:43.6226766Z CI=true 2025-12-04T09:29:43.6226923Z GITHUB_REPOSITORY_OWNER=pytorch 2025-12-04T09:29:43.6227159Z RUST_LOG=sccache::server=error 2025-12-04T09:29:43.6227344Z JOB_ID=57118543727 2025-12-04T09:29:43.6227502Z GITHUB_HEAD_REF= 2025-12-04T09:29:43.6227656Z GITHUB_ACTION_REF= 2025-12-04T09:29:43.6227855Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-12-04T09:29:43.6228104Z TEST_SHOWLOCALS=False 2025-12-04T09:29:43.6228287Z GITHUB_WORKFLOW=vllm-test 2025-12-04T09:29:43.6228479Z DEBIAN_FRONTEND=noninteractive 2025-12-04T09:29:43.6229039Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_045a3d82-8263-42b1-9b16-ddf4dd82f36c 2025-12-04T09:29:43.6229491Z NO_TD=False 2025-12-04T09:29:43.6229658Z SKIP_SCCACHE_INITIALIZATION=1 2025-12-04T09:29:43.6229867Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-12-04T09:29:43.6230091Z _=/usr/bin/env 2025-12-04T09:29:43.6230345Z OLDPWD=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda 2025-12-04T09:29:43.6230704Z ++ python -c 'import site; print(site.getsitepackages()[0])' 2025-12-04T09:29:43.6330167Z + TORCH_INSTALL_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch 2025-12-04T09:29:43.6331125Z + TORCH_BIN_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/bin 2025-12-04T09:29:43.6332062Z + TORCH_LIB_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/lib 2025-12-04T09:29:43.6332834Z + TORCH_TEST_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/test 2025-12-04T09:29:43.6333275Z + BUILD_DIR=build 2025-12-04T09:29:43.6333511Z + BUILD_RENAMED_DIR=build_renamed 2025-12-04T09:29:43.6333847Z + BUILD_BIN_DIR=build/bin 2025-12-04T09:29:43.6334162Z + SHARD_NUMBER=1 2025-12-04T09:29:43.6334428Z + NUM_TEST_SHARDS=1 2025-12-04T09:29:43.6334718Z + export TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:29:43.6335327Z + TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:29:43.6335672Z + export VALGRIND=ON 2025-12-04T09:29:43.6335951Z + VALGRIND=ON 2025-12-04T09:29:43.6336292Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *clang9* ]] 2025-12-04T09:29:43.6336664Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *xpu* ]] 2025-12-04T09:29:43.6336925Z + detect_cuda_arch 2025-12-04T09:29:43.6337127Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *cuda* ]] 2025-12-04T09:29:43.6337376Z + command -v nvidia-smi 2025-12-04T09:29:43.6337572Z /usr/bin/nvidia-smi 2025-12-04T09:29:43.6342013Z ++ nvidia-smi --query-gpu=compute_cap --format=csv 2025-12-04T09:29:43.6343417Z ++ tail -n 1 2025-12-04T09:29:43.6576196Z + TORCH_CUDA_ARCH_LIST=8.9 2025-12-04T09:29:43.6576551Z + export TORCH_CUDA_ARCH_LIST 2025-12-04T09:29:43.6576958Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *s390x* ]] 2025-12-04T09:29:43.6577299Z + [[ 0 == \1 ]] 2025-12-04T09:29:43.6577460Z + [[ True == \1 ]] 2025-12-04T09:29:43.6577670Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *bazel* ]] 2025-12-04T09:29:43.6580424Z ++ realpath build/custom_test_artifacts 2025-12-04T09:29:43.6759105Z + CUSTOM_TEST_ARTIFACT_BUILD_DIR=/var/lib/jenkins/workspace/build/custom_test_artifacts 2025-12-04T09:29:43.6759750Z + [[ -n '' ]] 2025-12-04T09:29:43.6759975Z + echo 'Environment variables' 2025-12-04T09:29:43.6760239Z Environment variables 2025-12-04T09:29:43.6760456Z + env 2025-12-04T09:29:43.6986415Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:29:43.6987570Z CONTINUE_THROUGH_ERROR=True 2025-12-04T09:29:43.6987985Z BUILD_ENVIRONMENT=linux-jammy-cuda12.8-py3.12-gcc11 2025-12-04T09:29:43.6988718Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-12-04T09:29:43.6989072Z HOSTNAME=dc9d197ee609 2025-12-04T09:29:43.6989608Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_045a3d82-8263-42b1-9b16-ddf4dd82f36c 2025-12-04T09:29:43.6990342Z GITHUB_ACTION=__run_3 2025-12-04T09:29:43.6990720Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-12-04T09:29:43.6990990Z GITHUB_RUN_NUMBER=4606 2025-12-04T09:29:43.6991229Z TEST_CONFIG=vllm_basic_models_test 2025-12-04T09:29:43.6991497Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-12-04T09:29:43.6991791Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-12-04T09:29:43.6992067Z SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:29:43.6992421Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-12-04T09:29:43.6992700Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-12-04T09:29:43.6992964Z GITHUB_REF_TYPE=branch 2025-12-04T09:29:43.6993146Z TORCH_CUDA_ARCH_LIST=8.9 2025-12-04T09:29:43.6993372Z BASE_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:29:43.6993606Z XLA_CUDA= 2025-12-04T09:29:43.6993769Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-12-04T09:29:43.6994141Z HUGGING_FACE_HUB_TOKEN=*** 2025-12-04T09:29:43.6994371Z *** 2025-12-04T09:29:43.6994532Z GITHUB_REPOSITORY_ID=65600975 2025-12-04T09:29:43.6994735Z GITHUB_ACTIONS=true 2025-12-04T09:29:43.6994906Z NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:29:43.6995137Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:29:43.6995462Z SHA1=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:29:43.6995801Z GITHUB_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:29:43.6996435Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/vllm.yml@refs/heads/main 2025-12-04T09:29:43.6996912Z UCC_HOME=/usr 2025-12-04T09:29:43.6997138Z TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:29:43.6997472Z VERBOSE_TEST_LOGS=False 2025-12-04T09:29:43.6997767Z GITHUB_REF=refs/heads/main 2025-12-04T09:29:43.6998034Z SHARD_NUMBER=1 2025-12-04T09:29:43.6998194Z GITHUB_REF_PROTECTED=true 2025-12-04T09:29:43.6998377Z HOME=/var/lib/jenkins 2025-12-04T09:29:43.6998576Z GITHUB_API_URL=https://api.github.com 2025-12-04T09:29:43.6998814Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-12-04T09:29:43.6999056Z UCX_COMMIT=7836b165abdbe468a2f607e7254011c07d788152 2025-12-04T09:29:43.6999297Z USE_SYSTEM_NCCL=1 2025-12-04T09:29:43.6999456Z NUM_TEST_SHARDS=1 2025-12-04T09:29:43.6999615Z UCX_HOME=/usr 2025-12-04T09:29:43.7000198Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_045a3d82-8263-42b1-9b16-ddf4dd82f36c 2025-12-04T09:29:43.7000850Z JOB_NAME=ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu, unstable) 2025-12-04T09:29:43.7001467Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_045a3d82-8263-42b1-9b16-ddf4dd82f36c 2025-12-04T09:29:43.7002031Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-12-04T09:29:43.7002390Z GITHUB_EVENT_NAME=schedule 2025-12-04T09:29:43.7002574Z DASHBOARD_TAG= 2025-12-04T09:29:43.7002736Z GITHUB_RUN_ID=19922031852 2025-12-04T09:29:43.7002917Z INSTALLED_OPENBLAS= 2025-12-04T09:29:43.7003348Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_045a3d82-8263-42b1-9b16-ddf4dd82f36c 2025-12-04T09:29:43.7003819Z GITHUB_ACTOR=pytorchmergebot 2025-12-04T09:29:43.7004013Z PR_NUMBER= 2025-12-04T09:29:43.7004164Z DESIRED_CUDA=12.8.1 2025-12-04T09:29:43.7004326Z GITHUB_RUN_ATTEMPT=1 2025-12-04T09:29:43.7004495Z VALGRIND=ON 2025-12-04T09:29:43.7004655Z ANACONDA_PYTHON_VERSION=3.12 2025-12-04T09:29:43.7004896Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-12-04T09:29:43.7005154Z TERM=vt100 2025-12-04T09:29:43.7005306Z INSTALLED_VISION=yes 2025-12-04T09:29:43.7005470Z BRANCH=main 2025-12-04T09:29:43.7005618Z SCCACHE_REGION=us-east-1 2025-12-04T09:29:43.7005808Z OPENSSL_ROOT_DIR=/opt/openssl 2025-12-04T09:29:43.7006106Z BUILD_AOT_INDUCTOR_TEST= 2025-12-04T09:29:43.7006282Z CUDA_PATH=/usr/local/cuda 2025-12-04T09:29:43.7006653Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-12-04T09:29:43.7007064Z GITHUB_SERVER_URL=https://github.com 2025-12-04T09:29:43.7007323Z UCC_COMMIT=430e241bf5d38cbc73fc7a6b89155397232e3f96 2025-12-04T09:29:43.7007566Z REENABLED_ISSUES= 2025-12-04T09:29:43.7007720Z DOCS= 2025-12-04T09:29:43.7007854Z SHLVL=1 2025-12-04T09:29:43.7007992Z MAX_JOBS=14 2025-12-04T09:29:43.7008157Z GITHUB_ACTOR_ID=97764156 2025-12-04T09:29:43.7008397Z GITHUB_WORKFLOW_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:29:43.7008670Z GITHUB_REF_NAME=main 2025-12-04T09:29:43.7008936Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-12-04T09:29:43.7009239Z GITHUB_JOB=test 2025-12-04T09:29:43.7009397Z NO_TEST_TIMEOUT=False 2025-12-04T09:29:43.7009568Z TD_DISTRIBUTED=False 2025-12-04T09:29:43.7009755Z GITHUB_REPOSITORY=pytorch/pytorch 2025-12-04T09:29:43.7009964Z GITHUB_RETENTION_DAYS=90 2025-12-04T09:29:43.7010151Z OPENSSL_DIR=/opt/openssl 2025-12-04T09:29:43.7010338Z GITHUB_ACTION_REPOSITORY= 2025-12-04T09:29:43.7010877Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:29:43.7011552Z GITHUB_BASE_REF= 2025-12-04T09:29:43.7011749Z INSTALLED_ACL= 2025-12-04T09:29:43.7012113Z ARTIFACTS_FILE_SUFFIX=test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_57118543727 2025-12-04T09:29:43.7012516Z CI=true 2025-12-04T09:29:43.7012684Z GITHUB_REPOSITORY_OWNER=pytorch 2025-12-04T09:29:43.7012927Z RUST_LOG=sccache::server=error 2025-12-04T09:29:43.7013119Z JOB_ID=57118543727 2025-12-04T09:29:43.7013279Z GITHUB_HEAD_REF= 2025-12-04T09:29:43.7013442Z GITHUB_ACTION_REF= 2025-12-04T09:29:43.7013647Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-12-04T09:29:43.7013900Z TEST_SHOWLOCALS=False 2025-12-04T09:29:43.7014098Z GITHUB_WORKFLOW=vllm-test 2025-12-04T09:29:43.7014289Z DEBIAN_FRONTEND=noninteractive 2025-12-04T09:29:43.7014727Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_045a3d82-8263-42b1-9b16-ddf4dd82f36c 2025-12-04T09:29:43.7015174Z NO_TD=False 2025-12-04T09:29:43.7015337Z SKIP_SCCACHE_INITIALIZATION=1 2025-12-04T09:29:43.7015550Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-12-04T09:29:43.7015865Z OLDPWD=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda 2025-12-04T09:29:43.7016266Z _=/usr/bin/env 2025-12-04T09:29:43.7016435Z + echo 'Testing pytorch' 2025-12-04T09:29:43.7016614Z Testing pytorch 2025-12-04T09:29:43.7016776Z + export LANG=C.UTF-8 2025-12-04T09:29:43.7016937Z + LANG=C.UTF-8 2025-12-04T09:29:43.7017097Z + PR_NUMBER= 2025-12-04T09:29:43.7017269Z + [[ vllm_basic_models_test == \d\e\f\a\u\l\t ]] 2025-12-04T09:29:43.7017542Z + [[ vllm_basic_models_test == \d\i\s\t\r\i\b\u\t\e\d ]] 2025-12-04T09:29:43.7017797Z + [[ vllm_basic_models_test == \s\l\o\w ]] 2025-12-04T09:29:43.7018077Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *slow-gradcheck* ]] 2025-12-04T09:29:43.7018372Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *cuda* ]] 2025-12-04T09:29:43.7018636Z + export PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-12-04T09:29:43.7018879Z + PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-12-04T09:29:43.7019109Z + [[ vllm_basic_models_test == *crossref* ]] 2025-12-04T09:29:43.7019474Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *rocm* ]] 2025-12-04T09:29:43.7019751Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *xpu* ]] 2025-12-04T09:29:43.7020028Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *-bazel-* ]] 2025-12-04T09:29:43.7020269Z + pip_install ninja==1.10.2 2025-12-04T09:29:43.7020527Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-12-04T09:29:43.7020854Z + python3 -m pip install --progress-bar off ninja==1.10.2 2025-12-04T09:29:44.2899786Z Collecting ninja==1.10.2 2025-12-04T09:29:44.3122825Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl.metadata (5.0 kB) 2025-12-04T09:29:44.3425510Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl (108 kB) 2025-12-04T09:29:44.5443570Z Installing collected packages: ninja 2025-12-04T09:29:44.5444069Z Attempting uninstall: ninja 2025-12-04T09:29:44.5470196Z Found existing installation: ninja 1.11.1.4 2025-12-04T09:29:44.5492191Z Uninstalling ninja-1.11.1.4: 2025-12-04T09:29:44.5628020Z Successfully uninstalled ninja-1.11.1.4 2025-12-04T09:29:44.6048266Z Successfully installed ninja-1.10.2 2025-12-04T09:29:44.6459939Z + export PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:29:44.6461497Z + PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:29:44.6462382Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *aarch64* ]] 2025-12-04T09:29:44.6462791Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *asan* ]] 2025-12-04T09:29:44.6463266Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *-debug* ]] 2025-12-04T09:29:44.6463645Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *-bazel-* ]] 2025-12-04T09:29:44.6464246Z + echo 'We are not in debug mode: linux-jammy-cuda12.8-py3.12-gcc11. Expect the assertion to pass' 2025-12-04T09:29:44.6464767Z We are not in debug mode: linux-jammy-cuda12.8-py3.12-gcc11. Expect the assertion to pass 2025-12-04T09:29:44.6465240Z + cd test 2025-12-04T09:29:44.6465484Z + python -c 'import torch; torch._C._crash_if_debug_asserts_fail(424242)' 2025-12-04T09:29:46.0998510Z + [[ vllm_basic_models_test == \n\o\g\p\u\_\N\O\_\A\V\X\2 ]] 2025-12-04T09:29:46.0998955Z + [[ vllm_basic_models_test == \n\o\g\p\u\_\A\V\X\5\1\2 ]] 2025-12-04T09:29:46.0999393Z + [[ vllm_basic_models_test == \l\e\g\a\c\y\_\n\v\i\d\i\a\_\d\r\i\v\e\r ]] 2025-12-04T09:29:46.1003459Z + DYNAMO_BENCHMARK_FLAGS=() 2025-12-04T09:29:46.1003906Z + [[ vllm_basic_models_test == *pr_time_benchmarks* ]] 2025-12-04T09:29:46.1004399Z + [[ vllm_basic_models_test == *dynamo_eager* ]] 2025-12-04T09:29:46.1004729Z + [[ vllm_basic_models_test == *aot_eager* ]] 2025-12-04T09:29:46.1005197Z + [[ vllm_basic_models_test == *aot_inductor* ]] 2025-12-04T09:29:46.1005546Z + [[ vllm_basic_models_test == *max_autotune_inductor* ]] 2025-12-04T09:29:46.1005888Z + [[ vllm_basic_models_test == *inductor* ]] 2025-12-04T09:29:46.1006553Z + [[ vllm_basic_models_test == *dynamic* ]] 2025-12-04T09:29:46.1006874Z + [[ vllm_basic_models_test == *cpu* ]] 2025-12-04T09:29:46.1007164Z + [[ vllm_basic_models_test == *xpu* ]] 2025-12-04T09:29:46.1007471Z + DYNAMO_BENCHMARK_FLAGS+=(--device cuda) 2025-12-04T09:29:46.1061462Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *libtorch* ]] 2025-12-04T09:29:46.1061853Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *-bazel-* ]] 2025-12-04T09:29:46.1064968Z + cd test 2025-12-04T09:29:46.1065769Z + python -c 'import torch; print(torch.__config__.show())' 2025-12-04T09:29:47.5807715Z PyTorch built with: 2025-12-04T09:29:47.5808009Z - GCC 11.4 2025-12-04T09:29:47.5808218Z - C++ Version: 201703 2025-12-04T09:29:47.5808729Z - Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-12-04T09:29:47.5809420Z - Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-12-04T09:29:47.5809823Z - OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-12-04T09:29:47.5810168Z - LAPACK is enabled (usually provided by MKL) 2025-12-04T09:29:47.5810463Z - NNPACK is enabled 2025-12-04T09:29:47.5810699Z - CPU capability usage: AVX2 2025-12-04T09:29:47.5810944Z - CUDA Runtime 12.8 2025-12-04T09:29:47.5811493Z - NVCC architecture flags: -gencode;arch=compute_80,code=sm_80;-gencode;arch=compute_89,code=sm_89;-gencode;arch=compute_90,code=sm_90 2025-12-04T09:29:47.5812089Z - CuDNN 91.0.2 (built against CUDA 12.9) 2025-12-04T09:29:47.5816684Z - Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, COMMIT_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32, CUDA_VERSION=12.8, CUDNN_VERSION=9.10.2, CXX_COMPILER=/opt/cache/bin/c++, CXX_FLAGS= -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOROCTRACER -DLIBKINETO_NOXPUPTI=ON -DUSE_FBGEMM -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -DC10_NODEPRECATED -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=range-loop-construct -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-unknown-pragmas -Wno-unused-parameter -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=old-style-cast -faligned-new -Werror -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, FORCE_FALLBACK_CUDA_MPI=1, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, TORCH_VERSION=2.10.0, USE_CUDA=ON, USE_CUDNN=ON, USE_CUSPARSELT=ON, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_GLOO=ON, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=ON, USE_NCCL=ON, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, USE_ROCM_KERNEL_ASSERT=OFF, USE_XCCL=OFF, USE_XPU=OFF, 2025-12-04T09:29:47.5820375Z 2025-12-04T09:29:47.8529042Z + cd test 2025-12-04T09:29:47.8529636Z + python -c 'import torch; print(torch.__config__.parallel_info())' 2025-12-04T09:29:49.0434742Z ATen/Parallel: 2025-12-04T09:29:49.0435218Z at::get_num_threads() : 8 2025-12-04T09:29:49.0435660Z at::get_num_interop_threads() : 8 2025-12-04T09:29:49.0435987Z OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-12-04T09:29:49.0436252Z omp_get_max_threads() : 8 2025-12-04T09:29:49.0436783Z Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-12-04T09:29:49.0437313Z mkl_get_max_threads() : 8 2025-12-04T09:29:49.0437681Z Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-12-04T09:29:49.0438078Z std::thread::hardware_concurrency() : 16 2025-12-04T09:29:49.0438383Z Environment variables: 2025-12-04T09:29:49.0438614Z OMP_NUM_THREADS : [not set] 2025-12-04T09:29:49.0438850Z MKL_NUM_THREADS : [not set] 2025-12-04T09:29:49.0439084Z ATen parallel backend: OpenMP 2025-12-04T09:29:49.0439248Z 2025-12-04T09:29:49.2741943Z + [[ vllm_basic_models_test == *numpy_2* ]] 2025-12-04T09:29:49.2743063Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *aarch64* ]] 2025-12-04T09:29:49.2743641Z + [[ vllm_basic_models_test == *backward* ]] 2025-12-04T09:29:49.2744438Z + [[ vllm_basic_models_test == *libtorch_agnostic_targetting* ]] 2025-12-04T09:29:49.2744829Z + [[ vllm_basic_models_test == *xla* ]] 2025-12-04T09:29:49.2745112Z + [[ vllm_basic_models_test == *vllm* ]] 2025-12-04T09:29:49.2745634Z + echo 'vLLM CI uses TORCH_CUDA_ARCH_LIST: 8.9' 2025-12-04T09:29:49.2746173Z vLLM CI uses TORCH_CUDA_ARCH_LIST: 8.9 2025-12-04T09:29:49.2746638Z + cd .ci/lumen_cli 2025-12-04T09:29:49.2747014Z + python -m pip install -e . 2025-12-04T09:29:49.5536589Z Obtaining file:///var/lib/jenkins/workspace/.ci/lumen_cli 2025-12-04T09:29:50.2672130Z Installing build dependencies ... [?25l- \ done 2025-12-04T09:29:50.4761751Z [?25h Checking if build backend supports build_editable ... [?25ldone 2025-12-04T09:29:50.7321977Z [?25h Getting requirements to build editable ... [?25l- done 2025-12-04T09:29:50.8939045Z [?25h Preparing editable metadata (pyproject.toml) ... [?25l- done 2025-12-04T09:29:50.9553775Z [?25hCollecting pyyaml==6.0.2 (from lumen-ci==0.1.0) 2025-12-04T09:29:50.9760152Z Downloading PyYAML-6.0.2-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB) 2025-12-04T09:29:51.0057567Z Collecting GitPython==3.1.45 (from lumen-ci==0.1.0) 2025-12-04T09:29:51.0084364Z Downloading gitpython-3.1.45-py3-none-any.whl.metadata (13 kB) 2025-12-04T09:29:51.0344299Z Collecting docker==7.1.0 (from lumen-ci==0.1.0) 2025-12-04T09:29:51.0374071Z Downloading docker-7.1.0-py3-none-any.whl.metadata (3.8 kB) 2025-12-04T09:29:51.0451015Z Requirement already satisfied: pytest==7.3.2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from lumen-ci==0.1.0) (7.3.2) 2025-12-04T09:29:51.2345862Z Collecting uv==0.9.6 (from lumen-ci==0.1.0) 2025-12-04T09:29:51.2388356Z Downloading uv-0.9.6-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (11 kB) 2025-12-04T09:29:51.2467327Z Requirement already satisfied: requests>=2.26.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from docker==7.1.0->lumen-ci==0.1.0) (2.32.5) 2025-12-04T09:29:51.2470255Z Requirement already satisfied: urllib3>=1.26.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from docker==7.1.0->lumen-ci==0.1.0) (2.5.0) 2025-12-04T09:29:51.2594089Z Collecting gitdb<5,>=4.0.1 (from GitPython==3.1.45->lumen-ci==0.1.0) 2025-12-04T09:29:51.2622158Z Downloading gitdb-4.0.12-py3-none-any.whl.metadata (1.2 kB) 2025-12-04T09:29:51.2694239Z Requirement already satisfied: iniconfig in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from pytest==7.3.2->lumen-ci==0.1.0) (2.3.0) 2025-12-04T09:29:51.2696982Z Requirement already satisfied: packaging in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from pytest==7.3.2->lumen-ci==0.1.0) (23.1) 2025-12-04T09:29:51.2701961Z Requirement already satisfied: pluggy<2.0,>=0.12 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from pytest==7.3.2->lumen-ci==0.1.0) (1.6.0) 2025-12-04T09:29:51.2821731Z Collecting smmap<6,>=3.0.1 (from gitdb<5,>=4.0.1->GitPython==3.1.45->lumen-ci==0.1.0) 2025-12-04T09:29:51.2847281Z Downloading smmap-5.0.2-py3-none-any.whl.metadata (4.3 kB) 2025-12-04T09:29:51.2934756Z Requirement already satisfied: charset_normalizer<4,>=2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->docker==7.1.0->lumen-ci==0.1.0) (3.4.4) 2025-12-04T09:29:51.2938744Z Requirement already satisfied: idna<4,>=2.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->docker==7.1.0->lumen-ci==0.1.0) (3.11) 2025-12-04T09:29:51.2944325Z Requirement already satisfied: certifi>=2017.4.17 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->docker==7.1.0->lumen-ci==0.1.0) (2025.11.12) 2025-12-04T09:29:51.3018235Z Downloading docker-7.1.0-py3-none-any.whl (147 kB) 2025-12-04T09:29:51.3182831Z Downloading gitpython-3.1.45-py3-none-any.whl (208 kB) 2025-12-04T09:29:51.3292379Z Downloading PyYAML-6.0.2-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (767 kB) 2025-12-04T09:29:51.3437163Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/767.5 kB ? eta -:--:-- 2025-12-04T09:29:51.3437904Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 767.5/767.5 kB 45.5 MB/s 0:00:00 2025-12-04T09:29:51.3464247Z [?25hDownloading uv-0.9.6-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (21.4 MB) 2025-12-04T09:29:51.5067330Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/21.4 MB ? eta -:--:-- 2025-12-04T09:29:51.5068183Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 21.4/21.4 MB 134.2 MB/s 0:00:00 2025-12-04T09:29:51.5096470Z [?25hDownloading gitdb-4.0.12-py3-none-any.whl (62 kB) 2025-12-04T09:29:51.5185601Z Downloading smmap-5.0.2-py3-none-any.whl (24 kB) 2025-12-04T09:29:51.5453588Z Building wheels for collected packages: lumen-ci 2025-12-04T09:29:51.7236649Z Building editable for lumen-ci (pyproject.toml) ... [?25l- done 2025-12-04T09:29:51.7243257Z [?25h Created wheel for lumen-ci: filename=lumen_ci-0.1.0-0.editable-py3-none-any.whl size=2707 sha256=1e986321abb5c9aa660b5b9b2dde43e1b9d56a5b3d68f7158958dea582e72568 2025-12-04T09:29:51.7244649Z Stored in directory: /tmp/pip-ephem-wheel-cache-1vvv357s/wheels/6d/29/6c/82214c85262c5bdb1d615291039f4ec9f974034103e7a303fa 2025-12-04T09:29:51.7257643Z Successfully built lumen-ci 2025-12-04T09:29:51.8833856Z Installing collected packages: uv, smmap, pyyaml, gitdb, docker, GitPython, lumen-ci 2025-12-04T09:29:52.0515955Z [?25l 2025-12-04T09:29:52.1305970Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/7 [uv] 2025-12-04T09:29:52.1306675Z  Attempting uninstall: pyyaml 2025-12-04T09:29:52.1355886Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/7 [uv] 2025-12-04T09:29:52.1356586Z  Found existing installation: PyYAML 6.0.3 2025-12-04T09:29:52.1391485Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/7 [uv] 2025-12-04T09:29:52.1392109Z  Uninstalling PyYAML-6.0.3: 2025-12-04T09:29:52.1534271Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/7 [uv] 2025-12-04T09:29:52.1534947Z  Successfully uninstalled PyYAML-6.0.3 2025-12-04T09:29:52.2195895Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/7 [uv] 2025-12-04T09:29:52.3900767Z  ━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━ 3/7 [gitdb] 2025-12-04T09:29:52.4213368Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━ 5/7 [GitPython] 2025-12-04T09:29:52.4214288Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 7/7 [lumen-ci] 2025-12-04T09:29:52.4214656Z [?25h 2025-12-04T09:29:52.4304567Z Successfully installed GitPython-3.1.45 docker-7.1.0 gitdb-4.0.12 lumen-ci-0.1.0 pyyaml-6.0.2 smmap-5.0.2 uv-0.9.6 2025-12-04T09:29:52.5164702Z + python -m cli.run test external vllm --test-plan vllm_basic_models_test --shard-id 1 --num-shards 1 2025-12-04T09:29:52.7144870Z 2025-12-04 09:29:52,713 [INFO] cli.lib.core.vllm.vllm_test: Display VllmTestParameters VllmTestParameters(torch_whls_path=PosixPath('/var/lib/jenkins/workspace/dist'), vllm_whls_path=PosixPath('/var/lib/jenkins/workspace/dist/external/vllm/wheels'), torch_cuda_arch_list='8.9', cleaning_script=PosixPath('/var/lib/jenkins/workspace/.github/ci_configs/vllm/use_existing_torch.py')) 2025-12-04T09:29:52.7146547Z 2025-12-04 09:29:52,714 [INFO] cli.lib.common.git_helper: Cloning vllm to vllm 2025-12-04T09:29:52.8932526Z 2025-12-04 09:29:52,892 [INFO] cli.lib.common.git_helper: Progress: 10% - remote: Counting objects: 10% (7/64) 2025-12-04T09:29:52.8933408Z 2025-12-04 09:29:52,892 [INFO] cli.lib.common.git_helper: Progress: 15% - remote: Counting objects: 15% (10/64) 2025-12-04T09:29:52.8934089Z 2025-12-04 09:29:52,892 [INFO] cli.lib.common.git_helper: Progress: 20% - remote: Counting objects: 20% (13/64) 2025-12-04T09:29:52.8934797Z 2025-12-04 09:29:52,892 [INFO] cli.lib.common.git_helper: Progress: 25% - remote: Counting objects: 25% (16/64) 2025-12-04T09:29:52.8935546Z 2025-12-04 09:29:52,893 [INFO] cli.lib.common.git_helper: Progress: 35% - remote: Counting objects: 35% (23/64) 2025-12-04T09:29:52.8936069Z 2025-12-04 09:29:52,893 [INFO] cli.lib.common.git_helper: Progress: 40% - remote: Counting objects: 40% (26/64) 2025-12-04T09:29:52.8936633Z 2025-12-04 09:29:52,893 [INFO] cli.lib.common.git_helper: Progress: 45% - remote: Counting objects: 45% (29/64) 2025-12-04T09:29:52.8937276Z 2025-12-04 09:29:52,893 [INFO] cli.lib.common.git_helper: Progress: 50% - remote: Counting objects: 50% (32/64) 2025-12-04T09:29:52.8937872Z 2025-12-04 09:29:52,893 [INFO] cli.lib.common.git_helper: Progress: 60% - remote: Counting objects: 60% (39/64) 2025-12-04T09:29:52.8938390Z 2025-12-04 09:29:52,893 [INFO] cli.lib.common.git_helper: Progress: 65% - remote: Counting objects: 65% (42/64) 2025-12-04T09:29:52.8939029Z 2025-12-04 09:29:52,893 [INFO] cli.lib.common.git_helper: Progress: 70% - remote: Counting objects: 70% (45/64) 2025-12-04T09:29:52.8939672Z 2025-12-04 09:29:52,893 [INFO] cli.lib.common.git_helper: Progress: 75% - remote: Counting objects: 75% (48/64) 2025-12-04T09:29:52.8940180Z 2025-12-04 09:29:52,893 [INFO] cli.lib.common.git_helper: Progress: 85% - remote: Counting objects: 85% (55/64) 2025-12-04T09:29:52.8940770Z 2025-12-04 09:29:52,893 [INFO] cli.lib.common.git_helper: Progress: 90% - remote: Counting objects: 90% (58/64) 2025-12-04T09:29:52.8941302Z 2025-12-04 09:29:52,893 [INFO] cli.lib.common.git_helper: Progress: 95% - remote: Counting objects: 95% (61/64) 2025-12-04T09:29:52.8942018Z 2025-12-04 09:29:52,893 [INFO] cli.lib.common.git_helper: Progress: 100% - remote: Counting objects: 100% (64/64) 2025-12-04T09:29:52.8942559Z 2025-12-04 09:29:52,893 [INFO] cli.lib.common.git_helper: Progress: 5% - remote: Compressing objects: 5% (3/54) 2025-12-04T09:29:52.8963199Z 2025-12-04 09:29:52,896 [INFO] cli.lib.common.git_helper: Progress: 20% - remote: Compressing objects: 20% (11/54) 2025-12-04T09:29:52.8964614Z 2025-12-04 09:29:52,896 [INFO] cli.lib.common.git_helper: Progress: 25% - remote: Compressing objects: 25% (14/54) 2025-12-04T09:29:52.8971920Z 2025-12-04 09:29:52,896 [INFO] cli.lib.common.git_helper: Progress: 35% - remote: Compressing objects: 35% (19/54) 2025-12-04T09:29:52.8972533Z 2025-12-04 09:29:52,896 [INFO] cli.lib.common.git_helper: Progress: 40% - remote: Compressing objects: 40% (22/54) 2025-12-04T09:29:52.8973302Z 2025-12-04 09:29:52,897 [INFO] cli.lib.common.git_helper: Progress: 50% - remote: Compressing objects: 50% (27/54) 2025-12-04T09:29:52.8973886Z 2025-12-04 09:29:52,897 [INFO] cli.lib.common.git_helper: Progress: 55% - remote: Compressing objects: 55% (30/54) 2025-12-04T09:29:52.8975855Z 2025-12-04 09:29:52,897 [INFO] cli.lib.common.git_helper: Progress: 70% - remote: Compressing objects: 70% (38/54) 2025-12-04T09:29:52.8997327Z 2025-12-04 09:29:52,899 [INFO] cli.lib.common.git_helper: Progress: 75% - remote: Compressing objects: 75% (41/54) 2025-12-04T09:29:52.8998061Z 2025-12-04 09:29:52,899 [INFO] cli.lib.common.git_helper: Progress: 85% - remote: Compressing objects: 85% (46/54) 2025-12-04T09:29:52.8998724Z 2025-12-04 09:29:52,899 [INFO] cli.lib.common.git_helper: Progress: 90% - remote: Compressing objects: 90% (49/54) 2025-12-04T09:29:52.8999409Z 2025-12-04 09:29:52,899 [INFO] cli.lib.common.git_helper: Progress: 100% - remote: Compressing objects: 100% (54/54) 2025-12-04T09:29:52.9307654Z 2025-12-04 09:29:52,930 [INFO] cli.lib.common.git_helper: Progress: 0% - Receiving objects: 0% (1/147633) 2025-12-04T09:29:53.0345825Z 2025-12-04 09:29:53,034 [INFO] cli.lib.common.git_helper: Progress: 5% - Receiving objects: 5% (7382/147633) 2025-12-04T09:29:53.2980553Z 2025-12-04 09:29:53,297 [INFO] cli.lib.common.git_helper: Progress: 10% - Receiving objects: 10% (14764/147633) 2025-12-04T09:29:53.5270567Z 2025-12-04 09:29:53,526 [INFO] cli.lib.common.git_helper: Progress: 15% - Receiving objects: 15% (22145/147633), 28.63 MiB | 57.24 MiB/s 2025-12-04T09:29:53.7568685Z 2025-12-04 09:29:53,756 [INFO] cli.lib.common.git_helper: Progress: 20% - Receiving objects: 20% (29527/147633), 28.63 MiB | 57.24 MiB/s 2025-12-04T09:29:53.8371464Z 2025-12-04 09:29:53,836 [INFO] cli.lib.common.git_helper: Progress: 25% - Receiving objects: 25% (36909/147633), 28.63 MiB | 57.24 MiB/s 2025-12-04T09:29:53.9285848Z 2025-12-04 09:29:53,928 [INFO] cli.lib.common.git_helper: Progress: 30% - Receiving objects: 30% (44290/147633), 54.34 MiB | 54.38 MiB/s 2025-12-04T09:29:54.0074253Z 2025-12-04 09:29:54,007 [INFO] cli.lib.common.git_helper: Progress: 35% - Receiving objects: 35% (51672/147633), 54.34 MiB | 54.38 MiB/s 2025-12-04T09:29:54.0780148Z 2025-12-04 09:29:54,077 [INFO] cli.lib.common.git_helper: Progress: 40% - Receiving objects: 40% (59054/147633), 54.34 MiB | 54.38 MiB/s 2025-12-04T09:29:54.2406762Z 2025-12-04 09:29:54,240 [INFO] cli.lib.common.git_helper: Progress: 45% - Receiving objects: 45% (66435/147633), 54.34 MiB | 54.38 MiB/s 2025-12-04T09:29:54.3746139Z 2025-12-04 09:29:54,374 [INFO] cli.lib.common.git_helper: Progress: 50% - Receiving objects: 50% (73817/147633), 54.34 MiB | 54.38 MiB/s 2025-12-04T09:29:54.5134858Z 2025-12-04 09:29:54,513 [INFO] cli.lib.common.git_helper: Progress: 55% - Receiving objects: 55% (81199/147633), 83.47 MiB | 55.68 MiB/s 2025-12-04T09:29:54.6293721Z 2025-12-04 09:29:54,629 [INFO] cli.lib.common.git_helper: Progress: 60% - Receiving objects: 60% (88580/147633), 83.47 MiB | 55.68 MiB/s 2025-12-04T09:29:54.7184837Z 2025-12-04 09:29:54,718 [INFO] cli.lib.common.git_helper: Progress: 65% - Receiving objects: 65% (95962/147633), 83.47 MiB | 55.68 MiB/s 2025-12-04T09:29:54.7552590Z 2025-12-04 09:29:54,754 [INFO] cli.lib.common.git_helper: Progress: 70% - Receiving objects: 70% (103344/147633), 83.47 MiB | 55.68 MiB/s 2025-12-04T09:29:54.8031494Z 2025-12-04 09:29:54,802 [INFO] cli.lib.common.git_helper: Progress: 75% - Receiving objects: 75% (110725/147633), 83.47 MiB | 55.68 MiB/s 2025-12-04T09:29:54.8542523Z 2025-12-04 09:29:54,853 [INFO] cli.lib.common.git_helper: Progress: 80% - Receiving objects: 80% (118107/147633), 83.47 MiB | 55.68 MiB/s 2025-12-04T09:29:54.9283388Z 2025-12-04 09:29:54,928 [INFO] cli.lib.common.git_helper: Progress: 85% - Receiving objects: 85% (125489/147633), 114.46 MiB | 57.25 MiB/s 2025-12-04T09:29:54.9718903Z 2025-12-04 09:29:54,971 [INFO] cli.lib.common.git_helper: Progress: 90% - Receiving objects: 90% (132870/147633), 114.46 MiB | 57.25 MiB/s 2025-12-04T09:29:55.0304155Z 2025-12-04 09:29:55,030 [INFO] cli.lib.common.git_helper: Progress: 95% - Receiving objects: 95% (140252/147633), 114.46 MiB | 57.25 MiB/s 2025-12-04T09:29:55.0998029Z 2025-12-04 09:29:55,099 [INFO] cli.lib.common.git_helper: Progress: 100% - Receiving objects: 100% (147633/147633), 114.46 MiB | 57.25 MiB/s 2025-12-04T09:29:55.1194672Z 2025-12-04 09:29:55,119 [INFO] cli.lib.common.git_helper: Resolving deltas: 0% (0/116129) 2025-12-04T09:29:55.1505232Z 2025-12-04 09:29:55,150 [INFO] cli.lib.common.git_helper: Progress: 5% - Resolving deltas: 5% (5807/116129) 2025-12-04T09:29:55.2072096Z 2025-12-04 09:29:55,206 [INFO] cli.lib.common.git_helper: Progress: 10% - Resolving deltas: 10% (11613/116129) 2025-12-04T09:29:55.2586597Z 2025-12-04 09:29:55,258 [INFO] cli.lib.common.git_helper: Progress: 15% - Resolving deltas: 15% (17420/116129) 2025-12-04T09:29:55.3015487Z 2025-12-04 09:29:55,301 [INFO] cli.lib.common.git_helper: Progress: 20% - Resolving deltas: 20% (23226/116129) 2025-12-04T09:29:55.3543001Z 2025-12-04 09:29:55,353 [INFO] cli.lib.common.git_helper: Progress: 25% - Resolving deltas: 25% (29033/116129) 2025-12-04T09:29:55.3946970Z 2025-12-04 09:29:55,394 [INFO] cli.lib.common.git_helper: Progress: 30% - Resolving deltas: 30% (34839/116129) 2025-12-04T09:29:55.4505782Z 2025-12-04 09:29:55,450 [INFO] cli.lib.common.git_helper: Progress: 35% - Resolving deltas: 35% (40646/116129) 2025-12-04T09:29:55.4997555Z 2025-12-04 09:29:55,499 [INFO] cli.lib.common.git_helper: Progress: 40% - Resolving deltas: 40% (46452/116129) 2025-12-04T09:29:55.5342396Z 2025-12-04 09:29:55,533 [INFO] cli.lib.common.git_helper: Progress: 45% - Resolving deltas: 45% (52259/116129) 2025-12-04T09:29:55.5557083Z 2025-12-04 09:29:55,555 [INFO] cli.lib.common.git_helper: Progress: 50% - Resolving deltas: 50% (58065/116129) 2025-12-04T09:29:55.5780028Z 2025-12-04 09:29:55,577 [INFO] cli.lib.common.git_helper: Progress: 55% - Resolving deltas: 55% (63871/116129) 2025-12-04T09:29:55.6101176Z 2025-12-04 09:29:55,609 [INFO] cli.lib.common.git_helper: Progress: 60% - Resolving deltas: 60% (69678/116129) 2025-12-04T09:29:55.6392578Z 2025-12-04 09:29:55,638 [INFO] cli.lib.common.git_helper: Progress: 65% - Resolving deltas: 65% (75484/116129) 2025-12-04T09:29:55.6749230Z 2025-12-04 09:29:55,674 [INFO] cli.lib.common.git_helper: Progress: 70% - Resolving deltas: 70% (81291/116129) 2025-12-04T09:29:55.7012746Z 2025-12-04 09:29:55,700 [INFO] cli.lib.common.git_helper: Progress: 75% - Resolving deltas: 75% (87097/116129) 2025-12-04T09:29:55.7312239Z 2025-12-04 09:29:55,730 [INFO] cli.lib.common.git_helper: Progress: 80% - Resolving deltas: 80% (92904/116129) 2025-12-04T09:29:55.7592673Z 2025-12-04 09:29:55,758 [INFO] cli.lib.common.git_helper: Progress: 85% - Resolving deltas: 85% (98710/116129) 2025-12-04T09:29:55.7819042Z 2025-12-04 09:29:55,781 [INFO] cli.lib.common.git_helper: Progress: 90% - Resolving deltas: 90% (104517/116129) 2025-12-04T09:29:55.8036428Z 2025-12-04 09:29:55,803 [INFO] cli.lib.common.git_helper: Progress: 95% - Resolving deltas: 95% (110323/116129) 2025-12-04T09:29:55.8172313Z 2025-12-04 09:29:55,816 [INFO] cli.lib.common.git_helper: Progress: 100% - Resolving deltas: 100% (116129/116129) 2025-12-04T09:29:56.5120160Z 2025-12-04 09:29:56,511 [INFO] cli.lib.common.git_helper: Checking out pinned vllm commit e5192819208c4d68194844b7dfafbc00020d0dea 2025-12-04T09:29:56.9230036Z 2025-12-04 09:29:56,922 [INFO] cli.lib.common.git_helper: Successfully cloned vllm 2025-12-04T09:29:56.9600061Z 2025-12-04 09:29:56,959 [INFO] cli.lib.core.vllm.vllm_test: Running vllm test with inputs: VllmTestParameters(torch_whls_path=PosixPath('/var/lib/jenkins/workspace/dist'), vllm_whls_path=PosixPath('/var/lib/jenkins/workspace/dist/external/vllm/wheels'), torch_cuda_arch_list='8.9', cleaning_script=PosixPath('/var/lib/jenkins/workspace/.github/ci_configs/vllm/use_existing_torch.py')) 2025-12-04T09:29:56.9618402Z 2025-12-04 09:29:56,961 [INFO] cli.lib.common.pip_helper: torch already exist with version: 2.10.0a0+gitffd9b0f 2025-12-04T09:29:56.9620782Z 2025-12-04 09:29:56,961 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0+617079d-cp312-cp312-linux_x86_64.whl... 2025-12-04T09:29:56.9622217Z 2025-12-04 09:29:56,961 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0+617079d-cp312-cp312-linux_x86_64.whl 2025-12-04T09:29:56.9623709Z 2025-12-04 09:29:56,962 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0+617079d-cp312-cp312-linux_x86_64.whl 2025-12-04T09:29:57.2472301Z Processing /var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0+617079d-cp312-cp312-linux_x86_64.whl 2025-12-04T09:29:57.2548389Z Requirement already satisfied: numpy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchvision==0.25.0a0+617079d) (1.26.2) 2025-12-04T09:29:57.2551537Z Requirement already satisfied: torch in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchvision==0.25.0a0+617079d) (2.10.0a0+gitffd9b0f) 2025-12-04T09:29:57.2555240Z Requirement already satisfied: pillow!=8.3.*,>=5.3.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchvision==0.25.0a0+617079d) (11.0.0) 2025-12-04T09:29:57.2606842Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+617079d) (3.18.0) 2025-12-04T09:29:57.2611266Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+617079d) (4.12.2) 2025-12-04T09:29:57.2615802Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+617079d) (78.1.1) 2025-12-04T09:29:57.2619815Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+617079d) (1.13.3) 2025-12-04T09:29:57.2623695Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+617079d) (2.8.8) 2025-12-04T09:29:57.2626635Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+617079d) (3.1.6) 2025-12-04T09:29:57.2635162Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+617079d) (2025.10.0) 2025-12-04T09:29:57.2734298Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch->torchvision==0.25.0a0+617079d) (1.3.0) 2025-12-04T09:29:57.2774405Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch->torchvision==0.25.0a0+617079d) (3.0.3) 2025-12-04T09:29:57.4373026Z Installing collected packages: torchvision 2025-12-04T09:29:57.8666891Z Successfully installed torchvision-0.25.0a0+617079d 2025-12-04T09:29:57.9126172Z 2025-12-04 09:29:57,912 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0+e90a398-cp312-cp312-linux_x86_64.whl... 2025-12-04T09:29:57.9127641Z 2025-12-04 09:29:57,912 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0+e90a398-cp312-cp312-linux_x86_64.whl 2025-12-04T09:29:57.9129204Z 2025-12-04 09:29:57,912 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0+e90a398-cp312-cp312-linux_x86_64.whl 2025-12-04T09:29:58.1965757Z Processing /var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0+e90a398-cp312-cp312-linux_x86_64.whl 2025-12-04T09:29:58.2022788Z Requirement already satisfied: torch in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchaudio==2.10.0a0+e90a398) (2.10.0a0+gitffd9b0f) 2025-12-04T09:29:58.2031238Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+e90a398) (3.18.0) 2025-12-04T09:29:58.2035411Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+e90a398) (4.12.2) 2025-12-04T09:29:58.2040643Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+e90a398) (78.1.1) 2025-12-04T09:29:58.2044518Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+e90a398) (1.13.3) 2025-12-04T09:29:58.2048225Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+e90a398) (2.8.8) 2025-12-04T09:29:58.2051239Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+e90a398) (3.1.6) 2025-12-04T09:29:58.2055184Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+e90a398) (2025.10.0) 2025-12-04T09:29:58.2159205Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch->torchaudio==2.10.0a0+e90a398) (1.3.0) 2025-12-04T09:29:58.2199207Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch->torchaudio==2.10.0a0+e90a398) (3.0.3) 2025-12-04T09:29:58.3816371Z Installing collected packages: torchaudio 2025-12-04T09:29:58.5297842Z Successfully installed torchaudio-2.10.0a0+e90a398 2025-12-04T09:29:58.5702787Z 2025-12-04 09:29:58,569 [INFO] cli.lib.core.vllm.vllm_test: Done. Installed torch and other torch-related wheels 2025-12-04T09:29:58.5703442Z 2025-12-04 09:29:58,569 [INFO] cli.lib.core.vllm.vllm_test: Installing vllm wheels 2025-12-04T09:29:58.5705082Z 2025-12-04 09:29:58,570 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/external/vllm/wheels/xformers/xformers-0.0.34+3f91ad6f.d20251204-cp39-abi3-linux_x86_64.whl... 2025-12-04T09:29:58.5706841Z 2025-12-04 09:29:58,570 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/xformers/xformers-0.0.34+3f91ad6f.d20251204-cp39-abi3-linux_x86_64.whl 2025-12-04T09:29:58.5708304Z 2025-12-04 09:29:58,570 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/xformers/xformers-0.0.34+3f91ad6f.d20251204-cp39-abi3-linux_x86_64.whl 2025-12-04T09:29:58.8537328Z Processing /var/lib/jenkins/workspace/dist/external/vllm/wheels/xformers/xformers-0.0.34+3f91ad6f.d20251204-cp39-abi3-linux_x86_64.whl 2025-12-04T09:29:59.0507426Z Requirement already satisfied: torch>=2.9 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from xformers==0.0.34+3f91ad6f.d20251204) (2.10.0a0+gitffd9b0f) 2025-12-04T09:29:59.0509351Z Requirement already satisfied: numpy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from xformers==0.0.34+3f91ad6f.d20251204) (1.26.2) 2025-12-04T09:29:59.0519765Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.9->xformers==0.0.34+3f91ad6f.d20251204) (3.18.0) 2025-12-04T09:29:59.0524073Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.9->xformers==0.0.34+3f91ad6f.d20251204) (4.12.2) 2025-12-04T09:29:59.0529495Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.9->xformers==0.0.34+3f91ad6f.d20251204) (78.1.1) 2025-12-04T09:29:59.0533514Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.9->xformers==0.0.34+3f91ad6f.d20251204) (1.13.3) 2025-12-04T09:29:59.0537996Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.9->xformers==0.0.34+3f91ad6f.d20251204) (2.8.8) 2025-12-04T09:29:59.0541212Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.9->xformers==0.0.34+3f91ad6f.d20251204) (3.1.6) 2025-12-04T09:29:59.0545043Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.9->xformers==0.0.34+3f91ad6f.d20251204) (2025.10.0) 2025-12-04T09:29:59.0651466Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch>=2.9->xformers==0.0.34+3f91ad6f.d20251204) (1.3.0) 2025-12-04T09:29:59.0691152Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch>=2.9->xformers==0.0.34+3f91ad6f.d20251204) (3.0.3) 2025-12-04T09:29:59.2344968Z Installing collected packages: xformers 2025-12-04T09:30:03.0827954Z Successfully installed xformers-0.0.34+3f91ad6f.d20251204 2025-12-04T09:30:03.1271813Z 2025-12-04 09:30:03,126 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev389+ge51928192.d20251204-cp38-abi3-linux_x86_64.whl... 2025-12-04T09:30:03.1273924Z 2025-12-04 09:30:03,126 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev389+ge51928192.d20251204-cp38-abi3-linux_x86_64.whl 2025-12-04T09:30:03.1275712Z 2025-12-04 09:30:03,126 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev389+ge51928192.d20251204-cp38-abi3-linux_x86_64.whl 2025-12-04T09:30:03.4112473Z Processing /var/lib/jenkins/workspace/dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev389+ge51928192.d20251204-cp38-abi3-linux_x86_64.whl 2025-12-04T09:30:04.0173457Z Collecting regex (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.0395949Z Downloading regex-2025.11.3-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl.metadata (40 kB) 2025-12-04T09:30:04.0510764Z Requirement already satisfied: cachetools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) (6.2.2) 2025-12-04T09:30:04.0513282Z Requirement already satisfied: psutil in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) (7.1.3) 2025-12-04T09:30:04.0891633Z Collecting sentencepiece (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.0924421Z Downloading sentencepiece-0.2.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl.metadata (10 kB) 2025-12-04T09:30:04.0994364Z Requirement already satisfied: numpy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) (1.26.2) 2025-12-04T09:30:04.0997653Z Requirement already satisfied: requests>=2.26.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) (2.32.5) 2025-12-04T09:30:04.1000216Z Requirement already satisfied: tqdm in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) (4.67.1) 2025-12-04T09:30:04.1507945Z Collecting blake3 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.1534591Z Downloading blake3-1.0.8-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.6 kB) 2025-12-04T09:30:04.1695716Z Collecting py-cpuinfo (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.1727429Z Downloading py_cpuinfo-9.0.0-py3-none-any.whl.metadata (794 bytes) 2025-12-04T09:30:04.2104312Z Collecting transformers>=4.55.2 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.2130838Z Downloading transformers-4.57.3-py3-none-any.whl.metadata (43 kB) 2025-12-04T09:30:04.4758480Z Collecting tokenizers>=0.21.1 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.4787974Z Downloading tokenizers-0.22.1-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.8 kB) 2025-12-04T09:30:04.4866415Z Requirement already satisfied: protobuf in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) (5.29.5) 2025-12-04T09:30:04.5420755Z Collecting fastapi>=0.115.0 (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.5452286Z Downloading fastapi-0.123.7-py3-none-any.whl.metadata (30 kB) 2025-12-04T09:30:04.5572847Z Requirement already satisfied: aiohttp in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) (3.13.2) 2025-12-04T09:30:04.6016019Z Collecting openai>=1.99.1 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.6043385Z Downloading openai-2.8.1-py3-none-any.whl.metadata (29 kB) 2025-12-04T09:30:04.6989609Z Collecting pydantic>=2.12.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.7020391Z Downloading pydantic-2.12.5-py3-none-any.whl.metadata (90 kB) 2025-12-04T09:30:04.7244010Z Collecting prometheus_client>=0.18.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.7275925Z Downloading prometheus_client-0.23.1-py3-none-any.whl.metadata (1.9 kB) 2025-12-04T09:30:04.7340765Z Requirement already satisfied: pillow in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) (11.0.0) 2025-12-04T09:30:04.7485178Z Collecting prometheus-fastapi-instrumentator>=7.0.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.7517187Z Downloading prometheus_fastapi_instrumentator-7.1.0-py3-none-any.whl.metadata (13 kB) 2025-12-04T09:30:04.7888834Z Collecting tiktoken>=0.6.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.7918216Z Downloading tiktoken-0.12.0-cp312-cp312-manylinux_2_28_x86_64.whl.metadata (6.7 kB) 2025-12-04T09:30:04.8166714Z Collecting lm-format-enforcer==0.11.3 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.8197574Z Downloading lm_format_enforcer-0.11.3-py3-none-any.whl.metadata (17 kB) 2025-12-04T09:30:04.8648932Z Collecting llguidance<0.8.0,>=0.7.11 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.8688972Z Downloading llguidance-0.7.30-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (10 kB) 2025-12-04T09:30:04.9006604Z Collecting outlines_core==0.2.11 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.9042510Z Downloading outlines_core-0.2.11-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.8 kB) 2025-12-04T09:30:04.9285116Z Collecting diskcache==5.6.3 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.9311299Z Downloading diskcache-5.6.3-py3-none-any.whl.metadata (20 kB) 2025-12-04T09:30:04.9479024Z Collecting lark==1.2.2 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.9503996Z Downloading lark-1.2.2-py3-none-any.whl.metadata (1.8 kB) 2025-12-04T09:30:04.9870550Z Collecting xgrammar==0.1.25 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:04.9916743Z Downloading xgrammar-0.1.25-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.5 kB) 2025-12-04T09:30:04.9994638Z Requirement already satisfied: typing_extensions>=4.10 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) (4.12.2) 2025-12-04T09:30:04.9997361Z Requirement already satisfied: filelock>=3.16.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) (3.18.0) 2025-12-04T09:30:05.0113645Z Collecting partial-json-parser (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.0155627Z Downloading partial_json_parser-0.2.1.1.post7-py3-none-any.whl.metadata (6.1 kB) 2025-12-04T09:30:05.1989967Z Collecting pyzmq>=25.0.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.2027348Z Downloading pyzmq-27.1.0-cp312-abi3-manylinux_2_26_x86_64.manylinux_2_28_x86_64.whl.metadata (6.0 kB) 2025-12-04T09:30:05.2533651Z Collecting msgspec (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.2566435Z Downloading msgspec-0.20.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl.metadata (5.5 kB) 2025-12-04T09:30:05.3068243Z Collecting gguf>=0.13.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.3100470Z Downloading gguf-0.17.1-py3-none-any.whl.metadata (4.3 kB) 2025-12-04T09:30:05.3312282Z Collecting mistral_common>=1.8.5 (from mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.3356073Z Downloading mistral_common-1.8.6-py3-none-any.whl.metadata (5.3 kB) 2025-12-04T09:30:05.3904084Z Collecting opencv-python-headless>=4.11.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.3931763Z Downloading opencv_python_headless-4.12.0.88-cp37-abi3-manylinux2014_x86_64.manylinux_2_17_x86_64.whl.metadata (19 kB) 2025-12-04T09:30:05.4006108Z Requirement already satisfied: pyyaml in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) (6.0.2) 2025-12-04T09:30:05.4011666Z Requirement already satisfied: six>=1.16.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) (1.17.0) 2025-12-04T09:30:05.4016208Z Requirement already satisfied: setuptools<80,>=77.0.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) (78.1.1) 2025-12-04T09:30:05.4116433Z Collecting einops (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.4143129Z Downloading einops-0.8.1-py3-none-any.whl.metadata (13 kB) 2025-12-04T09:30:05.4408886Z Collecting compressed-tensors==0.11.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.4443856Z Downloading compressed_tensors-0.11.0-py3-none-any.whl.metadata (7.0 kB) 2025-12-04T09:30:05.4618984Z Collecting depyf==0.19.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.4653732Z Downloading depyf-0.19.0-py3-none-any.whl.metadata (7.3 kB) 2025-12-04T09:30:05.4871528Z Collecting cloudpickle (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.4904116Z Downloading cloudpickle-3.1.2-py3-none-any.whl.metadata (7.1 kB) 2025-12-04T09:30:05.5566630Z Collecting watchfiles (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.5598343Z Downloading watchfiles-1.1.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.9 kB) 2025-12-04T09:30:05.5790373Z Collecting python-json-logger (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.5818710Z Downloading python_json_logger-4.0.0-py3-none-any.whl.metadata (4.0 kB) 2025-12-04T09:30:05.5898359Z Requirement already satisfied: scipy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) (1.14.1) 2025-12-04T09:30:05.5900983Z Requirement already satisfied: ninja in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) (1.10.2) 2025-12-04T09:30:05.6585932Z Collecting pybase64 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.6613924Z Downloading pybase64-1.4.2-cp312-cp312-manylinux1_x86_64.manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_5_x86_64.whl.metadata (8.7 kB) 2025-12-04T09:30:05.7022338Z Collecting cbor2 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.7049798Z Downloading cbor2-5.7.1-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl.metadata (5.4 kB) 2025-12-04T09:30:05.7509826Z Collecting setproctitle (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.7546624Z Downloading setproctitle-1.3.7-cp312-cp312-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl.metadata (10 kB) 2025-12-04T09:30:05.7775693Z Collecting openai-harmony>=0.0.3 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.7810760Z Downloading openai_harmony-0.0.8-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (8.0 kB) 2025-12-04T09:30:05.8559989Z Collecting numba==0.61.2 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.8586339Z Downloading numba-0.61.2-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl.metadata (2.8 kB) 2025-12-04T09:30:05.9502266Z Collecting ray>=2.48.0 (from ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:05.9535909Z Downloading ray-2.52.1-cp312-cp312-manylinux2014_x86_64.whl.metadata (21 kB) 2025-12-04T09:30:06.0203402Z Collecting flashinfer-python==0.4.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:06.0265966Z Downloading flashinfer_python-0.4.0.tar.gz (4.5 MB) 2025-12-04T09:30:06.0572957Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/4.5 MB ? eta -:--:-- 2025-12-04T09:30:06.0573630Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 4.5/4.5 MB 153.1 MB/s 0:00:00 2025-12-04T09:30:07.7999763Z [?25h Installing build dependencies ... [?25l- \ | / done 2025-12-04T09:30:08.3745623Z [?25h Getting requirements to build wheel ... [?25l- done 2025-12-04T09:30:09.1718328Z [?25h Preparing metadata (pyproject.toml) ... [?25l- done 2025-12-04T09:30:09.1747819Z [?25hRequirement already satisfied: torch>=1.7.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from compressed-tensors==0.11.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (2.10.0a0+gitffd9b0f) 2025-12-04T09:30:09.2273160Z Collecting frozendict (from compressed-tensors==0.11.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:09.2306333Z Downloading frozendict-2.4.7-py3-none-any.whl.metadata (23 kB) 2025-12-04T09:30:09.2499626Z Collecting astor (from depyf==0.19.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:09.2530922Z Downloading astor-0.8.1-py2.py3-none-any.whl.metadata (4.2 kB) 2025-12-04T09:30:09.2602005Z Requirement already satisfied: dill in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from depyf==0.19.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (0.3.7) 2025-12-04T09:30:09.2905023Z Collecting apache-tvm-ffi==0.1.0b15 (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:09.2915340Z Using cached apache_tvm_ffi-0.1.0b15-cp312-abi3-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl.metadata (2.0 kB) 2025-12-04T09:30:09.2934852Z Requirement already satisfied: click in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (8.3.1) 2025-12-04T09:30:09.3102470Z Collecting nvidia-cudnn-frontend>=1.13.0 (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:09.3149189Z Downloading nvidia_cudnn_frontend-1.16.0-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl.metadata (6.8 kB) 2025-12-04T09:30:09.3341296Z Collecting nvidia-cutlass-dsl>=4.2.1 (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:09.3381933Z Downloading nvidia_cutlass_dsl-4.3.2-cp312-cp312-manylinux_2_28_x86_64.whl.metadata (2.6 kB) 2025-12-04T09:30:09.3563309Z Collecting nvidia-ml-py (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:09.3590999Z Downloading nvidia_ml_py-13.580.82-py3-none-any.whl.metadata (9.6 kB) 2025-12-04T09:30:09.3786215Z Collecting packaging>=24.2 (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:09.3796030Z Using cached packaging-25.0-py3-none-any.whl.metadata (3.3 kB) 2025-12-04T09:30:09.3810899Z Requirement already satisfied: tabulate in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (0.9.0) 2025-12-04T09:30:09.3935795Z Collecting interegular>=0.3.2 (from lm-format-enforcer==0.11.3->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:09.3968178Z Downloading interegular-0.3.3-py37-none-any.whl.metadata (3.0 kB) 2025-12-04T09:30:09.4595774Z Collecting llvmlite<0.45,>=0.44.0dev0 (from numba==0.61.2->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:09.4625137Z Downloading llvmlite-0.44.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.0 kB) 2025-12-04T09:30:09.4781631Z Requirement already satisfied: triton in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from xgrammar==0.1.25->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (3.5.1+gitbfeb0668) 2025-12-04T09:30:09.5113822Z Collecting starlette<0.51.0,>=0.40.0 (from fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:09.5141351Z Downloading starlette-0.50.0-py3-none-any.whl.metadata (6.3 kB) 2025-12-04T09:30:09.5482498Z Collecting annotated-doc>=0.0.2 (from fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:09.5509628Z Downloading annotated_doc-0.0.4-py3-none-any.whl.metadata (6.6 kB) 2025-12-04T09:30:09.5683443Z Collecting annotated-types>=0.6.0 (from pydantic>=2.12.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:09.5712801Z Downloading annotated_types-0.7.0-py3-none-any.whl.metadata (15 kB) 2025-12-04T09:30:10.1211985Z Collecting pydantic-core==2.41.5 (from pydantic>=2.12.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.1241685Z Downloading pydantic_core-2.41.5-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (7.3 kB) 2025-12-04T09:30:10.1461817Z Collecting typing_extensions>=4.10 (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.1472249Z Using cached typing_extensions-4.15.0-py3-none-any.whl.metadata (3.3 kB) 2025-12-04T09:30:10.1594454Z Collecting typing-inspection>=0.4.2 (from pydantic>=2.12.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.1625518Z Downloading typing_inspection-0.4.2-py3-none-any.whl.metadata (2.6 kB) 2025-12-04T09:30:10.1891475Z Collecting anyio<5,>=3.6.2 (from starlette<0.51.0,>=0.40.0->fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.1918612Z Downloading anyio-4.12.0-py3-none-any.whl.metadata (4.3 kB) 2025-12-04T09:30:10.2010006Z Requirement already satisfied: idna>=2.8 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from anyio<5,>=3.6.2->starlette<0.51.0,>=0.40.0->fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (3.11) 2025-12-04T09:30:10.2159879Z Collecting fastapi-cli>=0.0.8 (from fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.2190520Z Downloading fastapi_cli-0.0.16-py3-none-any.whl.metadata (6.4 kB) 2025-12-04T09:30:10.2459554Z Collecting httpx<1.0.0,>=0.23.0 (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.2487096Z Downloading httpx-0.28.1-py3-none-any.whl.metadata (7.1 kB) 2025-12-04T09:30:10.2575699Z Requirement already satisfied: jinja2>=3.1.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (3.1.6) 2025-12-04T09:30:10.2692292Z Collecting python-multipart>=0.0.18 (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.2723083Z Downloading python_multipart-0.0.20-py3-none-any.whl.metadata (1.8 kB) 2025-12-04T09:30:10.2928492Z Collecting email-validator>=2.0.0 (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.2960869Z Downloading email_validator-2.3.0-py3-none-any.whl.metadata (26 kB) 2025-12-04T09:30:10.3284202Z Collecting uvicorn>=0.12.0 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.3321926Z Downloading uvicorn-0.38.0-py3-none-any.whl.metadata (6.8 kB) 2025-12-04T09:30:10.3408568Z Requirement already satisfied: certifi in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from httpx<1.0.0,>=0.23.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (2025.11.12) 2025-12-04T09:30:10.3580472Z Collecting httpcore==1.* (from httpx<1.0.0,>=0.23.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.3610978Z Downloading httpcore-1.0.9-py3-none-any.whl.metadata (21 kB) 2025-12-04T09:30:10.3817066Z Collecting h11>=0.16 (from httpcore==1.*->httpx<1.0.0,>=0.23.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.3850189Z Downloading h11-0.16.0-py3-none-any.whl.metadata (8.3 kB) 2025-12-04T09:30:10.4058066Z Collecting dnspython>=2.0.0 (from email-validator>=2.0.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.4089919Z Downloading dnspython-2.8.0-py3-none-any.whl.metadata (5.7 kB) 2025-12-04T09:30:10.4372558Z Collecting typer>=0.15.1 (from fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.4406254Z Downloading typer-0.20.0-py3-none-any.whl.metadata (16 kB) 2025-12-04T09:30:10.4699927Z Collecting rich-toolkit>=0.14.8 (from fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.4731097Z Downloading rich_toolkit-0.17.0-py3-none-any.whl.metadata (1.0 kB) 2025-12-04T09:30:10.4934219Z Collecting fastapi-cloud-cli>=0.1.1 (from fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.4963408Z Downloading fastapi_cloud_cli-0.5.2-py3-none-any.whl.metadata (3.3 kB) 2025-12-04T09:30:10.5943126Z Collecting rignore>=0.5.1 (from fastapi-cloud-cli>=0.1.1->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.5983977Z Downloading rignore-0.7.6-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.2 kB) 2025-12-04T09:30:10.6711366Z Collecting sentry-sdk>=2.20.0 (from fastapi-cloud-cli>=0.1.1->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.6739060Z Downloading sentry_sdk-2.47.0-py2.py3-none-any.whl.metadata (10 kB) 2025-12-04T09:30:10.7412408Z Collecting fastar>=0.5.0 (from fastapi-cloud-cli>=0.1.1->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.7453061Z Downloading fastar-0.8.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.0 kB) 2025-12-04T09:30:10.7610085Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2>=3.1.5->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (3.0.3) 2025-12-04T09:30:10.8013390Z Collecting jsonschema>=4.21.1 (from mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.8041556Z Downloading jsonschema-4.25.1-py3-none-any.whl.metadata (7.6 kB) 2025-12-04T09:30:10.8299519Z Collecting pydantic-extra-types>=2.10.5 (from pydantic-extra-types[pycountry]>=2.10.5->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.8331236Z Downloading pydantic_extra_types-2.10.6-py3-none-any.whl.metadata (4.0 kB) 2025-12-04T09:30:10.8442095Z Requirement already satisfied: attrs>=22.2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jsonschema>=4.21.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (25.4.0) 2025-12-04T09:30:10.8567069Z Collecting jsonschema-specifications>=2023.03.6 (from jsonschema>=4.21.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.8601062Z Downloading jsonschema_specifications-2025.9.1-py3-none-any.whl.metadata (2.9 kB) 2025-12-04T09:30:10.8892429Z Collecting referencing>=0.28.4 (from jsonschema>=4.21.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:10.8922445Z Downloading referencing-0.37.0-py3-none-any.whl.metadata (2.8 kB) 2025-12-04T09:30:11.1846041Z Collecting rpds-py>=0.7.1 (from jsonschema>=4.21.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:11.1875896Z Downloading rpds_py-0.30.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.1 kB) 2025-12-04T09:30:11.2319256Z Collecting cuda-python>=12.8 (from nvidia-cutlass-dsl>=4.2.1->flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:11.2351704Z Downloading cuda_python-13.0.3-py3-none-any.whl.metadata (4.7 kB) 2025-12-04T09:30:11.2595378Z Collecting cuda-bindings~=13.0.3 (from cuda-python>=12.8->nvidia-cutlass-dsl>=4.2.1->flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:11.2634459Z Downloading cuda_bindings-13.0.3-cp312-cp312-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl.metadata (2.3 kB) 2025-12-04T09:30:11.2703796Z Requirement already satisfied: cuda-pathfinder~=1.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from cuda-python>=12.8->nvidia-cutlass-dsl>=4.2.1->flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (1.3.2) 2025-12-04T09:30:11.2782358Z Requirement already satisfied: distro<2,>=1.7.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from openai>=1.99.1->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (1.9.0) 2025-12-04T09:30:11.3371936Z Collecting jiter<1,>=0.10.0 (from openai>=1.99.1->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:11.3403960Z Downloading jiter-0.12.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.2 kB) 2025-12-04T09:30:11.3760945Z Collecting sniffio (from openai>=1.99.1->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:11.3792619Z Downloading sniffio-1.3.1-py3-none-any.whl.metadata (3.9 kB) 2025-12-04T09:30:11.5494354Z Collecting numpy (from vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:11.5525115Z Downloading numpy-2.2.6-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (62 kB) 2025-12-04T09:30:11.6229329Z Collecting pycountry>=23 (from pydantic-extra-types[pycountry]>=2.10.5->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:11.6261851Z Downloading pycountry-24.6.1-py3-none-any.whl.metadata (12 kB) 2025-12-04T09:30:11.6605409Z Collecting click (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:11.6636657Z Downloading click-8.2.1-py3-none-any.whl.metadata (2.5 kB) 2025-12-04T09:30:11.6702569Z Requirement already satisfied: msgpack<2.0.0,>=1.0.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from ray>=2.48.0->ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (1.1.2) 2025-12-04T09:30:11.6960596Z Collecting cupy-cuda12x (from ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:11.6987855Z Downloading cupy_cuda12x-13.6.0-cp312-cp312-manylinux2014_x86_64.whl.metadata (2.4 kB) 2025-12-04T09:30:11.7211244Z Requirement already satisfied: charset_normalizer<4,>=2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (3.4.4) 2025-12-04T09:30:11.7217161Z Requirement already satisfied: urllib3<3,>=1.21.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (2.5.0) 2025-12-04T09:30:11.7639243Z Collecting rich>=13.7.1 (from rich-toolkit>=0.14.8->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:11.7671923Z Downloading rich-14.2.0-py3-none-any.whl.metadata (18 kB) 2025-12-04T09:30:11.7939585Z Collecting markdown-it-py>=2.2.0 (from rich>=13.7.1->rich-toolkit>=0.14.8->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:11.7974037Z Downloading markdown_it_py-4.0.0-py3-none-any.whl.metadata (7.3 kB) 2025-12-04T09:30:11.8074659Z Requirement already satisfied: pygments<3.0.0,>=2.13.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from rich>=13.7.1->rich-toolkit>=0.14.8->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (2.15.0) 2025-12-04T09:30:11.8209126Z Collecting mdurl~=0.1 (from markdown-it-py>=2.2.0->rich>=13.7.1->rich-toolkit>=0.14.8->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:11.8241406Z Downloading mdurl-0.1.2-py3-none-any.whl.metadata (1.6 kB) 2025-12-04T09:30:11.8895631Z Collecting huggingface-hub<2.0,>=0.16.4 (from tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:11.8921761Z Downloading huggingface_hub-1.1.7-py3-none-any.whl.metadata (13 kB) 2025-12-04T09:30:11.9121866Z Requirement already satisfied: fsspec>=2023.5.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from huggingface-hub<2.0,>=0.16.4->tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (2025.10.0) 2025-12-04T09:30:11.9441165Z Collecting hf-xet<2.0.0,>=1.2.0 (from huggingface-hub<2.0,>=0.16.4->tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:11.9466800Z Downloading hf_xet-1.2.0-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.9 kB) 2025-12-04T09:30:11.9696335Z Collecting shellingham (from huggingface-hub<2.0,>=0.16.4->tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:11.9731279Z Downloading shellingham-1.5.4-py2.py3-none-any.whl.metadata (3.5 kB) 2025-12-04T09:30:11.9940091Z Collecting typer-slim (from huggingface-hub<2.0,>=0.16.4->tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:11.9975613Z Downloading typer_slim-0.20.0-py3-none-any.whl.metadata (16 kB) 2025-12-04T09:30:12.0203175Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=1.7.0->compressed-tensors==0.11.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (1.13.3) 2025-12-04T09:30:12.0206688Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=1.7.0->compressed-tensors==0.11.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (2.8.8) 2025-12-04T09:30:12.0260170Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch>=1.7.0->compressed-tensors==0.11.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (1.3.0) 2025-12-04T09:30:12.0403897Z Collecting huggingface-hub<2.0,>=0.16.4 (from tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:12.0430672Z Downloading huggingface_hub-0.36.0-py3-none-any.whl.metadata (14 kB) 2025-12-04T09:30:12.1717682Z Collecting safetensors>=0.4.3 (from transformers>=4.55.2->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:12.1752902Z Downloading safetensors-0.7.0-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.1 kB) 2025-12-04T09:30:12.2531521Z Collecting httptools>=0.6.3 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:12.2563674Z Downloading httptools-0.7.1-cp312-cp312-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl.metadata (3.5 kB) 2025-12-04T09:30:12.2785153Z Collecting python-dotenv>=0.13 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:12.2814148Z Downloading python_dotenv-1.2.1-py3-none-any.whl.metadata (25 kB) 2025-12-04T09:30:12.3228423Z Collecting uvloop>=0.15.1 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:12.3259834Z Downloading uvloop-0.22.1-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl.metadata (4.9 kB) 2025-12-04T09:30:12.4905737Z Collecting websockets>=10.4 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:12.4940091Z Downloading websockets-15.0.1-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.8 kB) 2025-12-04T09:30:12.5176013Z Requirement already satisfied: aiohappyeyeballs>=2.5.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (2.6.1) 2025-12-04T09:30:12.5179679Z Requirement already satisfied: aiosignal>=1.4.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (1.4.0) 2025-12-04T09:30:12.5186323Z Requirement already satisfied: frozenlist>=1.1.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (1.8.0) 2025-12-04T09:30:12.5190535Z Requirement already satisfied: multidict<7.0,>=4.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (6.7.0) 2025-12-04T09:30:12.5194287Z Requirement already satisfied: propcache>=0.2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (0.4.1) 2025-12-04T09:30:12.5198312Z Requirement already satisfied: yarl<2.0,>=1.17.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (1.22.0) 2025-12-04T09:30:12.5775868Z Collecting fastrlock>=0.5 (from cupy-cuda12x->ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev389+ge51928192.d20251204) 2025-12-04T09:30:12.5811442Z Downloading fastrlock-0.8.3-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_28_x86_64.whl.metadata (7.7 kB) 2025-12-04T09:30:12.6016641Z Requirement already satisfied: soundfile>=0.12.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (0.13.1) 2025-12-04T09:30:12.6036530Z Requirement already satisfied: cffi>=1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from soundfile>=0.12.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (2.0.0) 2025-12-04T09:30:12.6056377Z Requirement already satisfied: pycparser in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from cffi>=1.0->soundfile>=0.12.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (2.23) 2025-12-04T09:30:12.6083204Z Requirement already satisfied: soxr>=0.5.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251204) (1.0.0) 2025-12-04T09:30:12.6665292Z Downloading compressed_tensors-0.11.0-py3-none-any.whl (179 kB) 2025-12-04T09:30:12.6774087Z Downloading depyf-0.19.0-py3-none-any.whl (39 kB) 2025-12-04T09:30:12.6873846Z Downloading diskcache-5.6.3-py3-none-any.whl (45 kB) 2025-12-04T09:30:12.6939886Z Using cached apache_tvm_ffi-0.1.0b15-cp312-abi3-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl (1.6 MB) 2025-12-04T09:30:12.6983697Z Downloading lark-1.2.2-py3-none-any.whl (111 kB) 2025-12-04T09:30:12.7074148Z Downloading lm_format_enforcer-0.11.3-py3-none-any.whl (45 kB) 2025-12-04T09:30:12.7175673Z Downloading numba-0.61.2-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (3.9 MB) 2025-12-04T09:30:12.7457213Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.9 MB ? eta -:--:-- 2025-12-04T09:30:12.7458119Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.9/3.9 MB 144.3 MB/s 0:00:00 2025-12-04T09:30:12.7506013Z [?25hDownloading outlines_core-0.2.11-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (2.3 MB) 2025-12-04T09:30:12.7672406Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/2.3 MB ? eta -:--:-- 2025-12-04T09:30:12.7673042Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.3/2.3 MB 142.8 MB/s 0:00:00 2025-12-04T09:30:12.7726382Z [?25hDownloading xgrammar-0.1.25-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (8.7 MB) 2025-12-04T09:30:12.8208395Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/8.7 MB ? eta -:--:-- 2025-12-04T09:30:12.8209037Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 8.7/8.7 MB 185.5 MB/s 0:00:00 2025-12-04T09:30:12.8274099Z [?25hDownloading llguidance-0.7.30-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (15.0 MB) 2025-12-04T09:30:12.9055224Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/15.0 MB ? eta -:--:-- 2025-12-04T09:30:12.9055922Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 15.0/15.0 MB 195.7 MB/s 0:00:00 2025-12-04T09:30:12.9089908Z [?25hDownloading llvmlite-0.44.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (42.4 MB) 2025-12-04T09:30:13.1115103Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/42.4 MB ? eta -:--:-- 2025-12-04T09:30:13.1615906Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 42.2/42.4 MB 236.7 MB/s eta 0:00:01 2025-12-04T09:30:13.1616574Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 42.4/42.4 MB 168.4 MB/s 0:00:00 2025-12-04T09:30:13.1647502Z [?25hDownloading fastapi-0.123.7-py3-none-any.whl (111 kB) 2025-12-04T09:30:13.1723532Z Downloading pydantic-2.12.5-py3-none-any.whl (463 kB) 2025-12-04T09:30:13.1822419Z Downloading pydantic_core-2.41.5-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (2.1 MB) 2025-12-04T09:30:13.1951948Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/2.1 MB ? eta -:--:-- 2025-12-04T09:30:13.1952904Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.1/2.1 MB 175.7 MB/s 0:00:00 2025-12-04T09:30:13.1989296Z [?25hDownloading starlette-0.50.0-py3-none-any.whl (74 kB) 2025-12-04T09:30:13.2058289Z Downloading anyio-4.12.0-py3-none-any.whl (113 kB) 2025-12-04T09:30:13.2125025Z Downloading annotated_doc-0.0.4-py3-none-any.whl (5.3 kB) 2025-12-04T09:30:13.2181707Z Downloading annotated_types-0.7.0-py3-none-any.whl (13 kB) 2025-12-04T09:30:13.2243301Z Downloading httpx-0.28.1-py3-none-any.whl (73 kB) 2025-12-04T09:30:13.2313903Z Downloading httpcore-1.0.9-py3-none-any.whl (78 kB) 2025-12-04T09:30:13.2384284Z Downloading email_validator-2.3.0-py3-none-any.whl (35 kB) 2025-12-04T09:30:13.2448390Z Downloading dnspython-2.8.0-py3-none-any.whl (331 kB) 2025-12-04T09:30:13.2543964Z Downloading fastapi_cli-0.0.16-py3-none-any.whl (12 kB) 2025-12-04T09:30:13.2603195Z Downloading fastapi_cloud_cli-0.5.2-py3-none-any.whl (23 kB) 2025-12-04T09:30:13.2684001Z Downloading fastar-0.8.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (821 kB) 2025-12-04T09:30:13.2785852Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/821.6 kB ? eta -:--:-- 2025-12-04T09:30:13.2786844Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 821.6/821.6 kB 84.6 MB/s 0:00:00 2025-12-04T09:30:13.2812631Z [?25hDownloading gguf-0.17.1-py3-none-any.whl (96 kB) 2025-12-04T09:30:13.2886810Z Downloading h11-0.16.0-py3-none-any.whl (37 kB) 2025-12-04T09:30:13.2963086Z Downloading interegular-0.3.3-py37-none-any.whl (23 kB) 2025-12-04T09:30:13.3053862Z Downloading mistral_common-1.8.6-py3-none-any.whl (6.5 MB) 2025-12-04T09:30:13.3430379Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/6.5 MB ? eta -:--:-- 2025-12-04T09:30:13.3431356Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 6.5/6.5 MB 178.2 MB/s 0:00:00 2025-12-04T09:30:13.3457612Z [?25hDownloading jsonschema-4.25.1-py3-none-any.whl (90 kB) 2025-12-04T09:30:13.3525189Z Downloading jsonschema_specifications-2025.9.1-py3-none-any.whl (18 kB) 2025-12-04T09:30:13.3617457Z Downloading nvidia_cudnn_frontend-1.16.0-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl (2.0 MB) 2025-12-04T09:30:13.3761376Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/2.0 MB ? eta -:--:-- 2025-12-04T09:30:13.3762338Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.0/2.0 MB 143.2 MB/s 0:00:00 2025-12-04T09:30:13.3815812Z [?25hDownloading nvidia_cutlass_dsl-4.3.2-cp312-cp312-manylinux_2_28_x86_64.whl (58.6 MB) 2025-12-04T09:30:13.5839801Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/58.6 MB ? eta -:--:-- 2025-12-04T09:30:13.6698171Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 58.5/58.6 MB 298.9 MB/s eta 0:00:01 2025-12-04T09:30:13.6699263Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 58.6/58.6 MB 204.2 MB/s 0:00:00 2025-12-04T09:30:13.6727321Z [?25hDownloading cuda_python-13.0.3-py3-none-any.whl (7.6 kB) 2025-12-04T09:30:13.6819245Z Downloading cuda_bindings-13.0.3-cp312-cp312-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl (12.1 MB) 2025-12-04T09:30:13.7468902Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/12.1 MB ? eta -:--:-- 2025-12-04T09:30:13.7469890Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 12.1/12.1 MB 190.2 MB/s 0:00:00 2025-12-04T09:30:13.7506720Z [?25hDownloading openai-2.8.1-py3-none-any.whl (1.0 MB) 2025-12-04T09:30:13.7599336Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.0 MB ? eta -:--:-- 2025-12-04T09:30:13.7600307Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.0/1.0 MB 112.3 MB/s 0:00:00 2025-12-04T09:30:13.7631473Z [?25hDownloading jiter-0.12.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (361 kB) 2025-12-04T09:30:13.7690246Z Using cached typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2025-12-04T09:30:13.7749356Z Downloading openai_harmony-0.0.8-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (3.0 MB) 2025-12-04T09:30:13.7932190Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.0 MB ? eta -:--:-- 2025-12-04T09:30:13.7932959Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.0/3.0 MB 171.9 MB/s 0:00:00 2025-12-04T09:30:13.7964116Z [?25hDownloading opencv_python_headless-4.12.0.88-cp37-abi3-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (54.0 MB) 2025-12-04T09:30:13.9822801Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/54.0 MB ? eta -:--:-- 2025-12-04T09:30:13.9823499Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 54.0/54.0 MB 293.6 MB/s 0:00:00 2025-12-04T09:30:13.9859293Z [?25hDownloading numpy-2.2.6-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (16.5 MB) 2025-12-04T09:30:14.0733738Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/16.5 MB ? eta -:--:-- 2025-12-04T09:30:14.0734413Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 16.5/16.5 MB 192.0 MB/s 0:00:00 2025-12-04T09:30:14.0743959Z [?25hUsing cached packaging-25.0-py3-none-any.whl (66 kB) 2025-12-04T09:30:14.0771906Z Downloading prometheus_client-0.23.1-py3-none-any.whl (61 kB) 2025-12-04T09:30:14.0835518Z Downloading prometheus_fastapi_instrumentator-7.1.0-py3-none-any.whl (19 kB) 2025-12-04T09:30:14.0906480Z Downloading pydantic_extra_types-2.10.6-py3-none-any.whl (40 kB) 2025-12-04T09:30:14.0987714Z Downloading pycountry-24.6.1-py3-none-any.whl (6.3 MB) 2025-12-04T09:30:14.1270485Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/6.3 MB ? eta -:--:-- 2025-12-04T09:30:14.1271686Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 6.3/6.3 MB 236.2 MB/s 0:00:00 2025-12-04T09:30:14.1300375Z [?25hDownloading python_multipart-0.0.20-py3-none-any.whl (24 kB) 2025-12-04T09:30:14.1365124Z Downloading pyzmq-27.1.0-cp312-abi3-manylinux_2_26_x86_64.manylinux_2_28_x86_64.whl (840 kB) 2025-12-04T09:30:14.1449131Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/841.0 kB ? eta -:--:-- 2025-12-04T09:30:14.1450142Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 841.0/841.0 kB 99.1 MB/s 0:00:00 2025-12-04T09:30:14.1479217Z [?25hDownloading ray-2.52.1-cp312-cp312-manylinux2014_x86_64.whl (72.3 MB) 2025-12-04T09:30:14.3506969Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/72.3 MB ? eta -:--:-- 2025-12-04T09:30:14.5519758Z  ━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━ 36.2/72.3 MB 181.3 MB/s eta 0:00:01 2025-12-04T09:30:14.6656857Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 72.1/72.3 MB 180.1 MB/s eta 0:00:01 2025-12-04T09:30:14.6657606Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 72.3/72.3 MB 139.9 MB/s 0:00:00 2025-12-04T09:30:14.6693346Z [?25hDownloading click-8.2.1-py3-none-any.whl (102 kB) 2025-12-04T09:30:14.6762847Z Downloading referencing-0.37.0-py3-none-any.whl (26 kB) 2025-12-04T09:30:14.6826566Z Downloading rich_toolkit-0.17.0-py3-none-any.whl (31 kB) 2025-12-04T09:30:14.6948119Z Downloading rich-14.2.0-py3-none-any.whl (243 kB) 2025-12-04T09:30:14.7034321Z Downloading markdown_it_py-4.0.0-py3-none-any.whl (87 kB) 2025-12-04T09:30:14.7098657Z Downloading mdurl-0.1.2-py3-none-any.whl (10.0 kB) 2025-12-04T09:30:14.7194257Z Downloading rignore-0.7.6-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (959 kB) 2025-12-04T09:30:14.7288048Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/959.8 kB ? eta -:--:-- 2025-12-04T09:30:14.7289402Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 959.8/959.8 kB 111.7 MB/s 0:00:00 2025-12-04T09:30:14.7319742Z [?25hDownloading rpds_py-0.30.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (394 kB) 2025-12-04T09:30:14.7401690Z Downloading sentry_sdk-2.47.0-py2.py3-none-any.whl (411 kB) 2025-12-04T09:30:14.7494477Z Downloading tiktoken-0.12.0-cp312-cp312-manylinux_2_28_x86_64.whl (1.2 MB) 2025-12-04T09:30:14.7588328Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.2 MB ? eta -:--:-- 2025-12-04T09:30:14.7589291Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.2/1.2 MB 132.8 MB/s 0:00:00 2025-12-04T09:30:14.7636534Z [?25hDownloading regex-2025.11.3-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl (803 kB) 2025-12-04T09:30:14.7720445Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/803.5 kB ? eta -:--:-- 2025-12-04T09:30:14.7721474Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 803.5/803.5 kB 96.6 MB/s 0:00:00 2025-12-04T09:30:14.7751344Z [?25hDownloading tokenizers-0.22.1-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (3.3 MB) 2025-12-04T09:30:14.7924790Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.3 MB ? eta -:--:-- 2025-12-04T09:30:14.7925783Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.3/3.3 MB 202.1 MB/s 0:00:00 2025-12-04T09:30:14.7960220Z [?25hDownloading hf_xet-1.2.0-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (3.3 MB) 2025-12-04T09:30:14.8134326Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.3 MB ? eta -:--:-- 2025-12-04T09:30:14.8135273Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.3/3.3 MB 208.6 MB/s 0:00:00 2025-12-04T09:30:14.8166848Z [?25hDownloading transformers-4.57.3-py3-none-any.whl (12.0 MB) 2025-12-04T09:30:14.8628553Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/12.0 MB ? eta -:--:-- 2025-12-04T09:30:14.8629535Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 12.0/12.0 MB 269.6 MB/s 0:00:00 2025-12-04T09:30:14.8660928Z [?25hDownloading huggingface_hub-0.36.0-py3-none-any.whl (566 kB) 2025-12-04T09:30:14.8740770Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/566.1 kB ? eta -:--:-- 2025-12-04T09:30:14.8741763Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 566.1/566.1 kB 58.8 MB/s 0:00:00 2025-12-04T09:30:14.8770459Z [?25hDownloading safetensors-0.7.0-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (507 kB) 2025-12-04T09:30:14.8851984Z Downloading typer-0.20.0-py3-none-any.whl (47 kB) 2025-12-04T09:30:14.8913474Z Downloading shellingham-1.5.4-py2.py3-none-any.whl (9.8 kB) 2025-12-04T09:30:14.8971733Z Downloading typing_inspection-0.4.2-py3-none-any.whl (14 kB) 2025-12-04T09:30:14.9033857Z Downloading uvicorn-0.38.0-py3-none-any.whl (68 kB) 2025-12-04T09:30:14.9095733Z Downloading httptools-0.7.1-cp312-cp312-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl (517 kB) 2025-12-04T09:30:14.9187287Z Downloading python_dotenv-1.2.1-py3-none-any.whl (21 kB) 2025-12-04T09:30:14.9254767Z Downloading uvloop-0.22.1-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl (4.4 MB) 2025-12-04T09:30:14.9465930Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/4.4 MB ? eta -:--:-- 2025-12-04T09:30:14.9466885Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 4.4/4.4 MB 224.1 MB/s 0:00:00 2025-12-04T09:30:14.9493335Z [?25hDownloading watchfiles-1.1.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (456 kB) 2025-12-04T09:30:14.9571389Z Downloading websockets-15.0.1-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (182 kB) 2025-12-04T09:30:14.9641275Z Downloading astor-0.8.1-py2.py3-none-any.whl (27 kB) 2025-12-04T09:30:14.9720740Z Downloading blake3-1.0.8-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (388 kB) 2025-12-04T09:30:14.9800748Z Downloading cbor2-5.7.1-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl (285 kB) 2025-12-04T09:30:14.9882568Z Downloading cloudpickle-3.1.2-py3-none-any.whl (22 kB) 2025-12-04T09:30:14.9963479Z Downloading cupy_cuda12x-13.6.0-cp312-cp312-manylinux2014_x86_64.whl (112.9 MB) 2025-12-04T09:30:15.1994031Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/112.9 MB ? eta -:--:-- 2025-12-04T09:30:15.4006539Z  ━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━ 57.7/112.9 MB 290.4 MB/s eta 0:00:01 2025-12-04T09:30:15.5636732Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 112.7/112.9 MB 286.1 MB/s eta 0:00:01 2025-12-04T09:30:15.5637777Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 112.9/112.9 MB 199.5 MB/s 0:00:00 2025-12-04T09:30:15.5669716Z [?25hDownloading fastrlock-0.8.3-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_28_x86_64.whl (53 kB) 2025-12-04T09:30:15.5756050Z Downloading einops-0.8.1-py3-none-any.whl (64 kB) 2025-12-04T09:30:15.5840505Z Downloading frozendict-2.4.7-py3-none-any.whl (16 kB) 2025-12-04T09:30:15.5926390Z Downloading msgspec-0.20.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl (224 kB) 2025-12-04T09:30:15.6034890Z Downloading nvidia_ml_py-13.580.82-py3-none-any.whl (49 kB) 2025-12-04T09:30:15.6117050Z Downloading partial_json_parser-0.2.1.1.post7-py3-none-any.whl (10 kB) 2025-12-04T09:30:15.6198356Z Downloading py_cpuinfo-9.0.0-py3-none-any.whl (22 kB) 2025-12-04T09:30:15.6282906Z Downloading pybase64-1.4.2-cp312-cp312-manylinux1_x86_64.manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_5_x86_64.whl (71 kB) 2025-12-04T09:30:15.6377356Z Downloading python_json_logger-4.0.0-py3-none-any.whl (15 kB) 2025-12-04T09:30:15.6465621Z Downloading sentencepiece-0.2.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl (1.4 MB) 2025-12-04T09:30:15.6626352Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.4 MB ? eta -:--:-- 2025-12-04T09:30:15.6627345Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.4/1.4 MB 83.5 MB/s 0:00:00 2025-12-04T09:30:15.6655236Z [?25hDownloading setproctitle-1.3.7-cp312-cp312-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl (32 kB) 2025-12-04T09:30:15.6737515Z Downloading sniffio-1.3.1-py3-none-any.whl (10 kB) 2025-12-04T09:30:16.0620140Z Building wheels for collected packages: flashinfer-python 2025-12-04T09:30:18.7032245Z Building wheel for flashinfer-python (pyproject.toml) ... [?25l- \ | / - \ | / - \ | / - \ | / done 2025-12-04T09:30:18.7084249Z [?25h Created wheel for flashinfer-python: filename=flashinfer_python-0.4.0-py3-none-any.whl size=6764886 sha256=42f2ff33ba680be0448900ea037cbff6d48cd2aaaf63d8eccd5e554220d0b908 2025-12-04T09:30:18.7085567Z Stored in directory: /var/lib/jenkins/.cache/pip/wheels/86/5d/0d/2752ded4190ef42ea316eafbbf63e32fd02342fe0764a0dfc6 2025-12-04T09:30:18.7169284Z Successfully built flashinfer-python 2025-12-04T09:30:19.0248523Z Installing collected packages: py-cpuinfo, nvidia-ml-py, fastrlock, websockets, uvloop, typing_extensions, sniffio, shellingham, setproctitle, sentry-sdk, sentencepiece, safetensors, rpds-py, rignore, regex, pyzmq, python-multipart, python-json-logger, python-dotenv, pycountry, pybase64, prometheus_client, partial-json-parser, packaging, outlines_core, nvidia-cudnn-frontend, numpy, msgspec, mdurl, llvmlite, llguidance, lark, jiter, interegular, httptools, hf-xet, h11, frozendict, fastar, einops, dnspython, diskcache, cuda-bindings, cloudpickle, click, cbor2, blake3, astor, annotated-types, annotated-doc, uvicorn, typing-inspection, tiktoken, referencing, pydantic-core, opencv-python-headless, numba, markdown-it-py, huggingface-hub, httpcore, gguf, email-validator, depyf, cupy-cuda12x, cuda-python, apache-tvm-ffi, anyio, watchfiles, tokenizers, starlette, rich, pydantic, nvidia-cutlass-dsl, jsonschema-specifications, httpx, typer, transformers, rich-toolkit, pydantic-extra-types, prometheus-fastapi-instrumentator, openai-harmony, openai, lm-format-enforcer, jsonschema, flashinfer-python, fastapi, xgrammar, ray, fastapi-cloud-cli, fastapi-cli, compressed-tensors, mistral_common, vllm 2025-12-04T09:30:19.1930604Z [?25l 2025-12-04T09:30:19.2710601Z  ━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  3/93 [websockets] 2025-12-04T09:30:19.2711109Z  Attempting uninstall: typing_extensions 2025-12-04T09:30:19.2766737Z ━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  3/93 [websockets] 2025-12-04T09:30:19.2767287Z  Found existing installation: typing_extensions 4.12.2 2025-12-04T09:30:19.2787534Z ━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  3/93 [websockets] 2025-12-04T09:30:19.2788012Z  Uninstalling typing_extensions-4.12.2: 2025-12-04T09:30:19.3605687Z ━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  3/93 [websockets] 2025-12-04T09:30:19.5280571Z  ━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  5/93 [typing_extensions] 2025-12-04T09:30:19.6955534Z  ━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  5/93 [typing_extensions] 2025-12-04T09:30:19.7671926Z  ━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  5/93 [typing_extensions] 2025-12-04T09:30:19.7672486Z  Successfully uninstalled typing_extensions-4.12.2 2025-12-04T09:30:19.8631304Z ━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  5/93 [typing_extensions] 2025-12-04T09:30:20.0346959Z  ━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  9/93 [sentry-sdk] 2025-12-04T09:30:20.2142877Z  ━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  9/93 [sentry-sdk] 2025-12-04T09:30:20.3817784Z  ━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 14/93 [regex] 2025-12-04T09:30:20.5494827Z  ━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 19/93 [pycountry] 2025-12-04T09:30:20.6271453Z  ━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 19/93 [pycountry] 2025-12-04T09:30:20.6271931Z  Attempting uninstall: packaging 2025-12-04T09:30:20.6316666Z ━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 19/93 [pycountry] 2025-12-04T09:30:20.6317199Z  Found existing installation: packaging 23.1 2025-12-04T09:30:20.6349411Z ━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 19/93 [pycountry] 2025-12-04T09:30:20.6349880Z  Uninstalling packaging-23.1: 2025-12-04T09:30:20.6451100Z ━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 19/93 [pycountry] 2025-12-04T09:30:20.6451610Z  Successfully uninstalled packaging-23.1 2025-12-04T09:30:20.7169420Z ━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 19/93 [pycountry] 2025-12-04T09:30:20.7883763Z  ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 24/93 [outlines_core] 2025-12-04T09:30:20.7884290Z  Attempting uninstall: numpy 2025-12-04T09:30:20.7964978Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 24/93 [outlines_core] 2025-12-04T09:30:20.7965494Z  Found existing installation: numpy 1.26.2 2025-12-04T09:30:20.8532945Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 24/93 [outlines_core] 2025-12-04T09:30:20.8533466Z  Uninstalling numpy-1.26.2: 2025-12-04T09:30:20.8847680Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 24/93 [outlines_core] 2025-12-04T09:30:21.0522999Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/93 [numpy] 2025-12-04T09:30:21.2003428Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/93 [numpy] 2025-12-04T09:30:21.2004008Z  Successfully uninstalled numpy-1.26.2 2025-12-04T09:30:21.2196686Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/93 [numpy] 2025-12-04T09:30:21.3872872Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/93 [numpy] 2025-12-04T09:30:21.5554959Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/93 [numpy] 2025-12-04T09:30:21.7916888Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/93 [numpy] 2025-12-04T09:30:21.9812053Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/93 [numpy] 2025-12-04T09:30:22.1505189Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/93 [numpy] 2025-12-04T09:30:22.3190220Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/93 [numpy] 2025-12-04T09:30:22.4866919Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/93 [numpy] 2025-12-04T09:30:22.6541705Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/93 [numpy] 2025-12-04T09:30:22.7619432Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/93 [numpy] 2025-12-04T09:30:22.7619909Z  Attempting uninstall: llvmlite 2025-12-04T09:30:22.7660151Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/93 [numpy] 2025-12-04T09:30:22.7660710Z  Found existing installation: llvmlite 0.43.0 2025-12-04T09:30:22.7712957Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/93 [numpy] 2025-12-04T09:30:22.7713420Z  Uninstalling llvmlite-0.43.0: 2025-12-04T09:30:22.8216849Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/93 [numpy] 2025-12-04T09:30:22.8565231Z  ━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━ 29/93 [llvmlite] 2025-12-04T09:30:22.8566031Z  Successfully uninstalled llvmlite-0.43.0 2025-12-04T09:30:22.9891822Z ━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━ 29/93 [llvmlite] 2025-12-04T09:30:23.1568364Z  ━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━ 29/93 [llvmlite] 2025-12-04T09:30:23.3243000Z  ━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━ 29/93 [llvmlite] 2025-12-04T09:30:23.4918752Z  ━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━ 29/93 [llvmlite] 2025-12-04T09:30:23.6435901Z  ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 30/93 [llguidance] 2025-12-04T09:30:23.6436369Z  Attempting uninstall: lark 2025-12-04T09:30:23.6476081Z ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 30/93 [llguidance] 2025-12-04T09:30:23.6476565Z  Found existing installation: lark 0.12.0 2025-12-04T09:30:23.6534783Z ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 30/93 [llguidance] 2025-12-04T09:30:23.6535238Z  Uninstalling lark-0.12.0: 2025-12-04T09:30:23.6591862Z ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 30/93 [llguidance] 2025-12-04T09:30:23.6789583Z  ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 31/93 [lark] 2025-12-04T09:30:23.6790062Z  Successfully uninstalled lark-0.12.0 2025-12-04T09:30:23.8268112Z ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 31/93 [lark] 2025-12-04T09:30:23.9961217Z  ━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━ 35/93 [hf-xet] 2025-12-04T09:30:24.1614974Z  ━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━ 40/93 [dnspython] 2025-12-04T09:30:24.1615458Z  Attempting uninstall: cuda-bindings 2025-12-04T09:30:24.1647524Z ━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━ 40/93 [dnspython] 2025-12-04T09:30:24.1656227Z  ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━ 42/93 [cuda-bindings] 2025-12-04T09:30:24.1656776Z  Found existing installation: cuda-bindings 12.9.4 2025-12-04T09:30:24.1719661Z ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━ 42/93 [cuda-bindings] 2025-12-04T09:30:24.1720125Z  Uninstalling cuda-bindings-12.9.4: 2025-12-04T09:30:24.3324728Z ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━ 42/93 [cuda-bindings] 2025-12-04T09:30:24.4999953Z  ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━ 42/93 [cuda-bindings] 2025-12-04T09:30:24.6676229Z  ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━ 42/93 [cuda-bindings] 2025-12-04T09:30:24.7206012Z  ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━ 42/93 [cuda-bindings] 2025-12-04T09:30:24.7206877Z  Successfully uninstalled cuda-bindings-12.9.4 2025-12-04T09:30:24.8353004Z ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━ 42/93 [cuda-bindings] 2025-12-04T09:30:24.9955763Z  ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━ 42/93 [cuda-bindings] 2025-12-04T09:30:24.9956227Z  Attempting uninstall: click 2025-12-04T09:30:24.9989987Z ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━ 42/93 [cuda-bindings] 2025-12-04T09:30:24.9990805Z  Found existing installation: click 8.3.1 2025-12-04T09:30:25.0024676Z ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━ 42/93 [cuda-bindings] 2025-12-04T09:30:25.0025105Z  Uninstalling click-8.3.1: 2025-12-04T09:30:25.0032648Z ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━ 42/93 [cuda-bindings] 2025-12-04T09:30:25.0153124Z  ━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━ 44/93 [click] 2025-12-04T09:30:25.0153604Z  Successfully uninstalled click-8.3.1 2025-12-04T09:30:25.1710485Z ━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━ 44/93 [click] 2025-12-04T09:30:25.3385658Z  ━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━ 50/93 [uvicorn] 2025-12-04T09:30:25.5059802Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/93 [opencv-python-headless] 2025-12-04T09:30:25.6734526Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/93 [opencv-python-headless] 2025-12-04T09:30:25.8409011Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/93 [opencv-python-headless] 2025-12-04T09:30:25.8998087Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/93 [opencv-python-headless] 2025-12-04T09:30:25.8998593Z  Attempting uninstall: numba 2025-12-04T09:30:25.9042963Z ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/93 [opencv-python-headless] 2025-12-04T09:30:25.9043637Z  Found existing installation: numba 0.60.0 2025-12-04T09:30:25.9759085Z ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/93 [opencv-python-headless] 2025-12-04T09:30:25.9759558Z  Uninstalling numba-0.60.0: 2025-12-04T09:30:26.0083617Z ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/93 [opencv-python-headless] 2025-12-04T09:30:26.1759363Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 56/93 [numba] 2025-12-04T09:30:26.3434487Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 56/93 [numba] 2025-12-04T09:30:26.3616568Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 56/93 [numba] 2025-12-04T09:30:26.3617086Z  Successfully uninstalled numba-0.60.0 2025-12-04T09:30:26.5109416Z ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 56/93 [numba] 2025-12-04T09:30:26.6836698Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 56/93 [numba] 2025-12-04T09:30:26.8529941Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 56/93 [numba] 2025-12-04T09:30:27.0221221Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 56/93 [numba] 2025-12-04T09:30:27.1906849Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 56/93 [numba] 2025-12-04T09:30:27.3595493Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 56/93 [numba] 2025-12-04T09:30:27.5272474Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 56/93 [numba] 2025-12-04T09:30:27.6948826Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 56/93 [numba] 2025-12-04T09:30:27.8712160Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 56/93 [numba] 2025-12-04T09:30:28.0421393Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 56/93 [numba] 2025-12-04T09:30:28.2095828Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 56/93 [numba] 2025-12-04T09:30:28.3774465Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 56/93 [numba] 2025-12-04T09:30:28.5455782Z  ━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━ 58/93 [huggingface-hub] 2025-12-04T09:30:28.7137171Z  ━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━ 58/93 [huggingface-hub] 2025-12-04T09:30:28.8814254Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━ 61/93 [email-validator] 2025-12-04T09:30:29.0488472Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 63/93 [cupy-cuda12x] 2025-12-04T09:30:29.2164850Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 63/93 [cupy-cuda12x] 2025-12-04T09:30:29.3839054Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 63/93 [cupy-cuda12x] 2025-12-04T09:30:29.5513511Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 63/93 [cupy-cuda12x] 2025-12-04T09:30:29.7188237Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 63/93 [cupy-cuda12x] 2025-12-04T09:30:29.8862591Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 63/93 [cupy-cuda12x] 2025-12-04T09:30:30.0538714Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 63/93 [cupy-cuda12x] 2025-12-04T09:30:30.2213455Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 63/93 [cupy-cuda12x] 2025-12-04T09:30:30.3897596Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 63/93 [cupy-cuda12x] 2025-12-04T09:30:30.5573200Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 63/93 [cupy-cuda12x] 2025-12-04T09:30:30.7248627Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━ 65/93 [apache-tvm-ffi] 2025-12-04T09:30:30.8925517Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━ 68/93 [tokenizers] 2025-12-04T09:30:31.0601954Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 70/93 [rich] 2025-12-04T09:30:31.2277631Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 71/93 [pydantic] 2025-12-04T09:30:31.3957238Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 72/93 [nvidia-cutlass-dsl] 2025-12-04T09:30:31.5629954Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 72/93 [nvidia-cutlass-dsl] 2025-12-04T09:30:31.7305037Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 72/93 [nvidia-cutlass-dsl] 2025-12-04T09:30:31.9006706Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 72/93 [nvidia-cutlass-dsl] 2025-12-04T09:30:32.0689840Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 72/93 [nvidia-cutlass-dsl] 2025-12-04T09:30:32.2368841Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 72/93 [nvidia-cutlass-dsl] 2025-12-04T09:30:32.4048883Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 72/93 [nvidia-cutlass-dsl] 2025-12-04T09:30:32.5722189Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:32.7397285Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:32.9101077Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:33.0777559Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:33.2464422Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:33.4181652Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:33.5862321Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:33.7549887Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:33.9299364Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:34.1003585Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:34.2690967Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:34.4398224Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:34.6089832Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:34.7779649Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:34.9510491Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:35.1190404Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:35.2882292Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:35.4561048Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:35.6240933Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:35.7973281Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:35.9648204Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:36.1323902Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:36.3000618Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:36.4675513Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:36.6381444Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:36.8094610Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:36.9829181Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:37.1508400Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:37.3183160Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:37.4902431Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:37.6577847Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:37.8254352Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:37.9932495Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:38.1607655Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 76/93 [transformers] 2025-12-04T09:30:38.3283047Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━ 77/93 [rich-toolkit] 2025-12-04T09:30:38.4958063Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━ 80/93 [openai-harmony] 2025-12-04T09:30:38.6641069Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 81/93 [openai] 2025-12-04T09:30:38.8316286Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 81/93 [openai] 2025-12-04T09:30:38.9990685Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 81/93 [openai] 2025-12-04T09:30:39.1668707Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 81/93 [openai] 2025-12-04T09:30:39.3344308Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 81/93 [openai] 2025-12-04T09:30:39.5020225Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━ 83/93 [jsonschema] 2025-12-04T09:30:39.6694748Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 84/93 [flashinfer-python] 2025-12-04T09:30:39.8434480Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 84/93 [flashinfer-python] 2025-12-04T09:30:40.0154212Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 84/93 [flashinfer-python] 2025-12-04T09:30:40.1829020Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 84/93 [flashinfer-python] 2025-12-04T09:30:40.3508442Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 84/93 [flashinfer-python] 2025-12-04T09:30:40.5188330Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 84/93 [flashinfer-python] 2025-12-04T09:30:40.6868666Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 84/93 [flashinfer-python] 2025-12-04T09:30:40.8548824Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━ 85/93 [fastapi] 2025-12-04T09:30:41.0224345Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━ 86/93 [xgrammar] 2025-12-04T09:30:41.1899297Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:41.3576285Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:41.5252124Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:41.6928187Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:41.8603804Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:42.0313375Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:42.1990873Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:42.3667734Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:42.5367257Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:42.7045809Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:42.8720838Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:43.0404442Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:43.2085073Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:43.3759813Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:43.5435653Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:43.7121466Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:43.8796419Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:44.0536597Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:44.2211401Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:44.3933692Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:44.5608213Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:44.7284713Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 87/93 [ray] 2025-12-04T09:30:44.8961209Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━ 90/93 [compressed-tensors] 2025-12-04T09:30:45.0634157Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:45.2308856Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:45.3982791Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:45.5657105Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:45.7331484Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:45.9006615Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:46.0680535Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:46.2354947Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:46.4029979Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:46.5704168Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:46.7379291Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:46.9055331Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:47.0730678Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:47.2406024Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:47.4080803Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:47.5754618Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:47.7429019Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:47.9103547Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:48.0778270Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:48.2452357Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:48.4126521Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:48.5801004Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:48.7474978Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:48.9149931Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:49.0824242Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:49.2498718Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:49.4172628Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:49.5847575Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:49.7522026Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:49.9196382Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:50.0870164Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:50.2545312Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:50.4251986Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:50.5930482Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:50.7604931Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:50.9302956Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:51.1003596Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:51.2687752Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:51.4364812Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:51.6057134Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:51.7748494Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:51.8190394Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 92/93 [vllm] 2025-12-04T09:30:51.8190982Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 93/93 [vllm] 2025-12-04T09:30:51.8191742Z [?25h 2025-12-04T09:30:51.8341747Z Successfully installed annotated-doc-0.0.4 annotated-types-0.7.0 anyio-4.12.0 apache-tvm-ffi-0.1.0b15 astor-0.8.1 blake3-1.0.8 cbor2-5.7.1 click-8.2.1 cloudpickle-3.1.2 compressed-tensors-0.11.0 cuda-bindings-13.0.3 cuda-python-13.0.3 cupy-cuda12x-13.6.0 depyf-0.19.0 diskcache-5.6.3 dnspython-2.8.0 einops-0.8.1 email-validator-2.3.0 fastapi-0.123.7 fastapi-cli-0.0.16 fastapi-cloud-cli-0.5.2 fastar-0.8.0 fastrlock-0.8.3 flashinfer-python-0.4.0 frozendict-2.4.7 gguf-0.17.1 h11-0.16.0 hf-xet-1.2.0 httpcore-1.0.9 httptools-0.7.1 httpx-0.28.1 huggingface-hub-0.36.0 interegular-0.3.3 jiter-0.12.0 jsonschema-4.25.1 jsonschema-specifications-2025.9.1 lark-1.2.2 llguidance-0.7.30 llvmlite-0.44.0 lm-format-enforcer-0.11.3 markdown-it-py-4.0.0 mdurl-0.1.2 mistral_common-1.8.6 msgspec-0.20.0 numba-0.61.2 numpy-2.2.6 nvidia-cudnn-frontend-1.16.0 nvidia-cutlass-dsl-4.3.2 nvidia-ml-py-13.580.82 openai-2.8.1 openai-harmony-0.0.8 opencv-python-headless-4.12.0.88 outlines_core-0.2.11 packaging-25.0 partial-json-parser-0.2.1.1.post7 prometheus-fastapi-instrumentator-7.1.0 prometheus_client-0.23.1 py-cpuinfo-9.0.0 pybase64-1.4.2 pycountry-24.6.1 pydantic-2.12.5 pydantic-core-2.41.5 pydantic-extra-types-2.10.6 python-dotenv-1.2.1 python-json-logger-4.0.0 python-multipart-0.0.20 pyzmq-27.1.0 ray-2.52.1 referencing-0.37.0 regex-2025.11.3 rich-14.2.0 rich-toolkit-0.17.0 rignore-0.7.6 rpds-py-0.30.0 safetensors-0.7.0 sentencepiece-0.2.1 sentry-sdk-2.47.0 setproctitle-1.3.7 shellingham-1.5.4 sniffio-1.3.1 starlette-0.50.0 tiktoken-0.12.0 tokenizers-0.22.1 transformers-4.57.3 typer-0.20.0 typing-inspection-0.4.2 typing_extensions-4.15.0 uvicorn-0.38.0 uvloop-0.22.1 vllm-0.11.0rc2.dev389+ge51928192.d20251204 watchfiles-1.1.1 websockets-15.0.1 xgrammar-0.1.25 2025-12-04T09:30:52.7172714Z 2025-12-04 09:30:52,716 [INFO] cli.lib.core.vllm.vllm_test: Done. Installed vllm wheels 2025-12-04T09:30:52.7174160Z 2025-12-04 09:30:52,717 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install -e tests/vllm_test_utils 2025-12-04T09:30:52.7175420Z 2025-12-04 09:30:52,717 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install -e tests/vllm_test_utils 2025-12-04T09:30:52.7942132Z Using Python 3.12.5 environment at: /opt/conda/envs/py_3.12 2025-12-04T09:30:53.4960328Z Resolved 1 package in 664ms 2025-12-04T09:30:53.4967944Z Building vllm-test-utils @ file:///var/lib/jenkins/workspace/vllm/tests/vllm_test_utils 2025-12-04T09:30:53.6730881Z Built vllm-test-utils @ file:///var/lib/jenkins/workspace/vllm/tests/vllm_test_utils 2025-12-04T09:30:53.6769223Z Prepared 1 package in 180ms 2025-12-04T09:30:53.6790835Z Installed 1 package in 2ms 2025-12-04T09:30:53.6800192Z + vllm-test-utils==0.1 (from file:///var/lib/jenkins/workspace/vllm/tests/vllm_test_utils) 2025-12-04T09:30:53.6974889Z 2025-12-04 09:30:53,697 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install hf_transfer 2025-12-04T09:30:53.6976245Z 2025-12-04 09:30:53,697 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install hf_transfer 2025-12-04T09:30:53.7372969Z Using Python 3.12.5 environment at: /opt/conda/envs/py_3.12 2025-12-04T09:30:53.7632716Z Resolved 1 package in 19ms 2025-12-04T09:30:53.7665639Z Downloading hf-transfer (3.4MiB) 2025-12-04T09:30:53.7984514Z Downloading hf-transfer 2025-12-04T09:30:53.7988037Z Prepared 1 package in 35ms 2025-12-04T09:30:53.8269650Z Installed 1 package in 28ms 2025-12-04T09:30:53.8270222Z + hf-transfer==0.1.9 2025-12-04T09:30:53.8305743Z 2025-12-04 09:30:53,830 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python use_existing_torch.py 2025-12-04T09:30:53.8538248Z >>> cleaning requirements/dev.txt 2025-12-04T09:30:53.8538729Z <<< done cleaning requirements/dev.txt 2025-12-04T09:30:53.8539024Z 2025-12-04T09:30:53.8539243Z >>> cleaning requirements/lint.txt 2025-12-04T09:30:53.8539583Z <<< done cleaning requirements/lint.txt 2025-12-04T09:30:53.8539923Z 2025-12-04T09:30:53.8540090Z >>> cleaning requirements/build.txt 2025-12-04T09:30:53.8540517Z removed: 2025-12-04T09:30:53.8541215Z <<< done cleaning requirements/build.txt 2025-12-04T09:30:53.8541500Z 2025-12-04T09:30:53.8541690Z >>> cleaning requirements/common.txt 2025-12-04T09:30:53.8542243Z <<< done cleaning requirements/common.txt 2025-12-04T09:30:53.8542618Z 2025-12-04T09:30:53.8542795Z >>> cleaning requirements/cpu-build.txt 2025-12-04T09:30:53.8543269Z removed: 2025-12-04T09:30:53.8543633Z <<< done cleaning requirements/cpu-build.txt 2025-12-04T09:30:53.8543972Z 2025-12-04T09:30:53.8544131Z >>> cleaning requirements/cpu.txt 2025-12-04T09:30:53.8544565Z removed: 2025-12-04T09:30:53.8544910Z <<< done cleaning requirements/cpu.txt 2025-12-04T09:30:53.8545227Z 2025-12-04T09:30:53.8545390Z >>> cleaning requirements/cuda.txt 2025-12-04T09:30:53.8545799Z removed: 2025-12-04T09:30:53.8546198Z <<< done cleaning requirements/cuda.txt 2025-12-04T09:30:53.8546451Z 2025-12-04T09:30:53.8546575Z >>> cleaning requirements/docs.txt 2025-12-04T09:30:53.8546892Z removed: 2025-12-04T09:30:53.8547655Z <<< done cleaning requirements/docs.txt 2025-12-04T09:30:53.8548168Z 2025-12-04T09:30:53.8548292Z >>> cleaning requirements/kv_connectors.txt 2025-12-04T09:30:53.8548581Z <<< done cleaning requirements/kv_connectors.txt 2025-12-04T09:30:53.8548756Z 2025-12-04T09:30:53.8548857Z >>> cleaning requirements/nightly_torch_test.txt 2025-12-04T09:30:53.8549227Z <<< done cleaning requirements/nightly_torch_test.txt 2025-12-04T09:30:53.8549412Z 2025-12-04T09:30:53.8549559Z >>> cleaning requirements/rocm-build.txt 2025-12-04T09:30:53.8549788Z removed: 2025-12-04T09:30:53.8550756Z <<< done cleaning requirements/rocm-build.txt 2025-12-04T09:30:53.8550992Z 2025-12-04T09:30:53.8551073Z >>> cleaning requirements/rocm-test.txt 2025-12-04T09:30:53.8551402Z <<< done cleaning requirements/rocm-test.txt 2025-12-04T09:30:53.8551560Z 2025-12-04T09:30:53.8551639Z >>> cleaning requirements/rocm.txt 2025-12-04T09:30:53.8552101Z <<< done cleaning requirements/rocm.txt 2025-12-04T09:30:53.8552289Z 2025-12-04T09:30:53.8552369Z >>> cleaning requirements/test.txt 2025-12-04T09:30:53.8554555Z removed: 2025-12-04T09:30:53.8558157Z <<< done cleaning requirements/test.txt 2025-12-04T09:30:53.8558323Z 2025-12-04T09:30:53.8558422Z >>> cleaning requirements/tpu.txt 2025-12-04T09:30:53.8558717Z removed: 2025-12-04T09:30:53.8559939Z <<< done cleaning requirements/tpu.txt 2025-12-04T09:30:53.8560102Z 2025-12-04T09:30:53.8560187Z >>> cleaning requirements/xpu.txt 2025-12-04T09:30:53.8560481Z removed: 2025-12-04T09:30:53.8561753Z <<< done cleaning requirements/xpu.txt 2025-12-04T09:30:53.8561905Z 2025-12-04T09:30:53.8562009Z >>> cleaning pyproject.toml 2025-12-04T09:30:53.8562299Z removed: 2025-12-04T09:30:53.8564252Z <<< done cleaning pyproject.toml 2025-12-04T09:30:53.8564505Z 2025-12-04T09:30:53.8597044Z 2025-12-04 09:30:53,859 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install -r requirements/common.txt 2025-12-04T09:30:53.8598451Z 2025-12-04 09:30:53,859 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install -r requirements/common.txt 2025-12-04T09:30:53.8948903Z Using Python 3.12.5 environment at: /opt/conda/envs/py_3.12 2025-12-04T09:30:53.9113727Z Audited 50 packages in 20ms 2025-12-04T09:30:53.9134980Z 2025-12-04 09:30:53,913 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install -r requirements/build.txt 2025-12-04T09:30:53.9136837Z 2025-12-04 09:30:53,913 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install -r requirements/build.txt 2025-12-04T09:30:53.9448184Z Using Python 3.12.5 environment at: /opt/conda/envs/py_3.12 2025-12-04T09:30:54.0159554Z Resolved 11 packages in 64ms 2025-12-04T09:30:54.0295683Z Prepared 1 package in 13ms 2025-12-04T09:30:54.0363536Z Installed 1 package in 6ms 2025-12-04T09:30:54.0364245Z + setuptools-scm==9.2.2 2025-12-04T09:30:54.0413956Z 2025-12-04 09:30:54,041 [INFO] cli.lib.core.vllm.vllm_test: generate test.txt from requirements/test.in with local torch whls 2025-12-04T09:30:54.4127338Z 2025-12-04 09:30:54,412 [INFO] cli.lib.core.vllm.vllm_test: [INFO] Updated requirements/test.in 2025-12-04T09:30:54.4132251Z 2025-12-04 09:30:54,412 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip compile requirements/test.in -o test.txt --index-strategy unsafe-best-match --constraint snapshot_constraint.txt --torch-backend cu128 2025-12-04T09:30:54.4675298Z Updating https://github.com/IBM/terratorch.git (1.1.rc3) 2025-12-04T09:30:56.7835351Z Updated https://github.com/IBM/terratorch.git (07184fcf91a1324f831ff521dd238d97fe350e3e) 2025-12-04T09:30:59.8506039Z Resolved 329 packages in 5.40s 2025-12-04T09:30:59.8506700Z # This file was autogenerated by uv via the following command: 2025-12-04T09:30:59.8507825Z # uv pip compile requirements/test.in -o test.txt --index-strategy unsafe-best-match --constraint snapshot_constraint.txt --torch-backend cu128 2025-12-04T09:30:59.8736722Z absl-py==2.1.0 2025-12-04T09:30:59.8737207Z  # via 2025-12-04T09:30:59.8737476Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8738000Z # rouge-score 2025-12-04T09:30:59.8738418Z accelerate==1.0.1 2025-12-04T09:30:59.8738810Z  # via 2025-12-04T09:30:59.8739044Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8739424Z # lm-eval 2025-12-04T09:30:59.8739817Z # peft 2025-12-04T09:30:59.8740017Z aenum==3.1.16 2025-12-04T09:30:59.8740235Z  # via 2025-12-04T09:30:59.8740596Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8741119Z # lightly 2025-12-04T09:30:59.8741473Z affine==2.4.0 2025-12-04T09:30:59.8741697Z  # via 2025-12-04T09:30:59.8741856Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8742206Z # rasterio 2025-12-04T09:30:59.8742530Z aiohappyeyeballs==2.6.1 2025-12-04T09:30:59.8742909Z  # via 2025-12-04T09:30:59.8743199Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8743624Z # aiohttp 2025-12-04T09:30:59.8743922Z aiohttp==3.13.0 2025-12-04T09:30:59.8744245Z  # via 2025-12-04T09:30:59.8744552Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8744918Z # aiohttp-cors 2025-12-04T09:30:59.8745184Z # datasets 2025-12-04T09:30:59.8745439Z # fsspec 2025-12-04T09:30:59.8745680Z # gpt-oss 2025-12-04T09:30:59.8745963Z # lm-eval 2025-12-04T09:30:59.8746172Z # ray 2025-12-04T09:30:59.8756263Z aiohttp-cors==0.8.1 2025-12-04T09:30:59.8756650Z  # via 2025-12-04T09:30:59.8756941Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8757358Z # ray 2025-12-04T09:30:59.8757626Z aiosignal==1.4.0 2025-12-04T09:30:59.8757835Z  # via 2025-12-04T09:30:59.8758012Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8758250Z # aiohttp 2025-12-04T09:30:59.8758444Z albucore==0.0.16 2025-12-04T09:30:59.8758643Z  # via 2025-12-04T09:30:59.8758831Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8759115Z # terratorch 2025-12-04T09:30:59.8759442Z albumentations==1.4.6 2025-12-04T09:30:59.8759752Z  # via 2025-12-04T09:30:59.8760028Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8760430Z # terratorch 2025-12-04T09:30:59.8760693Z alembic==1.16.4 2025-12-04T09:30:59.8760988Z  # via 2025-12-04T09:30:59.8761523Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8761951Z # mlflow 2025-12-04T09:30:59.8762224Z annotated-types==0.7.0 2025-12-04T09:30:59.8762571Z  # via 2025-12-04T09:30:59.8762862Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8763284Z # pydantic 2025-12-04T09:30:59.8763475Z antlr4-python3-runtime==4.9.3 2025-12-04T09:30:59.8763692Z  # via 2025-12-04T09:30:59.8763874Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8764225Z # hydra-core 2025-12-04T09:30:59.8764570Z # omegaconf 2025-12-04T09:30:59.8764833Z anyio==4.6.2.post1 2025-12-04T09:30:59.8765156Z  # via 2025-12-04T09:30:59.8765382Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8765661Z # httpx 2025-12-04T09:30:59.8765953Z # starlette 2025-12-04T09:30:59.8766228Z argcomplete==3.5.1 2025-12-04T09:30:59.8766468Z  # via 2025-12-04T09:30:59.8766731Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8767131Z # datamodel-code-generator 2025-12-04T09:30:59.8767490Z arrow==1.3.0 2025-12-04T09:30:59.8767763Z  # via 2025-12-04T09:30:59.8768009Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8768438Z # isoduration 2025-12-04T09:30:59.8768748Z attrs==24.2.0 2025-12-04T09:30:59.8769066Z  # via 2025-12-04T09:30:59.8769335Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8769626Z # aiohttp 2025-12-04T09:30:59.8769875Z # fiona 2025-12-04T09:30:59.8770381Z # hypothesis 2025-12-04T09:30:59.8770645Z # jsonlines 2025-12-04T09:30:59.8770903Z # jsonschema 2025-12-04T09:30:59.8771330Z # pytest-subtests 2025-12-04T09:30:59.8771577Z # rasterio 2025-12-04T09:30:59.8771938Z # referencing 2025-12-04T09:30:59.8772261Z audioread==3.0.1 2025-12-04T09:30:59.8772535Z  # via 2025-12-04T09:30:59.8772820Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8773211Z # librosa 2025-12-04T09:30:59.8773434Z backoff==2.2.1 2025-12-04T09:30:59.8773666Z  # via 2025-12-04T09:30:59.8773932Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8774203Z # -r requirements/test.in 2025-12-04T09:30:59.8774533Z # schemathesis 2025-12-04T09:30:59.8774840Z bitsandbytes==0.46.1 2025-12-04T09:30:59.8775174Z  # via 2025-12-04T09:30:59.8775420Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8775770Z # -r requirements/test.in 2025-12-04T09:30:59.8776146Z # lightning 2025-12-04T09:30:59.8776407Z black==24.10.0 2025-12-04T09:30:59.8776679Z  # via 2025-12-04T09:30:59.8776884Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8777230Z # datamodel-code-generator 2025-12-04T09:30:59.8777605Z blinker==1.9.0 2025-12-04T09:30:59.8777885Z  # via 2025-12-04T09:30:59.8778106Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8778394Z # flask 2025-12-04T09:30:59.8778663Z blobfile==3.0.0 2025-12-04T09:30:59.8778979Z  # via 2025-12-04T09:30:59.8779271Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8779700Z # -r requirements/test.in 2025-12-04T09:30:59.8780038Z bm25s==0.2.13 2025-12-04T09:30:59.8780357Z  # via 2025-12-04T09:30:59.8780527Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8780937Z # mteb 2025-12-04T09:30:59.8781213Z boto3==1.35.57 2025-12-04T09:30:59.8781447Z  # via 2025-12-04T09:30:59.8781723Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8782069Z # runai-model-streamer-s3 2025-12-04T09:30:59.8782461Z # tensorizer 2025-12-04T09:30:59.8782691Z botocore==1.35.57 2025-12-04T09:30:59.8783000Z  # via 2025-12-04T09:30:59.8783160Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8783432Z # boto3 2025-12-04T09:30:59.8783749Z # s3transfer 2025-12-04T09:30:59.8784016Z bounded-pool-executor==0.0.3 2025-12-04T09:30:59.8784365Z  # via 2025-12-04T09:30:59.8784641Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8785016Z # pqdm 2025-12-04T09:30:59.8785492Z buildkite-test-collector==0.1.9 2025-12-04T09:30:59.8785848Z  # via 2025-12-04T09:30:59.8786087Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8786436Z # -r requirements/test.in 2025-12-04T09:30:59.8786682Z cachetools==5.5.2 2025-12-04T09:30:59.8786935Z  # via 2025-12-04T09:30:59.8787189Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8787430Z # google-auth 2025-12-04T09:30:59.8787679Z # mlflow-skinny 2025-12-04T09:30:59.8788154Z certifi==2024.8.30 2025-12-04T09:30:59.8788454Z  # via 2025-12-04T09:30:59.8788741Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8789067Z # fiona 2025-12-04T09:30:59.8789318Z # httpcore 2025-12-04T09:30:59.8789559Z # httpx 2025-12-04T09:30:59.8789784Z # lightly 2025-12-04T09:30:59.8790008Z # pyogrio 2025-12-04T09:30:59.8790164Z # pyproj 2025-12-04T09:30:59.8790304Z # rasterio 2025-12-04T09:30:59.8790493Z # requests 2025-12-04T09:30:59.8790658Z cffi==1.17.1 2025-12-04T09:30:59.8790837Z  # via 2025-12-04T09:30:59.8791069Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8791408Z # soundfile 2025-12-04T09:30:59.8791686Z chardet==5.2.0 2025-12-04T09:30:59.8791934Z  # via 2025-12-04T09:30:59.8792208Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8792595Z # mbstrdecoder 2025-12-04T09:30:59.8792890Z charset-normalizer==3.4.0 2025-12-04T09:30:59.8793151Z  # via 2025-12-04T09:30:59.8793482Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8793720Z # requests 2025-12-04T09:30:59.8793882Z chz==0.3.0 2025-12-04T09:30:59.8794057Z  # via 2025-12-04T09:30:59.8794332Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8794740Z # gpt-oss 2025-12-04T09:30:59.8795020Z click==8.1.7 2025-12-04T09:30:59.8795307Z  # via 2025-12-04T09:30:59.8795519Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8795772Z # black 2025-12-04T09:30:59.8795973Z # click-plugins 2025-12-04T09:30:59.8796261Z # cligj 2025-12-04T09:30:59.8796494Z # fiona 2025-12-04T09:30:59.8796633Z # flask 2025-12-04T09:30:59.8796773Z # jiwer 2025-12-04T09:30:59.8796918Z # mlflow-skinny 2025-12-04T09:30:59.8797070Z # nltk 2025-12-04T09:30:59.8797214Z # rasterio 2025-12-04T09:30:59.8797423Z # ray 2025-12-04T09:30:59.8797581Z # schemathesis 2025-12-04T09:30:59.8797856Z # typer 2025-12-04T09:30:59.8798175Z # uvicorn 2025-12-04T09:30:59.8798426Z click-plugins==1.1.1.2 2025-12-04T09:30:59.8798690Z  # via 2025-12-04T09:30:59.8798921Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8799238Z # fiona 2025-12-04T09:30:59.8799496Z # rasterio 2025-12-04T09:30:59.8799683Z cligj==0.7.2 2025-12-04T09:30:59.8799972Z  # via 2025-12-04T09:30:59.8800236Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8800549Z # fiona 2025-12-04T09:30:59.8800837Z # rasterio 2025-12-04T09:30:59.8801131Z cloudpickle==3.1.1 2025-12-04T09:30:59.8801455Z  # via 2025-12-04T09:30:59.8801690Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8802036Z # mlflow-skinny 2025-12-04T09:30:59.8802282Z colorama==0.4.6 2025-12-04T09:30:59.8802570Z  # via 2025-12-04T09:30:59.8802819Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8803093Z # sacrebleu 2025-12-04T09:30:59.8803327Z # schemathesis 2025-12-04T09:30:59.8803539Z # tqdm-multiprocess 2025-12-04T09:30:59.8803820Z colorful==0.5.6 2025-12-04T09:30:59.8804002Z  # via 2025-12-04T09:30:59.8804177Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8804473Z # ray 2025-12-04T09:30:59.8804630Z contourpy==1.3.0 2025-12-04T09:30:59.8804799Z  # via 2025-12-04T09:30:59.8805050Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8805321Z # matplotlib 2025-12-04T09:30:59.8805562Z coverage==7.10.6 2025-12-04T09:30:59.8805756Z  # via 2025-12-04T09:30:59.8806101Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8806364Z # pytest-cov 2025-12-04T09:30:59.8806636Z cramjam==2.9.0 2025-12-04T09:30:59.8806910Z  # via 2025-12-04T09:30:59.8807126Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8807481Z # fastparquet 2025-12-04T09:30:59.8807784Z cupy-cuda12x==13.6.0 2025-12-04T09:30:59.8808028Z  # via 2025-12-04T09:30:59.8808191Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8808414Z # ray 2025-12-04T09:30:59.8808571Z cycler==0.12.1 2025-12-04T09:30:59.8808810Z  # via 2025-12-04T09:30:59.8809031Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8809376Z # matplotlib 2025-12-04T09:30:59.8809639Z databricks-sdk==0.59.0 2025-12-04T09:30:59.8809952Z  # via 2025-12-04T09:30:59.8810160Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8810491Z # mlflow-skinny 2025-12-04T09:30:59.8810692Z datamodel-code-generator==0.26.3 2025-12-04T09:30:59.8810911Z  # via 2025-12-04T09:30:59.8811076Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8811332Z # -r requirements/test.in 2025-12-04T09:30:59.8811535Z dataproperty==1.0.1 2025-12-04T09:30:59.8811725Z  # via 2025-12-04T09:30:59.8811884Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8812075Z # pytablewriter 2025-12-04T09:30:59.8812281Z # tabledata 2025-12-04T09:30:59.8812449Z datasets==3.0.2 2025-12-04T09:30:59.8812619Z  # via 2025-12-04T09:30:59.8812883Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8813073Z # evaluate 2025-12-04T09:30:59.8813227Z # lm-eval 2025-12-04T09:30:59.8813405Z # mteb 2025-12-04T09:30:59.8813565Z decorator==5.1.1 2025-12-04T09:30:59.8813739Z  # via 2025-12-04T09:30:59.8813891Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8814111Z # librosa 2025-12-04T09:30:59.8814273Z decord==0.6.0 2025-12-04T09:30:59.8814434Z  # via 2025-12-04T09:30:59.8814599Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8814849Z # -r requirements/test.in 2025-12-04T09:30:59.8815044Z dill==0.3.8 2025-12-04T09:30:59.8815212Z  # via 2025-12-04T09:30:59.8815379Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8815562Z # datasets 2025-12-04T09:30:59.8815712Z # evaluate 2025-12-04T09:30:59.8815860Z # lm-eval 2025-12-04T09:30:59.8816040Z # multiprocess 2025-12-04T09:30:59.8816222Z distlib==0.3.9 2025-12-04T09:30:59.8816396Z  # via 2025-12-04T09:30:59.8816558Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8816786Z # virtualenv 2025-12-04T09:30:59.8816958Z dnspython==2.7.0 2025-12-04T09:30:59.8817129Z  # via 2025-12-04T09:30:59.8817290Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8817539Z # email-validator 2025-12-04T09:30:59.8817725Z docker==7.1.0 2025-12-04T09:30:59.8817890Z  # via 2025-12-04T09:30:59.8818049Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8818241Z # gpt-oss 2025-12-04T09:30:59.8818420Z # mlflow 2025-12-04T09:30:59.8818587Z docopt==0.6.2 2025-12-04T09:30:59.8818760Z  # via 2025-12-04T09:30:59.8818913Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8819248Z # num2words 2025-12-04T09:30:59.8819432Z docstring-parser==0.17.0 2025-12-04T09:30:59.8819632Z  # via 2025-12-04T09:30:59.8819794Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8820020Z # jsonargparse 2025-12-04T09:30:59.8820215Z efficientnet-pytorch==0.7.1 2025-12-04T09:30:59.8820481Z  # via segmentation-models-pytorch 2025-12-04T09:30:59.8820713Z einops==0.8.1 2025-12-04T09:30:59.8820875Z  # via 2025-12-04T09:30:59.8821035Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8821240Z # -r requirements/test.in 2025-12-04T09:30:59.8821419Z # encodec 2025-12-04T09:30:59.8821571Z # terratorch 2025-12-04T09:30:59.8821724Z # torchgeo 2025-12-04T09:30:59.8821884Z # vector-quantize-pytorch 2025-12-04T09:30:59.8822206Z # vocos 2025-12-04T09:30:59.8822374Z einx==0.3.0 2025-12-04T09:30:59.8822539Z  # via 2025-12-04T09:30:59.8822716Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8822967Z # vector-quantize-pytorch 2025-12-04T09:30:59.8823190Z email-validator==2.2.0 2025-12-04T09:30:59.8823377Z  # via 2025-12-04T09:30:59.8823539Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8823762Z # pydantic 2025-12-04T09:30:59.8823927Z encodec==0.1.1 2025-12-04T09:30:59.8824108Z  # via 2025-12-04T09:30:59.8824268Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8824479Z # vocos 2025-12-04T09:30:59.8824644Z eval-type-backport==0.2.2 2025-12-04T09:30:59.8824856Z  # via 2025-12-04T09:30:59.8825018Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8825239Z # mteb 2025-12-04T09:30:59.8825398Z evaluate==0.4.3 2025-12-04T09:30:59.8825565Z  # via 2025-12-04T09:30:59.8825724Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8825950Z # lm-eval 2025-12-04T09:30:59.8826111Z fastapi==0.116.1 2025-12-04T09:30:59.8826290Z  # via 2025-12-04T09:30:59.8826461Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8826650Z # gpt-oss 2025-12-04T09:30:59.8826841Z # mlflow-skinny 2025-12-04T09:30:59.8827029Z fastparquet==2024.11.0 2025-12-04T09:30:59.8827213Z  # via 2025-12-04T09:30:59.8827371Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8827726Z # genai-perf 2025-12-04T09:30:59.8827900Z fastrlock==0.8.2 2025-12-04T09:30:59.8828081Z  # via 2025-12-04T09:30:59.8828244Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8828477Z # cupy-cuda12x 2025-12-04T09:30:59.8828668Z fastsafetensors==0.1.10 2025-12-04T09:30:59.8828860Z  # via 2025-12-04T09:30:59.8829018Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8829250Z # -r requirements/test.in 2025-12-04T09:30:59.8829450Z filelock==3.16.1 2025-12-04T09:30:59.8829628Z  # via 2025-12-04T09:30:59.8829788Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8829975Z # blobfile 2025-12-04T09:30:59.8830121Z # datasets 2025-12-04T09:30:59.8830272Z # huggingface-hub 2025-12-04T09:30:59.8830435Z # ray 2025-12-04T09:30:59.8830577Z # torch 2025-12-04T09:30:59.8830718Z # transformers 2025-12-04T09:30:59.8830912Z # virtualenv 2025-12-04T09:30:59.8831071Z fiona==1.10.1 2025-12-04T09:30:59.8831237Z  # via 2025-12-04T09:30:59.8831403Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8831626Z # torchgeo 2025-12-04T09:30:59.8831783Z flask==3.1.1 2025-12-04T09:30:59.8831960Z  # via 2025-12-04T09:30:59.8832118Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8832327Z # mlflow 2025-12-04T09:30:59.8832486Z fonttools==4.55.0 2025-12-04T09:30:59.8832661Z  # via 2025-12-04T09:30:59.8832816Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8833039Z # matplotlib 2025-12-04T09:30:59.8833212Z fqdn==1.5.1 2025-12-04T09:30:59.8833378Z  # via 2025-12-04T09:30:59.8833529Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8833753Z # jsonschema 2025-12-04T09:30:59.8833919Z frozendict==2.4.6 2025-12-04T09:30:59.8834091Z  # via 2025-12-04T09:30:59.8834249Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8834474Z # einx 2025-12-04T09:30:59.8834622Z frozenlist==1.5.0 2025-12-04T09:30:59.8834801Z  # via 2025-12-04T09:30:59.8834971Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8835157Z # aiohttp 2025-12-04T09:30:59.8835341Z # aiosignal 2025-12-04T09:30:59.8835508Z fsspec==2024.9.0 2025-12-04T09:30:59.8835678Z  # via 2025-12-04T09:30:59.8835838Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8836029Z # datasets 2025-12-04T09:30:59.8836168Z # evaluate 2025-12-04T09:30:59.8836333Z # fastparquet 2025-12-04T09:30:59.8836498Z # huggingface-hub 2025-12-04T09:30:59.8836748Z # lightning 2025-12-04T09:30:59.8836921Z # pytorch-lightning 2025-12-04T09:30:59.8837142Z # torch 2025-12-04T09:30:59.8837289Z ftfy==6.3.1 2025-12-04T09:30:59.8837452Z  # via 2025-12-04T09:30:59.8837608Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8837841Z # open-clip-torch 2025-12-04T09:30:59.8838023Z genai-perf==0.0.8 2025-12-04T09:30:59.8838210Z  # via 2025-12-04T09:30:59.8838368Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8838604Z # -r requirements/test.in 2025-12-04T09:30:59.8838807Z genson==1.3.0 2025-12-04T09:30:59.8838991Z  # via 2025-12-04T09:30:59.8839149Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8839402Z # datamodel-code-generator 2025-12-04T09:30:59.8839623Z geopandas==1.0.1 2025-12-04T09:30:59.8839794Z  # via 2025-12-04T09:30:59.8839952Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8840178Z # terratorch 2025-12-04T09:30:59.8840345Z gitdb==4.0.12 2025-12-04T09:30:59.8840521Z  # via 2025-12-04T09:30:59.8840679Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8840908Z # gitpython 2025-12-04T09:30:59.8841075Z gitpython==3.1.44 2025-12-04T09:30:59.8841252Z  # via 2025-12-04T09:30:59.8841403Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8841648Z # mlflow-skinny 2025-12-04T09:30:59.8841837Z google-api-core==2.24.2 2025-12-04T09:30:59.8842025Z  # via 2025-12-04T09:30:59.8842285Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8842482Z # google-cloud-core 2025-12-04T09:30:59.8842669Z # google-cloud-storage 2025-12-04T09:30:59.8842892Z # opencensus 2025-12-04T09:30:59.8843066Z google-auth==2.40.2 2025-12-04T09:30:59.8843251Z  # via 2025-12-04T09:30:59.8843419Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8843612Z # databricks-sdk 2025-12-04T09:30:59.8843785Z # google-api-core 2025-12-04T09:30:59.8843957Z # google-cloud-core 2025-12-04T09:30:59.8844139Z # google-cloud-storage 2025-12-04T09:30:59.8844387Z # runai-model-streamer-gcs 2025-12-04T09:30:59.8844605Z google-cloud-core==2.4.3 2025-12-04T09:30:59.8844803Z  # via 2025-12-04T09:30:59.8844959Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8845199Z # google-cloud-storage 2025-12-04T09:30:59.8845404Z google-cloud-storage==3.4.0 2025-12-04T09:30:59.8845611Z  # via 2025-12-04T09:30:59.8845772Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8846014Z # runai-model-streamer-gcs 2025-12-04T09:30:59.8846216Z google-crc32c==1.7.1 2025-12-04T09:30:59.8846396Z  # via 2025-12-04T09:30:59.8846556Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8846758Z # google-cloud-storage 2025-12-04T09:30:59.8846998Z # google-resumable-media 2025-12-04T09:30:59.8847218Z google-resumable-media==2.7.2 2025-12-04T09:30:59.8847426Z  # via 2025-12-04T09:30:59.8847599Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8847840Z # google-cloud-storage 2025-12-04T09:30:59.8848051Z googleapis-common-protos==1.70.0 2025-12-04T09:30:59.8848274Z  # via 2025-12-04T09:30:59.8848435Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8848662Z # google-api-core 2025-12-04T09:30:59.8848851Z gpt-oss==0.0.8 2025-12-04T09:30:59.8849033Z  # via 2025-12-04T09:30:59.8849189Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8849437Z # -r requirements/test.in 2025-12-04T09:30:59.8849643Z graphene==3.4.3 2025-12-04T09:30:59.8849821Z  # via 2025-12-04T09:30:59.8849982Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8850203Z # mlflow 2025-12-04T09:30:59.8850368Z graphql-core==3.2.6 2025-12-04T09:30:59.8850541Z  # via 2025-12-04T09:30:59.8850713Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8850906Z # graphene 2025-12-04T09:30:59.8851058Z # graphql-relay 2025-12-04T09:30:59.8851359Z # hypothesis-graphql 2025-12-04T09:30:59.8851567Z graphql-relay==3.2.0 2025-12-04T09:30:59.8851759Z  # via 2025-12-04T09:30:59.8851920Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8852143Z # graphene 2025-12-04T09:30:59.8852302Z greenlet==3.2.3 2025-12-04T09:30:59.8852477Z  # via 2025-12-04T09:30:59.8852648Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8852869Z # sqlalchemy 2025-12-04T09:30:59.8853041Z grpcio==1.71.0 2025-12-04T09:30:59.8853230Z  # via 2025-12-04T09:30:59.8853390Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8853609Z # ray 2025-12-04T09:30:59.8853764Z gunicorn==23.0.0 2025-12-04T09:30:59.8853930Z  # via 2025-12-04T09:30:59.8854094Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8854314Z # mlflow 2025-12-04T09:30:59.8854464Z h11==0.14.0 2025-12-04T09:30:59.8854629Z  # via 2025-12-04T09:30:59.8854791Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8854982Z # httpcore 2025-12-04T09:30:59.8855166Z # uvicorn 2025-12-04T09:30:59.8855328Z h5py==3.13.0 2025-12-04T09:30:59.8855494Z  # via 2025-12-04T09:30:59.8855652Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8855885Z # terratorch 2025-12-04T09:30:59.8856064Z harfile==0.3.0 2025-12-04T09:30:59.8856231Z  # via 2025-12-04T09:30:59.8856389Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8856616Z # schemathesis 2025-12-04T09:30:59.8856878Z hf-xet==1.1.7 2025-12-04T09:30:59.8857051Z  # via 2025-12-04T09:30:59.8857215Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8857444Z # huggingface-hub 2025-12-04T09:30:59.8857633Z hiredis==3.0.0 2025-12-04T09:30:59.8857801Z  # via 2025-12-04T09:30:59.8857953Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8858187Z # tensorizer 2025-12-04T09:30:59.8858356Z html2text==2025.4.15 2025-12-04T09:30:59.8858531Z  # via 2025-12-04T09:30:59.8858689Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8858915Z # gpt-oss 2025-12-04T09:30:59.8859134Z httpcore==1.0.6 2025-12-04T09:30:59.8859317Z  # via 2025-12-04T09:30:59.8859475Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8859687Z # httpx 2025-12-04T09:30:59.8859856Z httpx==0.27.2 2025-12-04T09:30:59.8860030Z  # via 2025-12-04T09:30:59.8860189Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8860397Z # -r requirements/test.in 2025-12-04T09:30:59.8860633Z # schemathesis 2025-12-04T09:30:59.8860821Z huggingface-hub==0.34.3 2025-12-04T09:30:59.8861008Z  # via 2025-12-04T09:30:59.8861172Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8861369Z # accelerate 2025-12-04T09:30:59.8861516Z # datasets 2025-12-04T09:30:59.8861669Z # evaluate 2025-12-04T09:30:59.8861826Z # open-clip-torch 2025-12-04T09:30:59.8861988Z # peft 2025-12-04T09:30:59.8862159Z # segmentation-models-pytorch 2025-12-04T09:30:59.8862382Z # sentence-transformers 2025-12-04T09:30:59.8862586Z # terratorch 2025-12-04T09:30:59.8862736Z # timm 2025-12-04T09:30:59.8862879Z # tokenizers 2025-12-04T09:30:59.8863029Z # transformers 2025-12-04T09:30:59.8863220Z # vocos 2025-12-04T09:30:59.8863372Z humanize==4.11.0 2025-12-04T09:30:59.8863540Z  # via 2025-12-04T09:30:59.8863707Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8863952Z # runai-model-streamer 2025-12-04T09:30:59.8864149Z hydra-core==1.3.2 2025-12-04T09:30:59.8864324Z  # via 2025-12-04T09:30:59.8864491Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8864684Z # lightly 2025-12-04T09:30:59.8864862Z # lightning 2025-12-04T09:30:59.8865032Z hypothesis==6.131.0 2025-12-04T09:30:59.8865211Z  # via 2025-12-04T09:30:59.8865363Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8865575Z # hypothesis-graphql 2025-12-04T09:30:59.8865771Z # hypothesis-jsonschema 2025-12-04T09:30:59.8866078Z # schemathesis 2025-12-04T09:30:59.8866277Z hypothesis-graphql==0.11.1 2025-12-04T09:30:59.8866494Z  # via 2025-12-04T09:30:59.8866648Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8866878Z # schemathesis 2025-12-04T09:30:59.8867067Z hypothesis-jsonschema==0.23.1 2025-12-04T09:30:59.8867274Z  # via 2025-12-04T09:30:59.8867439Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8867667Z # schemathesis 2025-12-04T09:30:59.8867847Z idna==3.10 2025-12-04T09:30:59.8868018Z  # via 2025-12-04T09:30:59.8868176Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8868361Z # anyio 2025-12-04T09:30:59.8868510Z # email-validator 2025-12-04T09:30:59.8868679Z # httpx 2025-12-04T09:30:59.8868821Z # jsonschema 2025-12-04T09:30:59.8868985Z # requests 2025-12-04T09:30:59.8869162Z # yarl 2025-12-04T09:30:59.8869317Z imageio==2.37.0 2025-12-04T09:30:59.8869495Z  # via 2025-12-04T09:30:59.8869659Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8869904Z # scikit-image 2025-12-04T09:30:59.8870093Z importlib-metadata==8.7.0 2025-12-04T09:30:59.8870297Z  # via 2025-12-04T09:30:59.8870460Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8870651Z # mlflow-skinny 2025-12-04T09:30:59.8870871Z # opentelemetry-api 2025-12-04T09:30:59.8871380Z importlib-resources==6.5.2 2025-12-04T09:30:59.8871591Z  # via 2025-12-04T09:30:59.8872006Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8872390Z # typeshed-client 2025-12-04T09:30:59.8872640Z inflect==5.6.2 2025-12-04T09:30:59.8872820Z  # via 2025-12-04T09:30:59.8872979Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8873222Z # datamodel-code-generator 2025-12-04T09:30:59.8873436Z iniconfig==2.0.0 2025-12-04T09:30:59.8873616Z  # via 2025-12-04T09:30:59.8873772Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8873991Z # pytest 2025-12-04T09:30:59.8874162Z isoduration==20.11.0 2025-12-04T09:30:59.8874343Z  # via 2025-12-04T09:30:59.8874502Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8874723Z # jsonschema 2025-12-04T09:30:59.8874890Z isort==5.13.2 2025-12-04T09:30:59.8875051Z  # via 2025-12-04T09:30:59.8875207Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8875453Z # datamodel-code-generator 2025-12-04T09:30:59.8875657Z itsdangerous==2.2.0 2025-12-04T09:30:59.8875849Z  # via 2025-12-04T09:30:59.8876010Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8876223Z # flask 2025-12-04T09:30:59.8876376Z jinja2==3.1.6 2025-12-04T09:30:59.8876544Z  # via 2025-12-04T09:30:59.8876695Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8876904Z # datamodel-code-generator 2025-12-04T09:30:59.8877108Z # flask 2025-12-04T09:30:59.8877248Z # mlflow 2025-12-04T09:30:59.8877420Z # torch 2025-12-04T09:30:59.8877574Z jiwer==3.0.5 2025-12-04T09:30:59.8877740Z  # via 2025-12-04T09:30:59.8877897Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8878142Z # -r requirements/test.in 2025-12-04T09:30:59.8878354Z jmespath==1.0.1 2025-12-04T09:30:59.8878532Z  # via 2025-12-04T09:30:59.8878692Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8878877Z # boto3 2025-12-04T09:30:59.8879057Z # botocore 2025-12-04T09:30:59.8879227Z joblib==1.4.2 2025-12-04T09:30:59.8879404Z  # via 2025-12-04T09:30:59.8879566Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8879763Z # librosa 2025-12-04T09:30:59.8879917Z # nltk 2025-12-04T09:30:59.8880094Z # scikit-learn 2025-12-04T09:30:59.8880281Z jsonargparse==4.35.0 2025-12-04T09:30:59.8880467Z  # via 2025-12-04T09:30:59.8880622Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8880816Z # lightning 2025-12-04T09:30:59.8881010Z # terratorch 2025-12-04T09:30:59.8881174Z jsonlines==4.0.0 2025-12-04T09:30:59.8881504Z  # via 2025-12-04T09:30:59.8881687Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8881911Z # lm-eval 2025-12-04T09:30:59.8882075Z jsonpointer==3.0.0 2025-12-04T09:30:59.8882261Z  # via 2025-12-04T09:30:59.8882417Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8882639Z # jsonschema 2025-12-04T09:30:59.8882807Z jsonschema==4.23.0 2025-12-04T09:30:59.8882977Z  # via 2025-12-04T09:30:59.8883136Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8883346Z # hypothesis-jsonschema 2025-12-04T09:30:59.8883535Z # mistral-common 2025-12-04T09:30:59.8883703Z # ray 2025-12-04T09:30:59.8883883Z # schemathesis 2025-12-04T09:30:59.8884073Z jsonschema-specifications==2024.10.1 2025-12-04T09:30:59.8884304Z  # via 2025-12-04T09:30:59.8884465Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8884694Z # jsonschema 2025-12-04T09:30:59.8884859Z junit-xml==1.9 2025-12-04T09:30:59.8885028Z  # via 2025-12-04T09:30:59.8885192Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8885424Z # schemathesis 2025-12-04T09:30:59.8885600Z kaleido==0.2.1 2025-12-04T09:30:59.8885770Z  # via 2025-12-04T09:30:59.8885925Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8886152Z # genai-perf 2025-12-04T09:30:59.8886321Z kiwisolver==1.4.7 2025-12-04T09:30:59.8886498Z  # via 2025-12-04T09:30:59.8886658Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8887769Z # matplotlib 2025-12-04T09:30:59.8887934Z kornia==0.8.1 2025-12-04T09:30:59.8888105Z  # via 2025-12-04T09:30:59.8888270Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8888486Z # torchgeo 2025-12-04T09:30:59.8888662Z kornia-rs==0.1.9 2025-12-04T09:30:59.8888850Z  # via 2025-12-04T09:30:59.8889004Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8889229Z # kornia 2025-12-04T09:30:59.8889392Z lazy-loader==0.4 2025-12-04T09:30:59.8889560Z  # via 2025-12-04T09:30:59.8889727Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8889929Z # librosa 2025-12-04T09:30:59.8890115Z # scikit-image 2025-12-04T09:30:59.8890284Z libnacl==2.1.0 2025-12-04T09:30:59.8890452Z  # via 2025-12-04T09:30:59.8890620Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8890838Z # tensorizer 2025-12-04T09:30:59.8891007Z librosa==0.10.2.post1 2025-12-04T09:30:59.8891190Z  # via 2025-12-04T09:30:59.8891349Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8891589Z # -r requirements/test.in 2025-12-04T09:30:59.8891791Z lightly==1.5.20 2025-12-04T09:30:59.8891969Z  # via 2025-12-04T09:30:59.8892128Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8892316Z # terratorch 2025-12-04T09:30:59.8892493Z # torchgeo 2025-12-04T09:30:59.8892659Z lightly-utils==0.0.2 2025-12-04T09:30:59.8892845Z  # via 2025-12-04T09:30:59.8892996Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8893218Z # lightly 2025-12-04T09:30:59.8893395Z lightning==2.5.1.post0 2025-12-04T09:30:59.8893577Z  # via 2025-12-04T09:30:59.8893734Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8893926Z # terratorch 2025-12-04T09:30:59.8894102Z # torchgeo 2025-12-04T09:30:59.8894278Z lightning-utilities==0.14.3 2025-12-04T09:30:59.8894487Z  # via 2025-12-04T09:30:59.8894647Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8894840Z # lightning 2025-12-04T09:30:59.8895003Z # pytorch-lightning 2025-12-04T09:30:59.8895214Z # torchmetrics 2025-12-04T09:30:59.8895384Z llvmlite==0.44.0 2025-12-04T09:30:59.8895561Z  # via 2025-12-04T09:30:59.8895719Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8895933Z # numba 2025-12-04T09:30:59.8896331Z lm-eval @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d 2025-12-04T09:30:59.8896788Z  # via 2025-12-04T09:30:59.8897032Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8897304Z # -r requirements/test.in 2025-12-04T09:30:59.8897511Z lxml==5.3.0 2025-12-04T09:30:59.8897675Z  # via 2025-12-04T09:30:59.8897836Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8898036Z # blobfile 2025-12-04T09:30:59.8898188Z # gpt-oss 2025-12-04T09:30:59.8898370Z # sacrebleu 2025-12-04T09:30:59.8898535Z mako==1.3.10 2025-12-04T09:30:59.8898698Z  # via 2025-12-04T09:30:59.8898862Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8899173Z # alembic 2025-12-04T09:30:59.8899351Z markdown==3.8.2 2025-12-04T09:30:59.8899534Z  # via 2025-12-04T09:30:59.8899695Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8899910Z # mlflow 2025-12-04T09:30:59.8900069Z markdown-it-py==3.0.0 2025-12-04T09:30:59.8900259Z  # via 2025-12-04T09:30:59.8900416Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8900635Z # rich 2025-12-04T09:30:59.8900797Z markupsafe==3.0.1 2025-12-04T09:30:59.8900972Z  # via 2025-12-04T09:30:59.8901130Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8901321Z # flask 2025-12-04T09:30:59.8901462Z # jinja2 2025-12-04T09:30:59.8901600Z # mako 2025-12-04T09:30:59.8901771Z # werkzeug 2025-12-04T09:30:59.8901950Z matplotlib==3.9.2 2025-12-04T09:30:59.8902121Z  # via 2025-12-04T09:30:59.8902279Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8902579Z # -r requirements/test.in 2025-12-04T09:30:59.8902763Z # lightning 2025-12-04T09:30:59.8902911Z # mlflow 2025-12-04T09:30:59.8903057Z # pycocotools 2025-12-04T09:30:59.8903251Z # torchgeo 2025-12-04T09:30:59.8903422Z mbstrdecoder==1.1.3 2025-12-04T09:30:59.8903601Z  # via 2025-12-04T09:30:59.8903753Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8903953Z # dataproperty 2025-12-04T09:30:59.8904119Z # pytablewriter 2025-12-04T09:30:59.8904308Z # typepy 2025-12-04T09:30:59.8904461Z mdurl==0.1.2 2025-12-04T09:30:59.8904629Z  # via 2025-12-04T09:30:59.8904789Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8905024Z # markdown-it-py 2025-12-04T09:30:59.8905217Z mistral-common==1.8.5 2025-12-04T09:30:59.8905407Z  # via 2025-12-04T09:30:59.8905560Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8905812Z # -r requirements/test.in 2025-12-04T09:30:59.8906020Z mlflow==2.22.0 2025-12-04T09:30:59.8906193Z  # via 2025-12-04T09:30:59.8906356Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8906582Z # terratorch 2025-12-04T09:30:59.8906751Z mlflow-skinny==2.22.0 2025-12-04T09:30:59.8906946Z  # via 2025-12-04T09:30:59.8907108Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8907319Z # mlflow 2025-12-04T09:30:59.8907481Z more-itertools==10.5.0 2025-12-04T09:30:59.8907668Z  # via 2025-12-04T09:30:59.8907821Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8908046Z # lm-eval 2025-12-04T09:30:59.8908213Z mpmath==1.3.0 2025-12-04T09:30:59.8908380Z  # via 2025-12-04T09:30:59.8908535Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8908747Z # sympy 2025-12-04T09:30:59.8908897Z msgpack==1.1.0 2025-12-04T09:30:59.8909056Z  # via 2025-12-04T09:30:59.8909206Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8909393Z # librosa 2025-12-04T09:30:59.8909565Z # ray 2025-12-04T09:30:59.8909718Z mteb==1.38.11 2025-12-04T09:30:59.8909882Z  # via 2025-12-04T09:30:59.8910030Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8910262Z # -r requirements/test.in 2025-12-04T09:30:59.8910468Z multidict==6.1.0 2025-12-04T09:30:59.8910637Z  # via 2025-12-04T09:30:59.8910791Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8910983Z # aiohttp 2025-12-04T09:30:59.8911147Z # yarl 2025-12-04T09:30:59.8911303Z multiprocess==0.70.16 2025-12-04T09:30:59.8911487Z  # via 2025-12-04T09:30:59.8911714Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8911918Z # datasets 2025-12-04T09:30:59.8912110Z # evaluate 2025-12-04T09:30:59.8912263Z munch==4.0.0 2025-12-04T09:30:59.8912432Z  # via 2025-12-04T09:30:59.8912589Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8912817Z # pretrainedmodels 2025-12-04T09:30:59.8913012Z mypy-extensions==1.0.0 2025-12-04T09:30:59.8913199Z  # via 2025-12-04T09:30:59.8913352Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8913569Z # black 2025-12-04T09:30:59.8913734Z networkx==3.2.1 2025-12-04T09:30:59.8913909Z  # via 2025-12-04T09:30:59.8914059Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8914251Z # scikit-image 2025-12-04T09:30:59.8914443Z # torch 2025-12-04T09:30:59.8914589Z nltk==3.9.1 2025-12-04T09:30:59.8914755Z  # via 2025-12-04T09:30:59.8923275Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8923593Z # rouge-score 2025-12-04T09:30:59.8923779Z num2words==0.5.14 2025-12-04T09:30:59.8923969Z  # via 2025-12-04T09:30:59.8924147Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8924400Z # -r requirements/test.in 2025-12-04T09:30:59.8924620Z numba==0.61.2 2025-12-04T09:30:59.8924800Z  # via 2025-12-04T09:30:59.8924969Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8925187Z # -r requirements/test.in 2025-12-04T09:30:59.8925551Z # librosa 2025-12-04T09:30:59.8925710Z numexpr==2.10.1 2025-12-04T09:30:59.8925885Z  # via 2025-12-04T09:30:59.8926045Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8926264Z # lm-eval 2025-12-04T09:30:59.8926420Z numpy==1.26.4 2025-12-04T09:30:59.8926590Z  # via 2025-12-04T09:30:59.8926745Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8926962Z # -r requirements/test.in 2025-12-04T09:30:59.8927153Z # accelerate 2025-12-04T09:30:59.8927297Z # albucore 2025-12-04T09:30:59.8927456Z # albumentations 2025-12-04T09:30:59.8927620Z # bitsandbytes 2025-12-04T09:30:59.8927780Z # bm25s 2025-12-04T09:30:59.8927919Z # contourpy 2025-12-04T09:30:59.8928071Z # cupy-cuda12x 2025-12-04T09:30:59.8928237Z # datasets 2025-12-04T09:30:59.8928382Z # decord 2025-12-04T09:30:59.8928519Z # einx 2025-12-04T09:30:59.8928653Z # encodec 2025-12-04T09:30:59.8928794Z # evaluate 2025-12-04T09:30:59.8928948Z # fastparquet 2025-12-04T09:30:59.8929099Z # genai-perf 2025-12-04T09:30:59.8929254Z # geopandas 2025-12-04T09:30:59.8929404Z # h5py 2025-12-04T09:30:59.8929536Z # imageio 2025-12-04T09:30:59.8929676Z # librosa 2025-12-04T09:30:59.8929816Z # lightly 2025-12-04T09:30:59.8929957Z # lightly-utils 2025-12-04T09:30:59.8930132Z # matplotlib 2025-12-04T09:30:59.8930294Z # mistral-common 2025-12-04T09:30:59.8930450Z # mlflow 2025-12-04T09:30:59.8930593Z # mteb 2025-12-04T09:30:59.8930734Z # numba 2025-12-04T09:30:59.8930869Z # numexpr 2025-12-04T09:30:59.8931028Z # opencv-python-headless 2025-12-04T09:30:59.8931230Z # pandas 2025-12-04T09:30:59.8931372Z # patsy 2025-12-04T09:30:59.8931525Z # peft 2025-12-04T09:30:59.8931675Z # pycocotools 2025-12-04T09:30:59.8931820Z # pyogrio 2025-12-04T09:30:59.8931963Z # rasterio 2025-12-04T09:30:59.8932107Z # rioxarray 2025-12-04T09:30:59.8932255Z # rouge-score 2025-12-04T09:30:59.8932419Z # runai-model-streamer 2025-12-04T09:30:59.8932603Z # sacrebleu 2025-12-04T09:30:59.8932753Z # scikit-image 2025-12-04T09:30:59.8932912Z # scikit-learn 2025-12-04T09:30:59.8933064Z # scipy 2025-12-04T09:30:59.8933220Z # segmentation-models-pytorch 2025-12-04T09:30:59.8933428Z # shapely 2025-12-04T09:30:59.8933574Z # soxr 2025-12-04T09:30:59.8933711Z # statsmodels 2025-12-04T09:30:59.8933861Z # tensorboardx 2025-12-04T09:30:59.8934103Z # tensorizer 2025-12-04T09:30:59.8934260Z # tifffile 2025-12-04T09:30:59.8934412Z # torchgeo 2025-12-04T09:30:59.8934560Z # torchmetrics 2025-12-04T09:30:59.8934715Z # torchvision 2025-12-04T09:30:59.8934862Z # transformers 2025-12-04T09:30:59.8935008Z # tritonclient 2025-12-04T09:30:59.8935158Z # vocos 2025-12-04T09:30:59.8935327Z # xarray 2025-12-04T09:30:59.8935476Z omegaconf==2.3.0 2025-12-04T09:30:59.8935651Z  # via 2025-12-04T09:30:59.8935806Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8936000Z # hydra-core 2025-12-04T09:30:59.8936181Z # lightning 2025-12-04T09:30:59.8936347Z open-clip-torch==2.32.0 2025-12-04T09:30:59.8936580Z  # via -r requirements/test.in 2025-12-04T09:30:59.8936817Z openai-harmony==0.0.4 2025-12-04T09:30:59.8937009Z  # via 2025-12-04T09:30:59.8937177Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8937402Z # gpt-oss 2025-12-04T09:30:59.8937560Z opencensus==0.11.4 2025-12-04T09:30:59.8937736Z  # via 2025-12-04T09:30:59.8937891Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8938098Z # ray 2025-12-04T09:30:59.8938258Z opencensus-context==0.1.3 2025-12-04T09:30:59.8938459Z  # via 2025-12-04T09:30:59.8938613Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8938833Z # opencensus 2025-12-04T09:30:59.8939029Z opencv-python-headless==4.11.0.86 2025-12-04T09:30:59.8939448Z  # via 2025-12-04T09:30:59.8939609Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8939814Z # -r requirements/test.in 2025-12-04T09:30:59.8940001Z # albucore 2025-12-04T09:30:59.8940152Z # albumentations 2025-12-04T09:30:59.8940365Z # mistral-common 2025-12-04T09:30:59.8940560Z opentelemetry-api==1.35.0 2025-12-04T09:30:59.8940757Z  # via 2025-12-04T09:30:59.8940922Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8941123Z # mlflow-skinny 2025-12-04T09:30:59.8941316Z # opentelemetry-exporter-prometheus 2025-12-04T09:30:59.8941548Z # opentelemetry-sdk 2025-12-04T09:30:59.8941801Z # opentelemetry-semantic-conventions 2025-12-04T09:30:59.8942061Z opentelemetry-exporter-prometheus==0.56b0 2025-12-04T09:30:59.8942300Z  # via 2025-12-04T09:30:59.8942461Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8942673Z # ray 2025-12-04T09:30:59.8942841Z opentelemetry-proto==1.36.0 2025-12-04T09:30:59.8943061Z  # via 2025-12-04T09:30:59.8943227Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8943438Z # ray 2025-12-04T09:30:59.8943600Z opentelemetry-sdk==1.35.0 2025-12-04T09:30:59.8943811Z  # via 2025-12-04T09:30:59.8943960Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8944152Z # mlflow-skinny 2025-12-04T09:30:59.8944342Z # opentelemetry-exporter-prometheus 2025-12-04T09:30:59.8944576Z # ray 2025-12-04T09:30:59.8944758Z opentelemetry-semantic-conventions==0.56b0 2025-12-04T09:30:59.8945008Z  # via 2025-12-04T09:30:59.8945171Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8945406Z # opentelemetry-sdk 2025-12-04T09:30:59.8945596Z packaging==24.2 2025-12-04T09:30:59.8945766Z  # via 2025-12-04T09:30:59.8945916Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8946105Z # accelerate 2025-12-04T09:30:59.8946251Z # black 2025-12-04T09:30:59.8946412Z # datamodel-code-generator 2025-12-04T09:30:59.8946606Z # datasets 2025-12-04T09:30:59.8946756Z # evaluate 2025-12-04T09:30:59.8946907Z # fastparquet 2025-12-04T09:30:59.8947060Z # geopandas 2025-12-04T09:30:59.8947204Z # gunicorn 2025-12-04T09:30:59.8947358Z # huggingface-hub 2025-12-04T09:30:59.8947526Z # hydra-core 2025-12-04T09:30:59.8947676Z # kornia 2025-12-04T09:30:59.8947815Z # lazy-loader 2025-12-04T09:30:59.8947964Z # lightning 2025-12-04T09:30:59.8948124Z # lightning-utilities 2025-12-04T09:30:59.8948297Z # matplotlib 2025-12-04T09:30:59.8948553Z # mlflow-skinny 2025-12-04T09:30:59.8948726Z # peft 2025-12-04T09:30:59.8948864Z # plotly 2025-12-04T09:30:59.8949008Z # pooch 2025-12-04T09:30:59.8949155Z # pyogrio 2025-12-04T09:30:59.8949296Z # pytest 2025-12-04T09:30:59.8949452Z # pytest-rerunfailures 2025-12-04T09:30:59.8949642Z # pytorch-lightning 2025-12-04T09:30:59.8949804Z # ray 2025-12-04T09:30:59.8949943Z # rioxarray 2025-12-04T09:30:59.8950103Z # scikit-image 2025-12-04T09:30:59.8950258Z # statsmodels 2025-12-04T09:30:59.8950423Z # tensorboardx 2025-12-04T09:30:59.8950582Z # torchmetrics 2025-12-04T09:30:59.8950735Z # transformers 2025-12-04T09:30:59.8950887Z # typepy 2025-12-04T09:30:59.8951071Z # xarray 2025-12-04T09:30:59.8951215Z pandas==2.2.3 2025-12-04T09:30:59.8951381Z  # via 2025-12-04T09:30:59.8951556Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8951749Z # datasets 2025-12-04T09:30:59.8951904Z # evaluate 2025-12-04T09:30:59.8952053Z # fastparquet 2025-12-04T09:30:59.8952202Z # genai-perf 2025-12-04T09:30:59.8952355Z # geopandas 2025-12-04T09:30:59.8952501Z # mlflow 2025-12-04T09:30:59.8952637Z # statsmodels 2025-12-04T09:30:59.8952784Z # torchgeo 2025-12-04T09:30:59.8952961Z # xarray 2025-12-04T09:30:59.8953123Z pathspec==0.12.1 2025-12-04T09:30:59.8953292Z  # via 2025-12-04T09:30:59.8953542Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8953767Z # black 2025-12-04T09:30:59.8953917Z pathvalidate==3.2.1 2025-12-04T09:30:59.8954094Z  # via 2025-12-04T09:30:59.8954249Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8954471Z # pytablewriter 2025-12-04T09:30:59.8954644Z patsy==1.0.1 2025-12-04T09:30:59.8954808Z  # via 2025-12-04T09:30:59.8954956Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8955177Z # statsmodels 2025-12-04T09:30:59.8955346Z peft==0.16.0 2025-12-04T09:30:59.8955509Z  # via 2025-12-04T09:30:59.8955674Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8955874Z # -r requirements/test.in 2025-12-04T09:30:59.8956088Z # lm-eval 2025-12-04T09:30:59.8956242Z pillow==10.4.0 2025-12-04T09:30:59.8956411Z  # via 2025-12-04T09:30:59.8956559Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8956747Z # genai-perf 2025-12-04T09:30:59.8956899Z # imageio 2025-12-04T09:30:59.8957058Z # lightly-utils 2025-12-04T09:30:59.8957223Z # matplotlib 2025-12-04T09:30:59.8957381Z # mistral-common 2025-12-04T09:30:59.8957542Z # scikit-image 2025-12-04T09:30:59.8957722Z # segmentation-models-pytorch 2025-12-04T09:30:59.8957937Z # sentence-transformers 2025-12-04T09:30:59.8958122Z # torchgeo 2025-12-04T09:30:59.8958314Z # torchvision 2025-12-04T09:30:59.8958494Z platformdirs==4.3.6 2025-12-04T09:30:59.8958677Z  # via 2025-12-04T09:30:59.8958834Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8959025Z # black 2025-12-04T09:30:59.8959168Z # pooch 2025-12-04T09:30:59.8959334Z # virtualenv 2025-12-04T09:30:59.8959502Z plotly==5.24.1 2025-12-04T09:30:59.8959666Z  # via 2025-12-04T09:30:59.8959815Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8960049Z # genai-perf 2025-12-04T09:30:59.8960213Z pluggy==1.5.0 2025-12-04T09:30:59.8960373Z  # via 2025-12-04T09:30:59.8960533Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8960717Z # pytest 2025-12-04T09:30:59.8960885Z # pytest-cov 2025-12-04T09:30:59.8961060Z polars==1.29.0 2025-12-04T09:30:59.8961230Z  # via 2025-12-04T09:30:59.8961380Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8961591Z # mteb 2025-12-04T09:30:59.8961740Z pooch==1.8.2 2025-12-04T09:30:59.8961898Z  # via 2025-12-04T09:30:59.8962052Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8962268Z # librosa 2025-12-04T09:30:59.8962516Z portalocker==2.10.1 2025-12-04T09:30:59.8962706Z  # via 2025-12-04T09:30:59.8962859Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8963202Z # sacrebleu 2025-12-04T09:30:59.8963404Z pqdm==0.2.0 2025-12-04T09:30:59.8963576Z  # via 2025-12-04T09:30:59.8963747Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8963987Z # -r requirements/test.in 2025-12-04T09:30:59.8964198Z pretrainedmodels==0.7.4 2025-12-04T09:30:59.8964393Z  # via 2025-12-04T09:30:59.8964544Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8964813Z # segmentation-models-pytorch 2025-12-04T09:30:59.8965049Z prometheus-client==0.22.0 2025-12-04T09:30:59.8965244Z  # via 2025-12-04T09:30:59.8965401Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8965622Z # opentelemetry-exporter-prometheus 2025-12-04T09:30:59.8965864Z # ray 2025-12-04T09:30:59.8966018Z propcache==0.2.0 2025-12-04T09:30:59.8966203Z  # via 2025-12-04T09:30:59.8966366Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8966557Z # aiohttp 2025-12-04T09:30:59.8966737Z # yarl 2025-12-04T09:30:59.8966903Z proto-plus==1.26.1 2025-12-04T09:30:59.8966992Z  # via 2025-12-04T09:30:59.8967067Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8967173Z # google-api-core 2025-12-04T09:30:59.8967237Z protobuf==5.28.3 2025-12-04T09:30:59.8967315Z  # via 2025-12-04T09:30:59.8967494Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8967560Z # google-api-core 2025-12-04T09:30:59.8967636Z # googleapis-common-protos 2025-12-04T09:30:59.8967703Z # mlflow-skinny 2025-12-04T09:30:59.8967774Z # opentelemetry-proto 2025-12-04T09:30:59.8967834Z # proto-plus 2025-12-04T09:30:59.8967897Z # ray 2025-12-04T09:30:59.8967956Z # tensorboardx 2025-12-04T09:30:59.8968052Z # tensorizer 2025-12-04T09:30:59.8968114Z psutil==6.1.0 2025-12-04T09:30:59.8968192Z  # via 2025-12-04T09:30:59.8968268Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8968336Z # accelerate 2025-12-04T09:30:59.8968392Z # peft 2025-12-04T09:30:59.8968487Z # tensorizer 2025-12-04T09:30:59.8968544Z py==1.11.0 2025-12-04T09:30:59.8968621Z  # via 2025-12-04T09:30:59.8968698Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8968797Z # pytest-forked 2025-12-04T09:30:59.8968859Z py-spy==0.4.0 2025-12-04T09:30:59.8968946Z  # via 2025-12-04T09:30:59.8969018Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8969097Z # ray 2025-12-04T09:30:59.8969163Z pyarrow==18.0.0 2025-12-04T09:30:59.8969239Z  # via 2025-12-04T09:30:59.8969308Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8969368Z # datasets 2025-12-04T09:30:59.8969427Z # genai-perf 2025-12-04T09:30:59.8969509Z # mlflow 2025-12-04T09:30:59.8969574Z pyasn1==0.6.1 2025-12-04T09:30:59.8969650Z  # via 2025-12-04T09:30:59.8969731Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8969801Z # pyasn1-modules 2025-12-04T09:30:59.8969884Z # rsa 2025-12-04T09:30:59.8969959Z pyasn1-modules==0.4.2 2025-12-04T09:30:59.8970035Z  # via 2025-12-04T09:30:59.8970117Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8970215Z # google-auth 2025-12-04T09:30:59.8970277Z pybind11==2.13.6 2025-12-04T09:30:59.8970354Z  # via 2025-12-04T09:30:59.8970435Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8970523Z # lm-eval 2025-12-04T09:30:59.8970586Z pycocotools==2.0.8 2025-12-04T09:30:59.8970669Z  # via 2025-12-04T09:30:59.8970736Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8970825Z # terratorch 2025-12-04T09:30:59.8970895Z pycountry==24.6.1 2025-12-04T09:30:59.8971211Z  # via 2025-12-04T09:30:59.8971311Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8971429Z # pydantic-extra-types 2025-12-04T09:30:59.8971491Z pycparser==2.22 2025-12-04T09:30:59.8971722Z  # via 2025-12-04T09:30:59.8971811Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8971899Z # cffi 2025-12-04T09:30:59.8971974Z pycryptodomex==3.22.0 2025-12-04T09:30:59.8972051Z  # via 2025-12-04T09:30:59.8972122Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8972213Z # blobfile 2025-12-04T09:30:59.8972275Z pydantic==2.12.0 2025-12-04T09:30:59.8972352Z  # via 2025-12-04T09:30:59.8972433Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8972503Z # -r requirements/test.in 2025-12-04T09:30:59.8972565Z # albumentations 2025-12-04T09:30:59.8972645Z # datamodel-code-generator 2025-12-04T09:30:59.8972706Z # fastapi 2025-12-04T09:30:59.8972771Z # gpt-oss 2025-12-04T09:30:59.8972827Z # lightly 2025-12-04T09:30:59.8972890Z # mistral-common 2025-12-04T09:30:59.8972961Z # mlflow-skinny 2025-12-04T09:30:59.8973022Z # mteb 2025-12-04T09:30:59.8973086Z # openai-harmony 2025-12-04T09:30:59.8973168Z # pydantic-extra-types 2025-12-04T09:30:59.8973252Z # ray 2025-12-04T09:30:59.8973317Z pydantic-core==2.41.1 2025-12-04T09:30:59.8973400Z  # via 2025-12-04T09:30:59.8973472Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8973559Z # pydantic 2025-12-04T09:30:59.8973640Z pydantic-extra-types==2.10.5 2025-12-04T09:30:59.8973719Z  # via 2025-12-04T09:30:59.8973800Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8974021Z # mistral-common 2025-12-04T09:30:59.8974085Z pygments==2.18.0 2025-12-04T09:30:59.8974168Z  # via 2025-12-04T09:30:59.8974250Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8974334Z # rich 2025-12-04T09:30:59.8974402Z pyogrio==0.11.0 2025-12-04T09:30:59.8974480Z  # via 2025-12-04T09:30:59.8974551Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8974646Z # geopandas 2025-12-04T09:30:59.8974710Z pyparsing==3.2.0 2025-12-04T09:30:59.8974788Z  # via 2025-12-04T09:30:59.8974867Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8974926Z # matplotlib 2025-12-04T09:30:59.8975013Z # rasterio 2025-12-04T09:30:59.8975078Z pyproj==3.7.1 2025-12-04T09:30:59.8975161Z  # via 2025-12-04T09:30:59.8975233Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8975295Z # geopandas 2025-12-04T09:30:59.8975354Z # rioxarray 2025-12-04T09:30:59.8975444Z # torchgeo 2025-12-04T09:30:59.8975514Z pyrate-limiter==3.7.0 2025-12-04T09:30:59.8975592Z  # via 2025-12-04T09:30:59.8975668Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8975759Z # schemathesis 2025-12-04T09:30:59.8975820Z pystemmer==3.0.0 2025-12-04T09:30:59.8975905Z  # via 2025-12-04T09:30:59.8975975Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8976055Z # mteb 2025-12-04T09:30:59.8976124Z pytablewriter==1.2.0 2025-12-04T09:30:59.8976201Z  # via 2025-12-04T09:30:59.8976273Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8976363Z # lm-eval 2025-12-04T09:30:59.8976421Z pytest==8.3.5 2025-12-04T09:30:59.8976497Z  # via 2025-12-04T09:30:59.8976569Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8976639Z # -r requirements/test.in 2025-12-04T09:30:59.8976718Z # buildkite-test-collector 2025-12-04T09:30:59.8976776Z # genai-perf 2025-12-04T09:30:59.8976839Z # pytest-asyncio 2025-12-04T09:30:59.8976903Z # pytest-cov 2025-12-04T09:30:59.8976967Z # pytest-forked 2025-12-04T09:30:59.8977037Z # pytest-mock 2025-12-04T09:30:59.8977112Z # pytest-rerunfailures 2025-12-04T09:30:59.8977171Z # pytest-shard 2025-12-04T09:30:59.8977235Z # pytest-subtests 2025-12-04T09:30:59.8977304Z # pytest-timeout 2025-12-04T09:30:59.8977363Z # schemathesis 2025-12-04T09:30:59.8977456Z # terratorch 2025-12-04T09:30:59.8977527Z pytest-asyncio==0.24.0 2025-12-04T09:30:59.8977605Z  # via 2025-12-04T09:30:59.8977760Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8977881Z # -r requirements/test.in 2025-12-04T09:30:59.8977946Z pytest-cov==6.3.0 2025-12-04T09:30:59.8978028Z  # via 2025-12-04T09:30:59.8978098Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8978201Z # -r requirements/test.in 2025-12-04T09:30:59.8978269Z pytest-forked==1.6.0 2025-12-04T09:30:59.8978355Z  # via 2025-12-04T09:30:59.8978425Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8978540Z # -r requirements/test.in 2025-12-04T09:30:59.8978603Z pytest-mock==3.14.0 2025-12-04T09:30:59.8978680Z  # via 2025-12-04T09:30:59.8978755Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8978844Z # genai-perf 2025-12-04T09:30:59.8978917Z pytest-rerunfailures==14.0 2025-12-04T09:30:59.8979001Z  # via 2025-12-04T09:30:59.8979070Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8979249Z # -r requirements/test.in 2025-12-04T09:30:59.8979322Z pytest-shard==0.1.2 2025-12-04T09:30:59.8979400Z  # via 2025-12-04T09:30:59.8979476Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8979580Z # -r requirements/test.in 2025-12-04T09:30:59.8979648Z pytest-subtests==0.14.1 2025-12-04T09:30:59.8979730Z  # via 2025-12-04T09:30:59.8979798Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8979891Z # schemathesis 2025-12-04T09:30:59.8979963Z pytest-timeout==2.3.1 2025-12-04T09:30:59.8980159Z  # via 2025-12-04T09:30:59.8980230Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8980352Z # -r requirements/test.in 2025-12-04T09:30:59.8980417Z python-box==7.3.2 2025-12-04T09:30:59.8980495Z  # via 2025-12-04T09:30:59.8980572Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8980661Z # terratorch 2025-12-04T09:30:59.8980737Z python-dateutil==2.9.0.post0 2025-12-04T09:30:59.8980815Z  # via 2025-12-04T09:30:59.8980884Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8980949Z # arrow 2025-12-04T09:30:59.8981008Z # botocore 2025-12-04T09:30:59.8981065Z # graphene 2025-12-04T09:30:59.8981126Z # lightly 2025-12-04T09:30:59.8981186Z # matplotlib 2025-12-04T09:30:59.8981243Z # pandas 2025-12-04T09:30:59.8981336Z # typepy 2025-12-04T09:30:59.8981406Z python-rapidjson==1.20 2025-12-04T09:30:59.8981485Z  # via 2025-12-04T09:30:59.8981564Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8981663Z # tritonclient 2025-12-04T09:30:59.8981732Z pytorch-lightning==2.6.0 2025-12-04T09:30:59.8981825Z  # via 2025-12-04T09:30:59.8981882Z # lightly 2025-12-04T09:30:59.8981972Z # lightning 2025-12-04T09:30:59.8982055Z pytrec-eval-terrier==0.5.7 2025-12-04T09:30:59.8982133Z  # via 2025-12-04T09:30:59.8982209Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8982291Z # mteb 2025-12-04T09:30:59.8982351Z pytz==2024.2 2025-12-04T09:30:59.8982438Z  # via 2025-12-04T09:30:59.8982512Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8982579Z # pandas 2025-12-04T09:30:59.8982672Z # typepy 2025-12-04T09:30:59.8982731Z pyyaml==6.0.2 2025-12-04T09:30:59.8982809Z  # via 2025-12-04T09:30:59.8982885Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8982943Z # accelerate 2025-12-04T09:30:59.8983006Z # albumentations 2025-12-04T09:30:59.8983082Z # datamodel-code-generator 2025-12-04T09:30:59.8983142Z # datasets 2025-12-04T09:30:59.8983201Z # genai-perf 2025-12-04T09:30:59.8983272Z # huggingface-hub 2025-12-04T09:30:59.8983332Z # jsonargparse 2025-12-04T09:30:59.8983392Z # lightning 2025-12-04T09:30:59.8983452Z # mlflow-skinny 2025-12-04T09:30:59.8983508Z # omegaconf 2025-12-04T09:30:59.8983569Z # peft 2025-12-04T09:30:59.8983637Z # pytorch-lightning 2025-12-04T09:30:59.8983693Z # ray 2025-12-04T09:30:59.8983757Z # responses 2025-12-04T09:30:59.8983816Z # schemathesis 2025-12-04T09:30:59.8983954Z # timm 2025-12-04T09:30:59.8984030Z # transformers 2025-12-04T09:30:59.8984119Z # vocos 2025-12-04T09:30:59.8984180Z rapidfuzz==3.12.1 2025-12-04T09:30:59.8984261Z  # via 2025-12-04T09:30:59.8984331Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8984411Z # jiwer 2025-12-04T09:30:59.8984475Z rasterio==1.4.3 2025-12-04T09:30:59.8984551Z  # via 2025-12-04T09:30:59.8984628Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8984689Z # rioxarray 2025-12-04T09:30:59.8984746Z # terratorch 2025-12-04T09:30:59.8984847Z # torchgeo 2025-12-04T09:30:59.8984904Z ray==2.48.0 2025-12-04T09:30:59.8984982Z  # via 2025-12-04T09:30:59.8985057Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8985161Z # -r requirements/test.in 2025-12-04T09:30:59.8985218Z redis==5.2.0 2025-12-04T09:30:59.8985310Z  # via 2025-12-04T09:30:59.8985383Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8985479Z # tensorizer 2025-12-04T09:30:59.8985548Z referencing==0.35.1 2025-12-04T09:30:59.8985626Z  # via 2025-12-04T09:30:59.8985696Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8985761Z # jsonschema 2025-12-04T09:30:59.8985880Z # jsonschema-specifications 2025-12-04T09:30:59.8986018Z regex==2024.9.11 2025-12-04T09:30:59.8986097Z  # via 2025-12-04T09:30:59.8986185Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8986322Z # nltk 2025-12-04T09:30:59.8986388Z # open-clip-torch 2025-12-04T09:30:59.8986453Z # sacrebleu 2025-12-04T09:30:59.8986508Z # tiktoken 2025-12-04T09:30:59.8986605Z # transformers 2025-12-04T09:30:59.8986673Z requests==2.32.3 2025-12-04T09:30:59.8986751Z  # via 2025-12-04T09:30:59.8986821Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8986900Z # buildkite-test-collector 2025-12-04T09:30:59.8986963Z # databricks-sdk 2025-12-04T09:30:59.8987026Z # datasets 2025-12-04T09:30:59.8987089Z # docker 2025-12-04T09:30:59.8987155Z # evaluate 2025-12-04T09:30:59.8987225Z # google-api-core 2025-12-04T09:30:59.8987295Z # google-cloud-storage 2025-12-04T09:30:59.8987353Z # gpt-oss 2025-12-04T09:30:59.8987422Z # huggingface-hub 2025-12-04T09:30:59.8987479Z # lightly 2025-12-04T09:30:59.8987535Z # lm-eval 2025-12-04T09:30:59.8987603Z # mistral-common 2025-12-04T09:30:59.8987673Z # mlflow-skinny 2025-12-04T09:30:59.8987728Z # mteb 2025-12-04T09:30:59.8987788Z # pooch 2025-12-04T09:30:59.8987841Z # ray 2025-12-04T09:30:59.8987898Z # responses 2025-12-04T09:30:59.8987965Z # schemathesis 2025-12-04T09:30:59.8988033Z # starlette-testclient 2025-12-04T09:30:59.8988098Z # tiktoken 2025-12-04T09:30:59.8988196Z # transformers 2025-12-04T09:30:59.8988257Z responses==0.25.3 2025-12-04T09:30:59.8988341Z  # via 2025-12-04T09:30:59.8988410Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8988504Z # genai-perf 2025-12-04T09:30:59.8988585Z rfc3339-validator==0.1.4 2025-12-04T09:30:59.8988665Z  # via 2025-12-04T09:30:59.8988735Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8988828Z # jsonschema 2025-12-04T09:30:59.8988884Z rfc3987==1.3.8 2025-12-04T09:30:59.8988961Z  # via 2025-12-04T09:30:59.8989037Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8989125Z # jsonschema 2025-12-04T09:30:59.8989187Z rich==13.9.4 2025-12-04T09:30:59.8989268Z  # via 2025-12-04T09:30:59.8989337Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8989405Z # genai-perf 2025-12-04T09:30:59.8989469Z # lightning 2025-12-04T09:30:59.8989525Z # mteb 2025-12-04T09:30:59.8989616Z # typer 2025-12-04T09:30:59.8989677Z rioxarray==0.19.0 2025-12-04T09:30:59.8989754Z  # via 2025-12-04T09:30:59.8989829Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8989919Z # terratorch 2025-12-04T09:30:59.8990062Z rouge-score==0.1.2 2025-12-04T09:30:59.8990162Z  # via 2025-12-04T09:30:59.8990235Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8990321Z # lm-eval 2025-12-04T09:30:59.8990392Z rpds-py==0.20.1 2025-12-04T09:30:59.8990469Z  # via 2025-12-04T09:30:59.8990537Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8990602Z # jsonschema 2025-12-04T09:30:59.8990692Z # referencing 2025-12-04T09:30:59.8990753Z rsa==4.9.1 2025-12-04T09:30:59.8990838Z  # via 2025-12-04T09:30:59.8990907Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8991001Z # google-auth 2025-12-04T09:30:59.8991059Z rtree==1.4.0 2025-12-04T09:30:59.8991135Z  # via 2025-12-04T09:30:59.8991210Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8991293Z # torchgeo 2025-12-04T09:30:59.8991364Z runai-model-streamer==0.14.0 2025-12-04T09:30:59.8991447Z  # via 2025-12-04T09:30:59.8991517Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8991631Z # -r requirements/test.in 2025-12-04T09:30:59.8991723Z runai-model-streamer-gcs==0.14.0 2025-12-04T09:30:59.8991803Z  # via 2025-12-04T09:30:59.8991874Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8991985Z # runai-model-streamer 2025-12-04T09:30:59.8992060Z runai-model-streamer-s3==0.14.0 2025-12-04T09:30:59.8992144Z  # via 2025-12-04T09:30:59.8992214Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8992408Z # runai-model-streamer 2025-12-04T09:30:59.8992475Z s3transfer==0.10.3 2025-12-04T09:30:59.8992555Z  # via 2025-12-04T09:30:59.8992624Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8992722Z # boto3 2025-12-04T09:30:59.8992784Z sacrebleu==2.4.3 2025-12-04T09:30:59.8992864Z  # via 2025-12-04T09:30:59.8992939Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8993023Z # lm-eval 2025-12-04T09:30:59.8993085Z safetensors==0.4.5 2025-12-04T09:30:59.8993172Z  # via 2025-12-04T09:30:59.8993252Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8993310Z # accelerate 2025-12-04T09:30:59.8993382Z # open-clip-torch 2025-12-04T09:30:59.8993439Z # peft 2025-12-04T09:30:59.8993505Z # timm 2025-12-04T09:30:59.8993599Z # transformers 2025-12-04T09:30:59.8993665Z schemathesis==3.39.15 2025-12-04T09:30:59.8993747Z  # via 2025-12-04T09:30:59.8993824Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8993934Z # -r requirements/test.in 2025-12-04T09:30:59.8994003Z scikit-image==0.25.2 2025-12-04T09:30:59.8994080Z  # via 2025-12-04T09:30:59.8994150Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8994250Z # albumentations 2025-12-04T09:30:59.8994311Z scikit-learn==1.5.2 2025-12-04T09:30:59.8994389Z  # via 2025-12-04T09:30:59.8994462Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8994523Z # albumentations 2025-12-04T09:30:59.8994579Z # librosa 2025-12-04T09:30:59.8994654Z # lm-eval 2025-12-04T09:30:59.8994710Z # mlflow 2025-12-04T09:30:59.8994774Z # mteb 2025-12-04T09:30:59.8994885Z # sentence-transformers 2025-12-04T09:30:59.8994942Z scipy==1.13.1 2025-12-04T09:30:59.8995024Z  # via 2025-12-04T09:30:59.8995093Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8995155Z # albumentations 2025-12-04T09:30:59.8995214Z # bm25s 2025-12-04T09:30:59.8995269Z # librosa 2025-12-04T09:30:59.8995326Z # mlflow 2025-12-04T09:30:59.8995392Z # mteb 2025-12-04T09:30:59.8995453Z # scikit-image 2025-12-04T09:30:59.8995512Z # scikit-learn 2025-12-04T09:30:59.8995589Z # sentence-transformers 2025-12-04T09:30:59.8995647Z # statsmodels 2025-12-04T09:30:59.8995731Z # vocos 2025-12-04T09:30:59.8995818Z segmentation-models-pytorch==0.4.0 2025-12-04T09:30:59.8995895Z  # via 2025-12-04T09:30:59.8995951Z # terratorch 2025-12-04T09:30:59.8996042Z # torchgeo 2025-12-04T09:30:59.8996238Z sentence-transformers==3.2.1 2025-12-04T09:30:59.8996336Z  # via 2025-12-04T09:30:59.8996410Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8996479Z # -r requirements/test.in 2025-12-04T09:30:59.8996567Z # mteb 2025-12-04T09:30:59.8996628Z setuptools==77.0.3 2025-12-04T09:30:59.8996704Z  # via 2025-12-04T09:30:59.8996791Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8996860Z # lightning-utilities 2025-12-04T09:30:59.8996926Z # pytablewriter 2025-12-04T09:30:59.8997016Z # torch 2025-12-04T09:30:59.8997077Z shapely==2.1.1 2025-12-04T09:30:59.8997153Z  # via 2025-12-04T09:30:59.8997231Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8997292Z # geopandas 2025-12-04T09:30:59.8997380Z # torchgeo 2025-12-04T09:30:59.8997449Z shellingham==1.5.4 2025-12-04T09:30:59.8997530Z  # via 2025-12-04T09:30:59.8997617Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8997705Z # typer 2025-12-04T09:30:59.8997764Z six==1.16.0 2025-12-04T09:30:59.8997847Z  # via 2025-12-04T09:30:59.8997915Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8997975Z # junit-xml 2025-12-04T09:30:59.8998037Z # lightly 2025-12-04T09:30:59.8998095Z # opencensus 2025-12-04T09:30:59.8998159Z # python-dateutil 2025-12-04T09:30:59.8998229Z # rfc3339-validator 2025-12-04T09:30:59.8998298Z # rouge-score 2025-12-04T09:30:59.8998500Z # segmentation-models-pytorch 2025-12-04T09:30:59.8998569Z smart-open==7.1.0 2025-12-04T09:30:59.8998648Z  # via 2025-12-04T09:30:59.8998730Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8998816Z # ray 2025-12-04T09:30:59.8998875Z smmap==5.0.2 2025-12-04T09:30:59.8998956Z  # via 2025-12-04T09:30:59.8999025Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8999108Z # gitdb 2025-12-04T09:30:59.8999172Z sniffio==1.3.1 2025-12-04T09:30:59.8999247Z  # via 2025-12-04T09:30:59.8999320Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8999381Z # anyio 2025-12-04T09:30:59.8999462Z # httpx 2025-12-04T09:30:59.8999530Z sortedcontainers==2.4.0 2025-12-04T09:30:59.8999611Z  # via 2025-12-04T09:30:59.8999679Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.8999767Z # hypothesis 2025-12-04T09:30:59.8999834Z soundfile==0.12.1 2025-12-04T09:30:59.8999909Z  # via 2025-12-04T09:30:59.8999981Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9000065Z # -r requirements/test.in 2025-12-04T09:30:59.9000121Z # librosa 2025-12-04T09:30:59.9000223Z # mistral-common 2025-12-04T09:30:59.9000284Z soxr==0.5.0.post1 2025-12-04T09:30:59.9000360Z  # via 2025-12-04T09:30:59.9000441Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9000497Z # librosa 2025-12-04T09:30:59.9000591Z # mistral-common 2025-12-04T09:30:59.9000660Z sqlalchemy==2.0.41 2025-12-04T09:30:59.9000738Z  # via 2025-12-04T09:30:59.9000843Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9000950Z # alembic 2025-12-04T09:30:59.9001082Z # mlflow 2025-12-04T09:30:59.9001147Z sqlitedict==2.1.0 2025-12-04T09:30:59.9001232Z  # via 2025-12-04T09:30:59.9001306Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9001391Z # lm-eval 2025-12-04T09:30:59.9001463Z sqlparse==0.5.3 2025-12-04T09:30:59.9001540Z  # via 2025-12-04T09:30:59.9001621Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9001718Z # mlflow-skinny 2025-12-04T09:30:59.9001780Z starlette==0.46.2 2025-12-04T09:30:59.9001868Z  # via 2025-12-04T09:30:59.9001941Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9001999Z # fastapi 2025-12-04T09:30:59.9002066Z # schemathesis 2025-12-04T09:30:59.9002174Z # starlette-testclient 2025-12-04T09:30:59.9002246Z starlette-testclient==0.4.1 2025-12-04T09:30:59.9002332Z  # via 2025-12-04T09:30:59.9002496Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9002606Z # schemathesis 2025-12-04T09:30:59.9002679Z statsmodels==0.14.4 2025-12-04T09:30:59.9002758Z  # via 2025-12-04T09:30:59.9002828Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9002924Z # genai-perf 2025-12-04T09:30:59.9002984Z structlog==25.4.0 2025-12-04T09:30:59.9003068Z  # via 2025-12-04T09:30:59.9003136Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9003225Z # gpt-oss 2025-12-04T09:30:59.9003288Z sympy==1.13.3 2025-12-04T09:30:59.9003366Z  # via 2025-12-04T09:30:59.9003435Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9003496Z # einx 2025-12-04T09:30:59.9003577Z # torch 2025-12-04T09:30:59.9003637Z tabledata==1.3.3 2025-12-04T09:30:59.9003717Z  # via 2025-12-04T09:30:59.9003785Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9003879Z # pytablewriter 2025-12-04T09:30:59.9003944Z tabulate==0.9.0 2025-12-04T09:30:59.9004025Z  # via 2025-12-04T09:30:59.9004107Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9004203Z # sacrebleu 2025-12-04T09:30:59.9004260Z tblib==3.1.0 2025-12-04T09:30:59.9004342Z  # via 2025-12-04T09:30:59.9004410Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9004515Z # -r requirements/test.in 2025-12-04T09:30:59.9004581Z tcolorpy==0.1.6 2025-12-04T09:30:59.9004658Z  # via 2025-12-04T09:30:59.9004726Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9004935Z # pytablewriter 2025-12-04T09:30:59.9004997Z tenacity==9.1.2 2025-12-04T09:30:59.9005074Z  # via 2025-12-04T09:30:59.9005155Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9005214Z # gpt-oss 2025-12-04T09:30:59.9005274Z # lm-eval 2025-12-04T09:30:59.9005364Z # plotly 2025-12-04T09:30:59.9005428Z tensorboardx==2.6.4 2025-12-04T09:30:59.9005507Z  # via 2025-12-04T09:30:59.9005582Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9005673Z # lightning 2025-12-04T09:30:59.9005740Z tensorizer==2.10.1 2025-12-04T09:30:59.9005817Z  # via 2025-12-04T09:30:59.9005896Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9006008Z # -r requirements/test.in 2025-12-04T09:30:59.9006068Z termcolor==3.1.0 2025-12-04T09:30:59.9006146Z  # via 2025-12-04T09:30:59.9006226Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9006311Z # gpt-oss 2025-12-04T09:30:59.9006569Z terratorch @ git+https://github.com/IBM/terratorch.git@07184fcf91a1324f831ff521dd238d97fe350e3e 2025-12-04T09:30:59.9006713Z  # via -r requirements/test.in 2025-12-04T09:30:59.9006780Z threadpoolctl==3.5.0 2025-12-04T09:30:59.9006859Z  # via 2025-12-04T09:30:59.9006935Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9007027Z # scikit-learn 2025-12-04T09:30:59.9007093Z tifffile==2025.3.30 2025-12-04T09:30:59.9007170Z  # via 2025-12-04T09:30:59.9007239Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9007308Z # scikit-image 2025-12-04T09:30:59.9007398Z # terratorch 2025-12-04T09:30:59.9007459Z tiktoken==0.12.0 2025-12-04T09:30:59.9007551Z  # via 2025-12-04T09:30:59.9007623Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9007681Z # gpt-oss 2025-12-04T09:30:59.9007747Z # lm-eval 2025-12-04T09:30:59.9007844Z # mistral-common 2025-12-04T09:30:59.9007901Z timm==1.0.17 2025-12-04T09:30:59.9007989Z  # via 2025-12-04T09:30:59.9008060Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9008128Z # -r requirements/test.in 2025-12-04T09:30:59.9008200Z # open-clip-torch 2025-12-04T09:30:59.9008280Z # segmentation-models-pytorch 2025-12-04T09:30:59.9008347Z # terratorch 2025-12-04T09:30:59.9008435Z # torchgeo 2025-12-04T09:30:59.9008497Z tokenizers==0.22.0 2025-12-04T09:30:59.9008593Z  # via 2025-12-04T09:30:59.9008669Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9008829Z # -r requirements/test.in 2025-12-04T09:30:59.9009016Z # transformers 2025-12-04T09:30:59.9009087Z tomli==2.2.1 2025-12-04T09:30:59.9009172Z  # via 2025-12-04T09:30:59.9009252Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9009344Z # schemathesis 2025-12-04T09:30:59.9009407Z tomli-w==1.2.0 2025-12-04T09:30:59.9009490Z  # via 2025-12-04T09:30:59.9009561Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9009654Z # schemathesis 2025-12-04T09:30:59.9010148Z torch @ file:///var/lib/jenkins/workspace/dist/torch-2.10.0a0%2Bgitffd9b0f-cp312-cp312-linux_x86_64.whl#sha256=a0c782d08e3b01a62f42b12b7832d890773e7e8ac303ff47ce3c04911a4563e9 2025-12-04T09:30:59.9010229Z  # via 2025-12-04T09:30:59.9010304Z # -r requirements/test.in 2025-12-04T09:30:59.9010365Z # accelerate 2025-12-04T09:30:59.9010424Z # bitsandbytes 2025-12-04T09:30:59.9010502Z # efficientnet-pytorch 2025-12-04T09:30:59.9010563Z # encodec 2025-12-04T09:30:59.9010634Z # fastsafetensors 2025-12-04T09:30:59.9010700Z # kornia 2025-12-04T09:30:59.9010756Z # lightly 2025-12-04T09:30:59.9010815Z # lightning 2025-12-04T09:30:59.9010878Z # lm-eval 2025-12-04T09:30:59.9010933Z # mteb 2025-12-04T09:30:59.9010998Z # open-clip-torch 2025-12-04T09:30:59.9011062Z # peft 2025-12-04T09:30:59.9011130Z # pretrainedmodels 2025-12-04T09:30:59.9011202Z # pytorch-lightning 2025-12-04T09:30:59.9011372Z # runai-model-streamer 2025-12-04T09:30:59.9011454Z # segmentation-models-pytorch 2025-12-04T09:30:59.9011575Z # sentence-transformers 2025-12-04T09:30:59.9011668Z # tensorizer 2025-12-04T09:30:59.9011728Z # terratorch 2025-12-04T09:30:59.9011790Z # timm 2025-12-04T09:30:59.9011847Z # torchaudio 2025-12-04T09:30:59.9011903Z # torchgeo 2025-12-04T09:30:59.9011968Z # torchmetrics 2025-12-04T09:30:59.9012028Z # torchvision 2025-12-04T09:30:59.9012105Z # vector-quantize-pytorch 2025-12-04T09:30:59.9012211Z # vocos 2025-12-04T09:30:59.9012750Z torchaudio @ file:///var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0%2Be90a398-cp312-cp312-linux_x86_64.whl#sha256=9fd6d716ab57b84fbbf009e8463967c8254f16e45d6b9c9618575442317da3e1 2025-12-04T09:30:59.9012836Z  # via 2025-12-04T09:30:59.9012908Z # -r requirements/test.in 2025-12-04T09:30:59.9012964Z # encodec 2025-12-04T09:30:59.9013051Z # vocos 2025-12-04T09:30:59.9013115Z torchgeo==0.7.0 2025-12-04T09:30:59.9013220Z  # via terratorch 2025-12-04T09:30:59.9013291Z torchmetrics==1.8.2 2025-12-04T09:30:59.9013374Z  # via 2025-12-04T09:30:59.9013432Z # lightning 2025-12-04T09:30:59.9013507Z # pytorch-lightning 2025-12-04T09:30:59.9013565Z # terratorch 2025-12-04T09:30:59.9013653Z # torchgeo 2025-12-04T09:30:59.9014199Z torchvision @ file:///var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0%2B617079d-cp312-cp312-linux_x86_64.whl#sha256=fc250b889759c63f19d9bb8fd5404f56b8b0b7c10d6e052aa34730d9dcb2b4a4 2025-12-04T09:30:59.9014278Z  # via 2025-12-04T09:30:59.9014353Z # -r requirements/test.in 2025-12-04T09:30:59.9014420Z # lightly 2025-12-04T09:30:59.9014493Z # open-clip-torch 2025-12-04T09:30:59.9014569Z # pretrainedmodels 2025-12-04T09:30:59.9014647Z # segmentation-models-pytorch 2025-12-04T09:30:59.9014709Z # terratorch 2025-12-04T09:30:59.9014773Z # timm 2025-12-04T09:30:59.9014863Z # torchgeo 2025-12-04T09:30:59.9014921Z tqdm==4.66.6 2025-12-04T09:30:59.9015003Z  # via 2025-12-04T09:30:59.9015076Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9015133Z # datasets 2025-12-04T09:30:59.9015197Z # evaluate 2025-12-04T09:30:59.9015261Z # huggingface-hub 2025-12-04T09:30:59.9015318Z # lightly 2025-12-04T09:30:59.9015382Z # lightning 2025-12-04T09:30:59.9015439Z # lm-eval 2025-12-04T09:30:59.9015501Z # mteb 2025-12-04T09:30:59.9015645Z # nltk 2025-12-04T09:30:59.9015720Z # open-clip-torch 2025-12-04T09:30:59.9015793Z # peft 2025-12-04T09:30:59.9015851Z # pqdm 2025-12-04T09:30:59.9015918Z # pretrainedmodels 2025-12-04T09:30:59.9015990Z # pytorch-lightning 2025-12-04T09:30:59.9016066Z # segmentation-models-pytorch 2025-12-04T09:30:59.9016139Z # sentence-transformers 2025-12-04T09:30:59.9016212Z # tqdm-multiprocess 2025-12-04T09:30:59.9016314Z # transformers 2025-12-04T09:30:59.9016384Z tqdm-multiprocess==0.0.11 2025-12-04T09:30:59.9016468Z  # via 2025-12-04T09:30:59.9016540Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9016625Z # lm-eval 2025-12-04T09:30:59.9016694Z transformers==4.56.2 2025-12-04T09:30:59.9016806Z  # via 2025-12-04T09:30:59.9016933Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9017009Z # -r requirements/test.in 2025-12-04T09:30:59.9017069Z # genai-perf 2025-12-04T09:30:59.9017135Z # lm-eval 2025-12-04T09:30:59.9017195Z # peft 2025-12-04T09:30:59.9017268Z # sentence-transformers 2025-12-04T09:30:59.9017405Z # transformers-stream-generator 2025-12-04T09:30:59.9017490Z transformers-stream-generator==0.0.5 2025-12-04T09:30:59.9017570Z  # via 2025-12-04T09:30:59.9017656Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9017763Z # -r requirements/test.in 2025-12-04T09:30:59.9017829Z tritonclient==2.51.0 2025-12-04T09:30:59.9018010Z  # via 2025-12-04T09:30:59.9018083Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9018152Z # -r requirements/test.in 2025-12-04T09:30:59.9018249Z # genai-perf 2025-12-04T09:30:59.9018311Z typepy==1.3.2 2025-12-04T09:30:59.9018394Z  # via 2025-12-04T09:30:59.9018463Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9018526Z # dataproperty 2025-12-04T09:30:59.9018594Z # pytablewriter 2025-12-04T09:30:59.9018682Z # tabledata 2025-12-04T09:30:59.9018748Z typer==0.15.2 2025-12-04T09:30:59.9018830Z  # via 2025-12-04T09:30:59.9018898Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9018996Z # fastsafetensors 2025-12-04T09:30:59.9019164Z types-python-dateutil==2.9.0.20241206 2025-12-04T09:30:59.9019246Z  # via 2025-12-04T09:30:59.9019317Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9019405Z # arrow 2025-12-04T09:30:59.9019472Z typeshed-client==2.8.2 2025-12-04T09:30:59.9019560Z  # via 2025-12-04T09:30:59.9019630Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9019723Z # jsonargparse 2025-12-04T09:30:59.9019797Z typing-extensions==4.15.0 2025-12-04T09:30:59.9019873Z  # via 2025-12-04T09:30:59.9019951Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9020021Z # aiosignal 2025-12-04T09:30:59.9020087Z # albumentations 2025-12-04T09:30:59.9020147Z # alembic 2025-12-04T09:30:59.9020207Z # chz 2025-12-04T09:30:59.9020266Z # fastapi 2025-12-04T09:30:59.9020324Z # graphene 2025-12-04T09:30:59.9020404Z # huggingface-hub 2025-12-04T09:30:59.9020460Z # librosa 2025-12-04T09:30:59.9020520Z # lightning 2025-12-04T09:30:59.9020596Z # lightning-utilities 2025-12-04T09:30:59.9020658Z # mistral-common 2025-12-04T09:30:59.9020721Z # mlflow-skinny 2025-12-04T09:30:59.9020781Z # mteb 2025-12-04T09:30:59.9020849Z # opentelemetry-api 2025-12-04T09:30:59.9020919Z # opentelemetry-sdk 2025-12-04T09:30:59.9021013Z # opentelemetry-semantic-conventions 2025-12-04T09:30:59.9021070Z # pqdm 2025-12-04T09:30:59.9021132Z # pydantic 2025-12-04T09:30:59.9021194Z # pydantic-core 2025-12-04T09:30:59.9021266Z # pydantic-extra-types 2025-12-04T09:30:59.9021339Z # pytorch-lightning 2025-12-04T09:30:59.9021397Z # sqlalchemy 2025-12-04T09:30:59.9021454Z # torch 2025-12-04T09:30:59.9021519Z # torchgeo 2025-12-04T09:30:59.9021576Z # typer 2025-12-04T09:30:59.9021717Z # typeshed-client 2025-12-04T09:30:59.9021845Z # typing-inspection 2025-12-04T09:30:59.9021916Z typing-inspection==0.4.2 2025-12-04T09:30:59.9021996Z  # via 2025-12-04T09:30:59.9022073Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9022159Z # pydantic 2025-12-04T09:30:59.9022223Z tzdata==2024.2 2025-12-04T09:30:59.9022301Z  # via 2025-12-04T09:30:59.9022372Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9022465Z # pandas 2025-12-04T09:30:59.9022530Z uri-template==1.3.0 2025-12-04T09:30:59.9022608Z  # via 2025-12-04T09:30:59.9022682Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9022771Z # jsonschema 2025-12-04T09:30:59.9022830Z urllib3==2.2.3 2025-12-04T09:30:59.9022917Z  # via 2025-12-04T09:30:59.9022986Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9023044Z # blobfile 2025-12-04T09:30:59.9023118Z # botocore 2025-12-04T09:30:59.9023175Z # docker 2025-12-04T09:30:59.9023235Z # lightly 2025-12-04T09:30:59.9023296Z # requests 2025-12-04T09:30:59.9023356Z # responses 2025-12-04T09:30:59.9023459Z # tritonclient 2025-12-04T09:30:59.9023521Z uvicorn==0.35.0 2025-12-04T09:30:59.9023600Z  # via 2025-12-04T09:30:59.9023677Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9023734Z # gpt-oss 2025-12-04T09:30:59.9023828Z # mlflow-skinny 2025-12-04T09:30:59.9023912Z vector-quantize-pytorch==1.27.7 2025-12-04T09:30:59.9024130Z  # via -r requirements/test.in 2025-12-04T09:30:59.9024195Z virtualenv==20.31.2 2025-12-04T09:30:59.9024282Z  # via 2025-12-04T09:30:59.9024356Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9024438Z # ray 2025-12-04T09:30:59.9024502Z vocos==0.1.0 2025-12-04T09:30:59.9024580Z  # via 2025-12-04T09:30:59.9024651Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9024772Z # -r requirements/test.in 2025-12-04T09:30:59.9024833Z wcwidth==0.2.13 2025-12-04T09:30:59.9024927Z  # via 2025-12-04T09:30:59.9024998Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9025081Z # ftfy 2025-12-04T09:30:59.9025150Z webcolors==24.11.1 2025-12-04T09:30:59.9025229Z  # via 2025-12-04T09:30:59.9025300Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9025394Z # jsonschema 2025-12-04T09:30:59.9025455Z werkzeug==3.1.3 2025-12-04T09:30:59.9025532Z  # via 2025-12-04T09:30:59.9025612Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9025679Z # flask 2025-12-04T09:30:59.9025774Z # schemathesis 2025-12-04T09:30:59.9025842Z word2number==1.1 2025-12-04T09:30:59.9025919Z  # via 2025-12-04T09:30:59.9025990Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9026079Z # lm-eval 2025-12-04T09:30:59.9026138Z wrapt==1.17.2 2025-12-04T09:30:59.9026219Z  # via 2025-12-04T09:30:59.9026289Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9026379Z # smart-open 2025-12-04T09:30:59.9026446Z xarray==2025.7.1 2025-12-04T09:30:59.9026534Z  # via 2025-12-04T09:30:59.9026607Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9026699Z # rioxarray 2025-12-04T09:30:59.9026758Z xxhash==3.5.0 2025-12-04T09:30:59.9026835Z  # via 2025-12-04T09:30:59.9026909Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9026965Z # datasets 2025-12-04T09:30:59.9027049Z # evaluate 2025-12-04T09:30:59.9027114Z yarl==1.17.1 2025-12-04T09:30:59.9027190Z  # via 2025-12-04T09:30:59.9027258Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9027319Z # aiohttp 2025-12-04T09:30:59.9027410Z # schemathesis 2025-12-04T09:30:59.9027473Z zipp==3.23.0 2025-12-04T09:30:59.9027550Z  # via 2025-12-04T09:30:59.9027619Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9027725Z # importlib-metadata 2025-12-04T09:30:59.9027786Z zstandard==0.23.0 2025-12-04T09:30:59.9027863Z  # via 2025-12-04T09:30:59.9028039Z # -c snapshot_constraint.txt 2025-12-04T09:30:59.9028138Z # lm-eval 2025-12-04T09:31:00.0168963Z 2025-12-04 09:31:00,016 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install -r test.txt 2025-12-04T09:31:00.0169332Z 2025-12-04 09:31:00,016 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install -r test.txt 2025-12-04T09:31:00.0500767Z Using Python 3.12.5 environment at: /opt/conda/envs/py_3.12 2025-12-04T09:31:00.1447836Z Resolved 329 packages in 88ms 2025-12-04T09:31:00.1648121Z Updating https://github.com/EleutherAI/lm-evaluation-harness.git (206b7722158f58c35b7ffcd53b035fdbdda5126d) 2025-12-04T09:31:00.2939396Z Building terratorch @ git+https://github.com/IBM/terratorch.git@07184fcf91a1324f831ff521dd238d97fe350e3e 2025-12-04T09:31:00.3395411Z Downloading virtualenv (5.8MiB) 2025-12-04T09:31:00.3397646Z Downloading pandas (12.1MiB) 2025-12-04T09:31:00.3399475Z Downloading black (1.7MiB) 2025-12-04T09:31:00.3401699Z Downloading transformers (11.1MiB) 2025-12-04T09:31:00.3403989Z Downloading rapidfuzz (3.0MiB) 2025-12-04T09:31:00.3406186Z Downloading plotly (18.2MiB) 2025-12-04T09:31:00.3407847Z Downloading numpy (17.1MiB) 2025-12-04T09:31:00.3410034Z Downloading zstandard (5.2MiB) 2025-12-04T09:31:00.3412109Z Downloading py-spy (2.6MiB) 2025-12-04T09:31:00.3414563Z Downloading kaleido (76.2MiB) 2025-12-04T09:31:00.3416427Z Downloading fastparquet (1.7MiB) 2025-12-04T09:31:00.3418081Z Downloading openai-harmony (2.9MiB) 2025-12-04T09:31:00.3420174Z Downloading botocore (12.1MiB) 2025-12-04T09:31:00.3426179Z Downloading hf-xet (3.0MiB) 2025-12-04T09:31:00.3427649Z Downloading fiona (16.4MiB) 2025-12-04T09:31:00.3429875Z Downloading timm (2.4MiB) 2025-12-04T09:31:00.3432366Z Downloading mlflow-skinny (6.0MiB) 2025-12-04T09:31:00.3434433Z Downloading opencv-python-headless (47.7MiB) 2025-12-04T09:31:00.3436507Z Downloading pyogrio (26.4MiB) 2025-12-04T09:31:00.3438566Z Downloading bitsandbytes (69.5MiB) 2025-12-04T09:31:00.3441392Z Downloading cramjam (2.2MiB) 2025-12-04T09:31:00.3443584Z Downloading statsmodels (10.2MiB) 2025-12-04T09:31:00.3445828Z Downloading pillow (4.3MiB) 2025-12-04T09:31:00.3448214Z Downloading runai-model-streamer-s3 (5.6MiB) 2025-12-04T09:31:00.3450958Z Downloading rasterio (21.2MiB) 2025-12-04T09:31:00.3453453Z Downloading scikit-image (14.3MiB) 2025-12-04T09:31:00.3455737Z Downloading mlflow (27.7MiB) 2025-12-04T09:31:00.3458010Z Downloading fonttools (4.7MiB) 2025-12-04T09:31:00.3461034Z Downloading pycryptodomex (2.2MiB) 2025-12-04T09:31:00.3463883Z Downloading scikit-learn (12.3MiB) 2025-12-04T09:31:00.3466810Z Downloading tokenizers (3.2MiB) 2025-12-04T09:31:00.3469564Z Downloading tritonclient (13.3MiB) 2025-12-04T09:31:00.3472280Z Downloading decord (13.0MiB) 2025-12-04T09:31:00.3475573Z Downloading shapely (3.0MiB) 2025-12-04T09:31:00.3477610Z Downloading pyarrow (38.2MiB) 2025-12-04T09:31:00.3480218Z Downloading mistral-common (6.2MiB) 2025-12-04T09:31:00.3482669Z Downloading pyproj (9.1MiB) 2025-12-04T09:31:00.3485539Z Downloading aiohttp (1.7MiB) 2025-12-04T09:31:00.3488108Z Downloading ray (66.9MiB) 2025-12-04T09:31:00.3491343Z Downloading python-rapidjson (1.6MiB) 2025-12-04T09:31:00.3502615Z Downloading scipy (36.4MiB) 2025-12-04T09:31:00.3505445Z Downloading kornia-rs (2.6MiB) 2025-12-04T09:31:00.3508447Z Downloading sqlalchemy (3.2MiB) 2025-12-04T09:31:00.3511066Z Downloading runai-model-streamer-gcs (4.2MiB) 2025-12-04T09:31:00.3514044Z Downloading polars (33.2MiB) 2025-12-04T09:31:00.3517321Z Downloading h5py (4.7MiB) 2025-12-04T09:31:00.3519771Z Downloading pydantic-core (2.0MiB) 2025-12-04T09:31:00.3522791Z Downloading matplotlib (7.9MiB) 2025-12-04T09:31:00.3526016Z Downloading grpcio (5.6MiB) 2025-12-04T09:31:00.3529074Z Downloading python-box (4.1MiB) 2025-12-04T09:31:01.1984144Z Downloading python-rapidjson 2025-12-04T09:31:01.2110753Z Downloading networkx (1.6MiB) 2025-12-04T09:31:01.2133345Z Downloading fastparquet 2025-12-04T09:31:01.2302142Z Downloading mteb (1.5MiB) 2025-12-04T09:31:01.2441325Z Downloading aiohttp 2025-12-04T09:31:01.2623100Z Downloading open-clip-torch (1.5MiB) 2025-12-04T09:31:01.3489810Z Downloading black 2025-12-04T09:31:01.3626454Z Downloading nltk (1.4MiB) 2025-12-04T09:31:01.3651850Z Downloading pydantic-core 2025-12-04T09:31:01.3827501Z Downloading kiwisolver (1.4MiB) 2025-12-04T09:31:01.4867773Z Downloading cramjam 2025-12-04T09:31:01.5035642Z Downloading fastsafetensors (1.4MiB) 2025-12-04T09:31:01.5344588Z Downloading pycryptodomex 2025-12-04T09:31:01.5497963Z Downloading xarray (1.3MiB) 2025-12-04T09:31:01.5605802Z Built terratorch @ git+https://github.com/IBM/terratorch.git@07184fcf91a1324f831ff521dd238d97fe350e3e 2025-12-04T09:31:01.6686535Z Downloading timm 2025-12-04T09:31:01.6813072Z Downloading py-spy 2025-12-04T09:31:01.6866344Z Downloading setuptools (1.2MiB) 2025-12-04T09:31:01.6988684Z Downloading pygments (1.1MiB) 2025-12-04T09:31:01.7051696Z Downloading kornia-rs 2025-12-04T09:31:01.7198309Z Downloading soundfile (1.1MiB) 2025-12-04T09:31:01.7539299Z Downloading openai-harmony 2025-12-04T09:31:01.7733485Z Downloading kornia (1.0MiB) 2025-12-04T09:31:01.8669007Z Downloading rapidfuzz 2025-12-04T09:31:01.8727291Z Downloading hf-xet 2025-12-04T09:31:01.9203608Z Downloading tokenizers 2025-12-04T09:31:01.9532904Z Downloading sqlalchemy 2025-12-04T09:31:01.9742807Z Downloading shapely 2025-12-04T09:31:02.0690285Z Downloading kiwisolver 2025-12-04T09:31:02.1654053Z Downloading fastsafetensors 2025-12-04T09:31:02.1706338Z Downloading open-clip-torch 2025-12-04T09:31:02.1757319Z Downloading xarray 2025-12-04T09:31:02.2164309Z Downloading nltk 2025-12-04T09:31:02.3090740Z Downloading soundfile 2025-12-04T09:31:02.3170686Z Downloading networkx 2025-12-04T09:31:02.3414968Z Downloading pygments 2025-12-04T09:31:02.4008807Z Downloading python-box 2025-12-04T09:31:02.4551853Z Downloading runai-model-streamer-gcs 2025-12-04T09:31:02.5144453Z Downloading pillow 2025-12-04T09:31:02.6037827Z Downloading setuptools 2025-12-04T09:31:02.6301559Z Downloading fonttools 2025-12-04T09:31:02.6725284Z Downloading kornia 2025-12-04T09:31:02.6946919Z Downloading h5py 2025-12-04T09:31:02.7524108Z Updated https://github.com/EleutherAI/lm-evaluation-harness.git (206b7722158f58c35b7ffcd53b035fdbdda5126d) 2025-12-04T09:31:02.7532070Z Building lm-eval @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d 2025-12-04T09:31:02.8491662Z Downloading zstandard 2025-12-04T09:31:03.1518531Z Downloading mteb 2025-12-04T09:31:03.1623001Z Downloading runai-model-streamer-s3 2025-12-04T09:31:03.2382956Z Downloading virtualenv 2025-12-04T09:31:03.2458966Z Downloading grpcio 2025-12-04T09:31:03.4361847Z Downloading mistral-common 2025-12-04T09:31:03.7447784Z Building encodec==0.1.1 2025-12-04T09:31:03.7469511Z Building antlr4-python3-runtime==4.9.3 2025-12-04T09:31:03.7473963Z Building pretrainedmodels==0.7.4 2025-12-04T09:31:03.7476384Z Building docopt==0.6.2 2025-12-04T09:31:03.7502201Z Building rouge-score==0.1.2 2025-12-04T09:31:03.7536169Z Building efficientnet-pytorch==0.7.1 2025-12-04T09:31:03.7552839Z Building sqlitedict==2.1.0 2025-12-04T09:31:03.7614526Z Building transformers-stream-generator==0.0.5 2025-12-04T09:31:03.7615341Z Building word2number==1.1 2025-12-04T09:31:03.8550640Z Downloading mlflow-skinny 2025-12-04T09:31:04.1344673Z Downloading matplotlib 2025-12-04T09:31:04.1837817Z Downloading pyproj 2025-12-04T09:31:04.3136442Z Built efficientnet-pytorch==0.7.1 2025-12-04T09:31:04.3831008Z Built sqlitedict==2.1.0 2025-12-04T09:31:04.3886437Z Built transformers-stream-generator==0.0.5 2025-12-04T09:31:04.4083041Z Built encodec==0.1.1 2025-12-04T09:31:04.4164706Z Built docopt==0.6.2 2025-12-04T09:31:04.4400116Z Built word2number==1.1 2025-12-04T09:31:04.4618554Z Built rouge-score==0.1.2 2025-12-04T09:31:04.5074469Z Built antlr4-python3-runtime==4.9.3 2025-12-04T09:31:04.5101198Z Downloading statsmodels 2025-12-04T09:31:04.5145672Z Built pretrainedmodels==0.7.4 2025-12-04T09:31:04.8316900Z Downloading pandas 2025-12-04T09:31:04.8738527Z Downloading decord 2025-12-04T09:31:04.9022785Z Downloading scikit-learn 2025-12-04T09:31:05.0042860Z Downloading tritonclient 2025-12-04T09:31:05.0633746Z Downloading botocore 2025-12-04T09:31:05.0874844Z Downloading transformers 2025-12-04T09:31:05.1240768Z Downloading scikit-image 2025-12-04T09:31:05.2779255Z Downloading fiona 2025-12-04T09:31:05.3031420Z Downloading numpy 2025-12-04T09:31:05.6831940Z Downloading rasterio 2025-12-04T09:31:05.8716560Z Downloading pyogrio 2025-12-04T09:31:06.1937133Z Downloading polars 2025-12-04T09:31:06.3219353Z Downloading mlflow 2025-12-04T09:31:07.0711450Z Downloading opencv-python-headless 2025-12-04T09:31:07.1268388Z Downloading scipy 2025-12-04T09:31:07.1576945Z Downloading bitsandbytes 2025-12-04T09:31:07.1860395Z Downloading pyarrow 2025-12-04T09:31:07.3690823Z Downloading kaleido 2025-12-04T09:31:07.4580853Z Downloading ray 2025-12-04T09:31:09.0349530Z Downloading plotly 2025-12-04T09:31:17.2548769Z Built lm-eval @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d 2025-12-04T09:31:17.4906064Z Prepared 295 packages in 17.33s 2025-12-04T09:31:20.5172732Z Uninstalled 99 packages in 3.02s 2025-12-04T09:31:21.3220880Z Installed 295 packages in 804ms 2025-12-04T09:31:21.3224560Z - absl-py==2.3.1 2025-12-04T09:31:21.3227499Z + absl-py==2.1.0 2025-12-04T09:31:21.3228117Z + accelerate==1.0.1 2025-12-04T09:31:21.3228720Z + aenum==3.1.16 2025-12-04T09:31:21.3229348Z + affine==2.4.0 2025-12-04T09:31:21.3229847Z - aiohttp==3.13.2 2025-12-04T09:31:21.3230917Z + aiohttp==3.13.0 2025-12-04T09:31:21.3231765Z + aiohttp-cors==0.8.1 2025-12-04T09:31:21.3232795Z + albucore==0.0.16 2025-12-04T09:31:21.3234183Z + albumentations==1.4.6 2025-12-04T09:31:21.3235251Z + alembic==1.16.4 2025-12-04T09:31:21.3235845Z + antlr4-python3-runtime==4.9.3 2025-12-04T09:31:21.3236917Z - anyio==4.12.0 2025-12-04T09:31:21.3238581Z + anyio==4.6.2.post1 2025-12-04T09:31:21.3239232Z + argcomplete==3.5.1 2025-12-04T09:31:21.3240145Z + arrow==1.3.0 2025-12-04T09:31:21.3240907Z - attrs==25.4.0 2025-12-04T09:31:21.3241744Z + attrs==24.2.0 2025-12-04T09:31:21.3242976Z - audioread==3.1.0 2025-12-04T09:31:21.3243597Z + audioread==3.0.1 2025-12-04T09:31:21.3244671Z + backoff==2.2.1 2025-12-04T09:31:21.3245323Z + bitsandbytes==0.46.1 2025-12-04T09:31:21.3246268Z + black==24.10.0 2025-12-04T09:31:21.3247305Z + blinker==1.9.0 2025-12-04T09:31:21.3247839Z + blobfile==3.0.0 2025-12-04T09:31:21.3248889Z + bm25s==0.2.13 2025-12-04T09:31:21.3249611Z - boto3==1.35.42 2025-12-04T09:31:21.3251260Z + boto3==1.35.57 2025-12-04T09:31:21.3252005Z - botocore==1.35.99 2025-12-04T09:31:21.3253485Z + botocore==1.35.57 2025-12-04T09:31:21.3254305Z + bounded-pool-executor==0.0.3 2025-12-04T09:31:21.3255305Z + buildkite-test-collector==0.1.9 2025-12-04T09:31:21.3256906Z - cachetools==6.2.2 2025-12-04T09:31:21.3257885Z + cachetools==5.5.2 2025-12-04T09:31:21.3258822Z - certifi==2025.11.12 2025-12-04T09:31:21.3259899Z + certifi==2024.8.30 2025-12-04T09:31:21.3261029Z - cffi==2.0.0 2025-12-04T09:31:21.3261570Z + cffi==1.17.1 2025-12-04T09:31:21.3262394Z + chardet==5.2.0 2025-12-04T09:31:21.3262903Z - charset-normalizer==3.4.4 2025-12-04T09:31:21.3263523Z + charset-normalizer==3.4.0 2025-12-04T09:31:21.3264203Z + chz==0.3.0 2025-12-04T09:31:21.3264988Z - click==8.2.1 2025-12-04T09:31:21.3265419Z + click==8.1.7 2025-12-04T09:31:21.3266590Z + click-plugins==1.1.1.2 2025-12-04T09:31:21.3267079Z + cligj==0.7.2 2025-12-04T09:31:21.3268740Z - cloudpickle==3.1.2 2025-12-04T09:31:21.3269142Z + cloudpickle==3.1.1 2025-12-04T09:31:21.3270449Z + colorful==0.5.6 2025-12-04T09:31:21.3271307Z + contourpy==1.3.0 2025-12-04T09:31:21.3273781Z + coverage==7.10.6 2025-12-04T09:31:21.3274309Z + cramjam==2.9.0 2025-12-04T09:31:21.3274721Z + cycler==0.12.1 2025-12-04T09:31:21.3275304Z + databricks-sdk==0.59.0 2025-12-04T09:31:21.3276677Z + datamodel-code-generator==0.26.3 2025-12-04T09:31:21.3277206Z + dataproperty==1.0.1 2025-12-04T09:31:21.3278070Z + datasets==3.0.2 2025-12-04T09:31:21.3278881Z - decorator==5.2.1 2025-12-04T09:31:21.3279881Z + decorator==5.1.1 2025-12-04T09:31:21.3280647Z + decord==0.6.0 2025-12-04T09:31:21.3282216Z - dill==0.3.7 2025-12-04T09:31:21.3282808Z + dill==0.3.8 2025-12-04T09:31:21.3284086Z + distlib==0.3.9 2025-12-04T09:31:21.3285353Z - dnspython==2.8.0 2025-12-04T09:31:21.3286159Z + dnspython==2.7.0 2025-12-04T09:31:21.3286752Z + docopt==0.6.2 2025-12-04T09:31:21.3287375Z + docstring-parser==0.17.0 2025-12-04T09:31:21.3288524Z + efficientnet-pytorch==0.7.1 2025-12-04T09:31:21.3289276Z + einx==0.3.0 2025-12-04T09:31:21.3289780Z - email-validator==2.3.0 2025-12-04T09:31:21.3290948Z + email-validator==2.2.0 2025-12-04T09:31:21.3291405Z + encodec==0.1.1 2025-12-04T09:31:21.3292122Z + eval-type-backport==0.2.2 2025-12-04T09:31:21.3293099Z + evaluate==0.4.3 2025-12-04T09:31:21.3294237Z - fastapi==0.123.7 2025-12-04T09:31:21.3295312Z + fastapi==0.116.1 2025-12-04T09:31:21.3296462Z + fastparquet==2024.11.0 2025-12-04T09:31:21.3297317Z - fastrlock==0.8.3 2025-12-04T09:31:21.3298447Z + fastrlock==0.8.2 2025-12-04T09:31:21.3299272Z + fastsafetensors==0.1.10 2025-12-04T09:31:21.3300167Z - filelock==3.18.0 2025-12-04T09:31:21.3301311Z + filelock==3.16.1 2025-12-04T09:31:21.3301724Z + fiona==1.10.1 2025-12-04T09:31:21.3302708Z + flask==3.1.1 2025-12-04T09:31:21.3303365Z + fonttools==4.55.0 2025-12-04T09:31:21.3304273Z + fqdn==1.5.1 2025-12-04T09:31:21.3305168Z - frozendict==2.4.7 2025-12-04T09:31:21.3305719Z + frozendict==2.4.6 2025-12-04T09:31:21.3307034Z - frozenlist==1.8.0 2025-12-04T09:31:21.3308037Z + frozenlist==1.5.0 2025-12-04T09:31:21.3308683Z - fsspec==2025.10.0 2025-12-04T09:31:21.3309916Z + fsspec==2024.9.0 2025-12-04T09:31:21.3310618Z + ftfy==6.3.1 2025-12-04T09:31:21.3311830Z + genai-perf==0.0.8 2025-12-04T09:31:21.3312886Z + genson==1.3.0 2025-12-04T09:31:21.3314593Z + geopandas==1.0.1 2025-12-04T09:31:21.3315161Z - gitpython==3.1.45 2025-12-04T09:31:21.3316856Z + gitpython==3.1.44 2025-12-04T09:31:21.3317561Z + google-api-core==2.24.2 2025-12-04T09:31:21.3318460Z - google-auth==2.43.0 2025-12-04T09:31:21.3319753Z + google-auth==2.40.2 2025-12-04T09:31:21.3320609Z + google-cloud-core==2.4.3 2025-12-04T09:31:21.3322273Z + google-cloud-storage==3.4.0 2025-12-04T09:31:21.3322780Z + google-crc32c==1.7.1 2025-12-04T09:31:21.3324095Z + google-resumable-media==2.7.2 2025-12-04T09:31:21.3324675Z + googleapis-common-protos==1.70.0 2025-12-04T09:31:21.3325502Z + gpt-oss==0.0.8 2025-12-04T09:31:21.3326691Z + graphene==3.4.3 2025-12-04T09:31:21.3327155Z + graphql-core==3.2.6 2025-12-04T09:31:21.3328528Z + graphql-relay==3.2.0 2025-12-04T09:31:21.3329510Z + greenlet==3.2.3 2025-12-04T09:31:21.3330684Z - grpcio==1.76.0 2025-12-04T09:31:21.3331322Z + grpcio==1.71.0 2025-12-04T09:31:21.3331985Z + gunicorn==23.0.0 2025-12-04T09:31:21.3333134Z - h11==0.16.0 2025-12-04T09:31:21.3334130Z + h11==0.14.0 2025-12-04T09:31:21.3335132Z + h5py==3.13.0 2025-12-04T09:31:21.3336735Z + harfile==0.3.0 2025-12-04T09:31:21.3337332Z - hf-xet==1.2.0 2025-12-04T09:31:21.3338710Z + hf-xet==1.1.7 2025-12-04T09:31:21.3339530Z + hiredis==3.0.0 2025-12-04T09:31:21.3339946Z + html2text==2025.4.15 2025-12-04T09:31:21.3340768Z - httpcore==1.0.9 2025-12-04T09:31:21.3341717Z + httpcore==1.0.6 2025-12-04T09:31:21.3342276Z - httpx==0.28.1 2025-12-04T09:31:21.3343313Z + httpx==0.27.2 2025-12-04T09:31:21.3344018Z - huggingface-hub==0.36.0 2025-12-04T09:31:21.3345042Z + huggingface-hub==0.34.3 2025-12-04T09:31:21.3345694Z + humanize==4.11.0 2025-12-04T09:31:21.3346657Z + hydra-core==1.3.2 2025-12-04T09:31:21.3347402Z - hypothesis==6.56.4 2025-12-04T09:31:21.3348188Z + hypothesis==6.131.0 2025-12-04T09:31:21.3349173Z + hypothesis-graphql==0.11.1 2025-12-04T09:31:21.3349897Z + hypothesis-jsonschema==0.23.1 2025-12-04T09:31:21.3350750Z - idna==3.11 2025-12-04T09:31:21.3351597Z + idna==3.10 2025-12-04T09:31:21.3352575Z - imageio==2.37.2 2025-12-04T09:31:21.3353532Z + imageio==2.37.0 2025-12-04T09:31:21.3355132Z + importlib-metadata==8.7.0 2025-12-04T09:31:21.3355585Z + importlib-resources==6.5.2 2025-12-04T09:31:21.3356273Z + inflect==5.6.2 2025-12-04T09:31:21.3358205Z - iniconfig==2.3.0 2025-12-04T09:31:21.3359887Z + iniconfig==2.0.0 2025-12-04T09:31:21.3360359Z + isoduration==20.11.0 2025-12-04T09:31:21.3360882Z + isort==5.13.2 2025-12-04T09:31:21.3361413Z + itsdangerous==2.2.0 2025-12-04T09:31:21.3362344Z + jiwer==3.0.5 2025-12-04T09:31:21.3363289Z - joblib==1.5.2 2025-12-04T09:31:21.3364367Z + joblib==1.4.2 2025-12-04T09:31:21.3365506Z + jsonargparse==4.35.0 2025-12-04T09:31:21.3365996Z + jsonlines==4.0.0 2025-12-04T09:31:21.3367372Z + jsonpointer==3.0.0 2025-12-04T09:31:21.3367916Z - jsonschema==4.25.1 2025-12-04T09:31:21.3368920Z + jsonschema==4.23.0 2025-12-04T09:31:21.3369831Z - jsonschema-specifications==2025.9.1 2025-12-04T09:31:21.3371214Z + jsonschema-specifications==2024.10.1 2025-12-04T09:31:21.3371971Z + junit-xml==1.9 2025-12-04T09:31:21.3372622Z + kaleido==0.2.1 2025-12-04T09:31:21.3373638Z + kiwisolver==1.4.7 2025-12-04T09:31:21.3374394Z + kornia==0.8.1 2025-12-04T09:31:21.3375659Z + kornia-rs==0.1.9 2025-12-04T09:31:21.3376161Z + libnacl==2.1.0 2025-12-04T09:31:21.3377351Z - librosa==0.10.2 2025-12-04T09:31:21.3378276Z + librosa==0.10.2.post1 2025-12-04T09:31:21.3379632Z + lightly==1.5.20 2025-12-04T09:31:21.3380769Z + lightly-utils==0.0.2 2025-12-04T09:31:21.3381874Z + lightning==2.5.1.post0 2025-12-04T09:31:21.3382618Z + lightning-utilities==0.14.3 2025-12-04T09:31:21.3384487Z + lm-eval==0.4.9.1 (from git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d) 2025-12-04T09:31:21.3385405Z + mako==1.3.10 2025-12-04T09:31:21.3386344Z - markdown==3.10 2025-12-04T09:31:21.3387334Z + markdown==3.8.2 2025-12-04T09:31:21.3388011Z - markdown-it-py==4.0.0 2025-12-04T09:31:21.3389252Z + markdown-it-py==3.0.0 2025-12-04T09:31:21.3390206Z - markupsafe==3.0.3 2025-12-04T09:31:21.3391189Z + markupsafe==3.0.1 2025-12-04T09:31:21.3392600Z + matplotlib==3.9.2 2025-12-04T09:31:21.3393146Z + mbstrdecoder==1.1.3 2025-12-04T09:31:21.3394009Z - mistral-common==1.8.6 2025-12-04T09:31:21.3395086Z + mistral-common==1.8.5 2025-12-04T09:31:21.3396211Z + mlflow==2.22.0 2025-12-04T09:31:21.3397353Z + mlflow-skinny==2.22.0 2025-12-04T09:31:21.3398444Z + more-itertools==10.5.0 2025-12-04T09:31:21.3399430Z - msgpack==1.1.2 2025-12-04T09:31:21.3400075Z + msgpack==1.1.0 2025-12-04T09:31:21.3401075Z + mteb==1.38.11 2025-12-04T09:31:21.3402091Z - multidict==6.7.0 2025-12-04T09:31:21.3403170Z + multidict==6.1.0 2025-12-04T09:31:21.3404377Z + multiprocess==0.70.16 2025-12-04T09:31:21.3405125Z + munch==4.0.0 2025-12-04T09:31:21.3406689Z - mypy-extensions==1.1.0 2025-12-04T09:31:21.3407526Z + mypy-extensions==1.0.0 2025-12-04T09:31:21.3408369Z - networkx==2.8.8 2025-12-04T09:31:21.3409948Z + networkx==3.2.1 2025-12-04T09:31:21.3411055Z + nltk==3.9.1 2025-12-04T09:31:21.3412277Z + num2words==0.5.14 2025-12-04T09:31:21.3413255Z + numexpr==2.10.1 2025-12-04T09:31:21.3415121Z - numpy==2.2.6 2025-12-04T09:31:21.3415623Z + numpy==1.26.4 2025-12-04T09:31:21.3417053Z + omegaconf==2.3.0 2025-12-04T09:31:21.3417702Z + open-clip-torch==2.32.0 2025-12-04T09:31:21.3419228Z - openai-harmony==0.0.8 2025-12-04T09:31:21.3420279Z + openai-harmony==0.0.4 2025-12-04T09:31:21.3421189Z + opencensus==0.11.4 2025-12-04T09:31:21.3422145Z + opencensus-context==0.1.3 2025-12-04T09:31:21.3422978Z - opencv-python-headless==4.12.0.88 2025-12-04T09:31:21.3424422Z + opencv-python-headless==4.11.0.86 2025-12-04T09:31:21.3425236Z + opentelemetry-api==1.35.0 2025-12-04T09:31:21.3425825Z + opentelemetry-exporter-prometheus==0.56b0 2025-12-04T09:31:21.3426455Z + opentelemetry-proto==1.36.0 2025-12-04T09:31:21.3427040Z + opentelemetry-sdk==1.35.0 2025-12-04T09:31:21.3427866Z + opentelemetry-semantic-conventions==0.56b0 2025-12-04T09:31:21.3428492Z - packaging==25.0 2025-12-04T09:31:21.3429005Z + packaging==24.2 2025-12-04T09:31:21.3429659Z - pandas==2.0.3 2025-12-04T09:31:21.3430268Z + pandas==2.2.3 2025-12-04T09:31:21.3431391Z + pathvalidate==3.2.1 2025-12-04T09:31:21.3432094Z + patsy==1.0.1 2025-12-04T09:31:21.3433368Z + peft==0.16.0 2025-12-04T09:31:21.3434377Z - pillow==11.0.0 2025-12-04T09:31:21.3435758Z + pillow==10.4.0 2025-12-04T09:31:21.3436185Z - platformdirs==4.5.0 2025-12-04T09:31:21.3436842Z + platformdirs==4.3.6 2025-12-04T09:31:21.3437142Z + plotly==5.24.1 2025-12-04T09:31:21.3437684Z - pluggy==1.6.0 2025-12-04T09:31:21.3438828Z + pluggy==1.5.0 2025-12-04T09:31:21.3439285Z + polars==1.29.0 2025-12-04T09:31:21.3439747Z + portalocker==2.10.1 2025-12-04T09:31:21.3440493Z + pqdm==0.2.0 2025-12-04T09:31:21.3441489Z + pretrainedmodels==0.7.4 2025-12-04T09:31:21.3442032Z - prometheus-client==0.23.1 2025-12-04T09:31:21.3442693Z + prometheus-client==0.22.0 2025-12-04T09:31:21.3443665Z - propcache==0.4.1 2025-12-04T09:31:21.3444410Z + propcache==0.2.0 2025-12-04T09:31:21.3445603Z + proto-plus==1.26.1 2025-12-04T09:31:21.3446213Z - protobuf==5.29.5 2025-12-04T09:31:21.3446889Z + protobuf==5.28.3 2025-12-04T09:31:21.3447641Z - psutil==7.1.3 2025-12-04T09:31:21.3448161Z + psutil==6.1.0 2025-12-04T09:31:21.3448898Z + py==1.11.0 2025-12-04T09:31:21.3449718Z + py-spy==0.4.0 2025-12-04T09:31:21.3450186Z + pyarrow==18.0.0 2025-12-04T09:31:21.3451277Z + pybind11==2.13.6 2025-12-04T09:31:21.3451894Z + pycocotools==2.0.8 2025-12-04T09:31:21.3452806Z - pycparser==2.23 2025-12-04T09:31:21.3453300Z + pycparser==2.22 2025-12-04T09:31:21.3453963Z + pycryptodomex==3.22.0 2025-12-04T09:31:21.3454643Z - pydantic==2.12.5 2025-12-04T09:31:21.3455361Z + pydantic==2.12.0 2025-12-04T09:31:21.3456674Z - pydantic-core==2.41.5 2025-12-04T09:31:21.3457305Z + pydantic-core==2.41.1 2025-12-04T09:31:21.3457939Z - pydantic-extra-types==2.10.6 2025-12-04T09:31:21.3458923Z + pydantic-extra-types==2.10.5 2025-12-04T09:31:21.3459540Z - pygments==2.15.0 2025-12-04T09:31:21.3460287Z + pygments==2.18.0 2025-12-04T09:31:21.3460959Z + pyogrio==0.11.0 2025-12-04T09:31:21.3462569Z + pyparsing==3.2.0 2025-12-04T09:31:21.3462902Z + pyproj==3.7.1 2025-12-04T09:31:21.3463430Z + pyrate-limiter==3.7.0 2025-12-04T09:31:21.3464878Z + pystemmer==3.0.0 2025-12-04T09:31:21.3465415Z + pytablewriter==1.2.0 2025-12-04T09:31:21.3466111Z - pytest==7.3.2 2025-12-04T09:31:21.3466872Z + pytest==8.3.5 2025-12-04T09:31:21.3467973Z + pytest-asyncio==0.24.0 2025-12-04T09:31:21.3468776Z + pytest-cov==6.3.0 2025-12-04T09:31:21.3469700Z + pytest-forked==1.6.0 2025-12-04T09:31:21.3470916Z + pytest-mock==3.14.0 2025-12-04T09:31:21.3471580Z + pytest-shard==0.1.2 2025-12-04T09:31:21.3472360Z - pytest-subtests==0.13.1 2025-12-04T09:31:21.3473436Z + pytest-subtests==0.14.1 2025-12-04T09:31:21.3473857Z + pytest-timeout==2.3.1 2025-12-04T09:31:21.3474521Z + python-box==7.3.2 2025-12-04T09:31:21.3474862Z + python-rapidjson==1.20 2025-12-04T09:31:21.3475856Z + pytorch-lightning==2.6.0 2025-12-04T09:31:21.3476325Z + pytrec-eval-terrier==0.5.7 2025-12-04T09:31:21.3477394Z - pytz==2025.2 2025-12-04T09:31:21.3478009Z + pytz==2024.2 2025-12-04T09:31:21.3478739Z + rapidfuzz==3.12.1 2025-12-04T09:31:21.3479792Z + rasterio==1.4.3 2025-12-04T09:31:21.3480558Z - ray==2.52.1 2025-12-04T09:31:21.3481682Z + ray==2.48.0 2025-12-04T09:31:21.3482417Z - redis==7.1.0 2025-12-04T09:31:21.3482883Z + redis==5.2.0 2025-12-04T09:31:21.3483712Z - referencing==0.37.0 2025-12-04T09:31:21.3484404Z + referencing==0.35.1 2025-12-04T09:31:21.3484953Z - regex==2025.11.3 2025-12-04T09:31:21.3486359Z + regex==2024.9.11 2025-12-04T09:31:21.3486891Z - requests==2.32.5 2025-12-04T09:31:21.3487942Z + requests==2.32.3 2025-12-04T09:31:21.3488658Z + responses==0.25.3 2025-12-04T09:31:21.3489467Z + rfc3339-validator==0.1.4 2025-12-04T09:31:21.3490192Z + rfc3987==1.3.8 2025-12-04T09:31:21.3491013Z - rich==14.2.0 2025-12-04T09:31:21.3491898Z + rich==13.9.4 2025-12-04T09:31:21.3492767Z + rioxarray==0.19.0 2025-12-04T09:31:21.3493448Z + rouge-score==0.1.2 2025-12-04T09:31:21.3494455Z - rpds-py==0.30.0 2025-12-04T09:31:21.3495707Z + rpds-py==0.20.1 2025-12-04T09:31:21.3496253Z + rtree==1.4.0 2025-12-04T09:31:21.3496926Z + runai-model-streamer==0.14.0 2025-12-04T09:31:21.3497720Z + runai-model-streamer-gcs==0.14.0 2025-12-04T09:31:21.3498791Z + runai-model-streamer-s3==0.14.0 2025-12-04T09:31:21.3499456Z - s3transfer==0.10.4 2025-12-04T09:31:21.3500664Z + s3transfer==0.10.3 2025-12-04T09:31:21.3501057Z + sacrebleu==2.4.3 2025-12-04T09:31:21.3501790Z - safetensors==0.7.0 2025-12-04T09:31:21.3502787Z + safetensors==0.4.5 2025-12-04T09:31:21.3503816Z + schemathesis==3.39.15 2025-12-04T09:31:21.3504485Z - scikit-image==0.22.0 2025-12-04T09:31:21.3505304Z + scikit-image==0.25.2 2025-12-04T09:31:21.3506409Z - scikit-learn==1.7.2 2025-12-04T09:31:21.3507055Z + scikit-learn==1.5.2 2025-12-04T09:31:21.3507571Z - scipy==1.14.1 2025-12-04T09:31:21.3508147Z + scipy==1.13.1 2025-12-04T09:31:21.3509145Z + segmentation-models-pytorch==0.4.0 2025-12-04T09:31:21.3510038Z + sentence-transformers==3.2.1 2025-12-04T09:31:21.3511281Z - setuptools==78.1.1 2025-12-04T09:31:21.3512553Z + setuptools==77.0.3 2025-12-04T09:31:21.3513484Z + shapely==2.1.1 2025-12-04T09:31:21.3514689Z - six==1.17.0 2025-12-04T09:31:21.3515175Z + six==1.16.0 2025-12-04T09:31:21.3516444Z + smart-open==7.1.0 2025-12-04T09:31:21.3517628Z - soundfile==0.13.1 2025-12-04T09:31:21.3518902Z + soundfile==0.12.1 2025-12-04T09:31:21.3519866Z - soxr==1.0.0 2025-12-04T09:31:21.3521049Z + soxr==0.5.0.post1 2025-12-04T09:31:21.3522287Z + sqlalchemy==2.0.41 2025-12-04T09:31:21.3523380Z + sqlitedict==2.1.0 2025-12-04T09:31:21.3523874Z + sqlparse==0.5.3 2025-12-04T09:31:21.3525108Z - starlette==0.50.0 2025-12-04T09:31:21.3525808Z + starlette==0.46.2 2025-12-04T09:31:21.3526661Z + starlette-testclient==0.4.1 2025-12-04T09:31:21.3527076Z + statsmodels==0.14.4 2025-12-04T09:31:21.3528017Z + structlog==25.4.0 2025-12-04T09:31:21.3528801Z + tabledata==1.3.3 2025-12-04T09:31:21.3529595Z + tblib==3.1.0 2025-12-04T09:31:21.3530250Z + tcolorpy==0.1.6 2025-12-04T09:31:21.3530969Z + tenacity==9.1.2 2025-12-04T09:31:21.3531738Z + tensorboardx==2.6.4 2025-12-04T09:31:21.3532420Z + tensorizer==2.10.1 2025-12-04T09:31:21.3533227Z + termcolor==3.1.0 2025-12-04T09:31:21.3535205Z + terratorch==1.0.2 (from git+https://github.com/IBM/terratorch.git@07184fcf91a1324f831ff521dd238d97fe350e3e) 2025-12-04T09:31:21.3536026Z - threadpoolctl==3.6.0 2025-12-04T09:31:21.3536964Z + threadpoolctl==3.5.0 2025-12-04T09:31:21.3538106Z - tifffile==2025.10.16 2025-12-04T09:31:21.3539398Z + tifffile==2025.3.30 2025-12-04T09:31:21.3540888Z + timm==1.0.17 2025-12-04T09:31:21.3541875Z - tokenizers==0.22.1 2025-12-04T09:31:21.3543373Z + tokenizers==0.22.0 2025-12-04T09:31:21.3543979Z + tomli==2.2.1 2025-12-04T09:31:21.3544771Z + tomli-w==1.2.0 2025-12-04T09:31:21.3546205Z - torch==2.10.0a0+gitffd9b0f (from file:///var/lib/jenkins/workspace/dist/torch-2.10.0a0%2Bgitffd9b0f-cp312-cp312-linux_x86_64.whl) 2025-12-04T09:31:21.3547945Z + torch==2.10.0a0+gitffd9b0f (from file:///var/lib/jenkins/workspace/dist/torch-2.10.0a0+gitffd9b0f-cp312-cp312-linux_x86_64.whl#sha256=a0c782d08e3b01a62f42b12b7832d890773e7e8ac303ff47ce3c04911a4563e9) 2025-12-04T09:31:21.3549983Z - torchaudio==2.10.0a0+e90a398 (from file:///var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0%2Be90a398-cp312-cp312-linux_x86_64.whl) 2025-12-04T09:31:21.3551577Z + torchaudio==2.10.0a0+e90a398 (from file:///var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0+e90a398-cp312-cp312-linux_x86_64.whl#sha256=9fd6d716ab57b84fbbf009e8463967c8254f16e45d6b9c9618575442317da3e1) 2025-12-04T09:31:21.3552564Z + torchgeo==0.7.0 2025-12-04T09:31:21.3553002Z + torchmetrics==1.8.2 2025-12-04T09:31:21.3553873Z - torchvision==0.25.0a0+617079d (from file:///var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0%2B617079d-cp312-cp312-linux_x86_64.whl) 2025-12-04T09:31:21.3555773Z + torchvision==0.25.0a0+617079d (from file:///var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0+617079d-cp312-cp312-linux_x86_64.whl#sha256=fc250b889759c63f19d9bb8fd5404f56b8b0b7c10d6e052aa34730d9dcb2b4a4) 2025-12-04T09:31:21.3556861Z - tqdm==4.67.1 2025-12-04T09:31:21.3557208Z + tqdm==4.66.6 2025-12-04T09:31:21.3558137Z + tqdm-multiprocess==0.0.11 2025-12-04T09:31:21.3559505Z - transformers==4.57.3 2025-12-04T09:31:21.3560123Z + transformers==4.56.2 2025-12-04T09:31:21.3560828Z + transformers-stream-generator==0.0.5 2025-12-04T09:31:21.3561692Z + tritonclient==2.51.0 2025-12-04T09:31:21.3562211Z + typepy==1.3.2 2025-12-04T09:31:21.3563291Z - typer==0.20.0 2025-12-04T09:31:21.3563598Z + typer==0.15.2 2025-12-04T09:31:21.3565046Z + types-python-dateutil==2.9.0.20241206 2025-12-04T09:31:21.3565749Z + typeshed-client==2.8.2 2025-12-04T09:31:21.3566319Z - tzdata==2025.2 2025-12-04T09:31:21.3567237Z + tzdata==2024.2 2025-12-04T09:31:21.3568210Z + uri-template==1.3.0 2025-12-04T09:31:21.3568917Z - urllib3==2.5.0 2025-12-04T09:31:21.3569524Z + urllib3==2.2.3 2025-12-04T09:31:21.3570252Z - uvicorn==0.38.0 2025-12-04T09:31:21.3571375Z + uvicorn==0.35.0 2025-12-04T09:31:21.3572155Z + vector-quantize-pytorch==1.27.7 2025-12-04T09:31:21.3572875Z + virtualenv==20.31.2 2025-12-04T09:31:21.3573776Z + vocos==0.1.0 2025-12-04T09:31:21.3575004Z + wcwidth==0.2.13 2025-12-04T09:31:21.3575932Z + webcolors==24.11.1 2025-12-04T09:31:21.3576749Z - werkzeug==3.1.4 2025-12-04T09:31:21.3578425Z + werkzeug==3.1.3 2025-12-04T09:31:21.3578925Z + word2number==1.1 2025-12-04T09:31:21.3579639Z - wrapt==2.0.1 2025-12-04T09:31:21.3581097Z + wrapt==1.17.2 2025-12-04T09:31:21.3582018Z + xarray==2025.7.1 2025-12-04T09:31:21.3582830Z + xxhash==3.5.0 2025-12-04T09:31:21.3584474Z - yarl==1.22.0 2025-12-04T09:31:21.3585439Z + yarl==1.17.1 2025-12-04T09:31:21.3586777Z + zipp==3.23.0 2025-12-04T09:31:21.3587674Z + zstandard==0.23.0 2025-12-04T09:31:21.5301475Z 2025-12-04 09:31:21,529 [INFO] cli.lib.core.vllm.vllm_test: Done. installed requirements for test dependencies 2025-12-04T09:31:21.5302372Z 2025-12-04 09:31:21,529 [INFO] cli.lib.core.vllm.vllm_test: Double check installed packages 2025-12-04T09:31:21.5348657Z 2025-12-04 09:31:21,534 [INFO] cli.lib.common.pip_helper: torch already exist with version: 2.10.0a0+gitffd9b0f 2025-12-04T09:31:21.5352041Z 2025-12-04 09:31:21,534 [INFO] cli.lib.common.pip_helper: xformers already exist with version: 0.0.34+3f91ad6f.d20251204 2025-12-04T09:31:21.5355830Z 2025-12-04 09:31:21,535 [INFO] cli.lib.common.pip_helper: torchvision already exist with version: 0.25.0a0+617079d 2025-12-04T09:31:21.5360732Z 2025-12-04 09:31:21,535 [INFO] cli.lib.common.pip_helper: torchaudio already exist with version: 2.10.0a0+e90a398 2025-12-04T09:31:21.5365238Z 2025-12-04 09:31:21,536 [INFO] cli.lib.common.pip_helper: vllm already exist with version: 0.11.0rc2.dev389+ge51928192.d20251204 2025-12-04T09:31:21.5365861Z 2025-12-04 09:31:21,536 [INFO] cli.lib.core.vllm.vllm_test: Done. checked installed packages 2025-12-04T09:31:21.5366360Z 2025-12-04 09:31:21,536 [INFO] cli.lib.core.vllm.lib: run vllm tests..... 2025-12-04T09:31:21.5366832Z 2025-12-04 09:31:21,536 [INFO] cli.lib.core.vllm.lib: Running tests: Basic models test 2025-12-04T09:31:21.5367358Z 2025-12-04 09:31:21,536 [INFO] cli.lib.core.vllm.lib: Running step: pytest -v -s models/test_transformers.py 2025-12-04T09:31:21.5367878Z 2025-12-04 09:31:21,536 [INFO] cli.lib.common.utils: [shell] pytest -v -s models/test_transformers.py 2025-12-04T09:31:31.6728799Z INFO 12-04 09:31:31 [__init__.py:224] Automatically detected platform cuda. 2025-12-04T09:31:34.1253430Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_asyncio/plugin.py:208: PytestDeprecationWarning: The configuration option "asyncio_default_fixture_loop_scope" is unset. 2025-12-04T09:31:34.1255748Z The event loop scope for asynchronous fixtures will default to the fixture caching scope. Future versions of pytest-asyncio will default the loop scope for asynchronous fixtures to function scope. Set the default fixture loop scope explicitly in order to avoid unexpected behavior in the future. Valid fixture loop scopes are: "function", "class", "module", "package", "session" 2025-12-04T09:31:34.1257411Z 2025-12-04T09:31:34.1257692Z warnings.warn(PytestDeprecationWarning(_DEFAULT_FIXTURE_LOOP_SCOPE_UNSET)) 2025-12-04T09:31:34.1324375Z ============================= test session starts ============================== 2025-12-04T09:31:34.1325117Z platform linux -- Python 3.12.5, pytest-8.3.5, pluggy-1.5.0 -- /opt/conda/envs/py_3.12/bin/python 2025-12-04T09:31:34.1418888Z cachedir: .pytest_cache 2025-12-04T09:31:34.1419715Z hypothesis profile 'ci' -> database=None, deadline=None, print_blob=True, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:31:34.1420250Z rootdir: /var/lib/jenkins/workspace/vllm 2025-12-04T09:31:34.1420493Z configfile: pyproject.toml 2025-12-04T09:31:34.1421310Z plugins: cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, xdist-3.3.1, xdoctest-1.3.0, hydra-core-1.3.2, anyio-4.6.2.post1, buildkite-test-collector-0.1.9, forked-1.6.0, shard-0.1.2, mock-3.14.0, schemathesis-3.39.15, subtests-0.14.1, hypothesis-6.131.0, cov-6.3.0, asyncio-0.24.0, timeout-2.3.1 2025-12-04T09:31:34.1422198Z asyncio: mode=Mode.STRICT, default_loop_scope=None 2025-12-04T09:31:36.9164949Z collecting ... WARNING 12-04 09:31:36 [interface.py:514] Current platform cuda does not have '_pytestfixturefunction' attribute. 2025-12-04T09:31:37.3885390Z WARNING 12-04 09:31:37 [interface.py:514] Current platform cuda does not have '__test__' attribute. 2025-12-04T09:31:37.3885968Z WARNING 12-04 09:31:37 [interface.py:514] Current platform cuda does not have '__bases__' attribute. 2025-12-04T09:31:37.3886462Z WARNING 12-04 09:31:37 [interface.py:514] Current platform cuda does not have '__test__' attribute. 2025-12-04T09:31:37.3886991Z WARNING 12-04 09:31:37 [interface.py:514] Current platform cuda does not have '_schemathesis_test' attribute. 2025-12-04T09:31:37.3915737Z  2025-12-04T09:31:37.3919558Z collecting 12 items  2025-12-04T09:31:37.3920134Z collected 12 items  2025-12-04T09:31:37.3925256Z Running 12 items in this shard: tests/models/test_transformers.py::test_models[meta-llama/Llama-3.2-1B-Instruct-transformers], tests/models/test_transformers.py::test_models[hmellor/Ilama-3.2-1B-auto], tests/models/test_transformers.py::test_models[allenai/OLMoE-1B-7B-0924-transformers], tests/models/test_transformers.py::test_hybrid_attention, tests/models/test_transformers.py::test_distributed, tests/models/test_transformers.py::test_quantization[5-32-TheBloke/TinyLlama-1.1B-Chat-v0.3-AWQ-quantization_kwargs0], tests/models/test_transformers.py::test_quantization[5-32-TheBloke/TinyLlama-1.1B-Chat-v0.3-GPTQ-quantization_kwargs1], tests/models/test_transformers.py::test_quantization[5-32-meta-llama/Llama-3.2-1B-Instruct-quantization_kwargs2], tests/models/test_transformers.py::test_embed_loading[Qwen/Qwen3-Embedding-0.6B], tests/models/test_transformers.py::test_embed_loading[meta-llama/Llama-3.2-1B-Instruct], tests/models/test_transformers.py::test_pooling[TransformersEmbeddingModel], tests/models/test_transformers.py::test_pooling[TransformersForSequenceClassification] 2025-12-04T09:31:37.3930454Z 2025-12-04T09:31:37.4181081Z models/test_transformers.py::test_models[meta-llama/Llama-3.2-1B-Instruct-transformers] INFO 12-04 09:31:37 [utils.py:239] non-default args: {'trust_remote_code': True, 'seed': 0, 'max_model_len': 1024, 'block_size': 16, 'disable_log_stats': True, 'enable_chunked_prefill': False, 'compilation_config': {'level': None, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': None, 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': None, 'local_cache_dir': None}, 'model_impl': 'transformers', 'model': 'meta-llama/Llama-3.2-1B-Instruct'} 2025-12-04T09:31:37.5850827Z 2025-12-04T09:31:37.5852109Z config.json: 0% 0.00/877 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-12-04T09:31:49.2357668Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:31:49.2386814Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:31:49.2395522Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:31:49.2404015Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:31:49.2411559Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:31:49.2419603Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:31:49.2422466Z (EngineCore_DP0 pid=764) INFO 12-04 09:31:49 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:31:49.5501389Z (EngineCore_DP0 pid=764) INFO 12-04 09:31:49 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:31:49.6007977Z (EngineCore_DP0 pid=764) INFO 12-04 09:31:49 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-12-04T09:31:49.9025566Z (EngineCore_DP0 pid=764) INFO 12-04 09:31:49 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:31:49.9027185Z (EngineCore_DP0 pid=764) INFO 12-04 09:31:49 [transformers.py:493] Using Transformers backend. 2025-12-04T09:31:49.9032910Z (EngineCore_DP0 pid=764) `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:31:50.1048231Z (EngineCore_DP0 pid=764) INFO 12-04 09:31:50 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:31:50.2424038Z (EngineCore_DP0 pid=764) INFO 12-04 09:31:50 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-12-04T09:31:50.3343175Z (EngineCore_DP0 pid=764) 2025-12-04T09:31:50.8941500Z model.safetensors: 0% 0.00/2.47G [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-12-04T09:33:31.3602778Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:33:31.3632683Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:33:31.3638413Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:33:31.3645477Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:33:31.3652248Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:33:31.3659373Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:33:31.3661813Z (EngineCore_DP0 pid=1309) INFO 12-04 09:33:31 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:33:31.6185508Z (EngineCore_DP0 pid=1309) INFO 12-04 09:33:31 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:33:31.8203742Z (EngineCore_DP0 pid=1309) WARNING 12-04 09:33:31 [utils.py:188] TransformersForCausalLM has no vLLM implementation, falling back to Transformers implementation. Some features may not be supported and performance may not be optimal. 2025-12-04T09:33:31.8207306Z (EngineCore_DP0 pid=1309) INFO 12-04 09:33:31 [gpu_model_runner.py:2840] Starting to load model hmellor/Ilama-3.2-1B... 2025-12-04T09:33:32.0724552Z (EngineCore_DP0 pid=1309) INFO 12-04 09:33:32 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:33:32.0725500Z (EngineCore_DP0 pid=1309) INFO 12-04 09:33:32 [transformers.py:493] Using Transformers backend. 2025-12-04T09:33:32.1022486Z (EngineCore_DP0 pid=1309) `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:33:32.2822670Z (EngineCore_DP0 pid=1309) INFO 12-04 09:33:32 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:33:32.4032985Z (EngineCore_DP0 pid=1309) INFO 12-04 09:33:32 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-12-04T09:33:32.4853770Z (EngineCore_DP0 pid=1309) 2025-12-04T09:33:33.1971771Z model.safetensors: 0% 0.00/4.94G [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-12-04T09:34:32.6916523Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:34:32.6946419Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:34:32.6952972Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:34:32.6959799Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:34:32.6966422Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:34:32.6973576Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:34:32.6976021Z (EngineCore_DP0 pid=1700) INFO 12-04 09:34:32 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:34:32.9529859Z (EngineCore_DP0 pid=1700) INFO 12-04 09:34:32 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:34:33.0075581Z (EngineCore_DP0 pid=1700) INFO 12-04 09:34:33 [gpu_model_runner.py:2840] Starting to load model hmellor/tiny-random-Gemma2ForCausalLM... 2025-12-04T09:34:33.2524860Z (EngineCore_DP0 pid=1700) INFO 12-04 09:34:33 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:34:33.2525850Z (EngineCore_DP0 pid=1700) INFO 12-04 09:34:33 [transformers.py:493] Using Transformers backend. 2025-12-04T09:34:33.2527733Z (EngineCore_DP0 pid=1700) `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:34:33.3534454Z (EngineCore_DP0 pid=1700) INFO 12-04 09:34:33 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:34:33.4146973Z (EngineCore_DP0 pid=1700) INFO 12-04 09:34:33 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-12-04T09:34:33.5446370Z (EngineCore_DP0 pid=1700) 2025-12-04T09:34:33.5950673Z model.safetensors: 0% 0.00/16.9M [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-12-04T09:34:57.7992838Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:34:57.8021934Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:34:57.8029351Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:34:57.8036467Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:34:57.8043439Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:34:57.8050559Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:34:57.8053106Z (EngineCore_DP0 pid=2001) INFO 12-04 09:34:57 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:34:58.0637794Z (EngineCore_DP0 pid=2001) INFO 12-04 09:34:58 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:34:58.0940021Z (EngineCore_DP0 pid=2001) INFO 12-04 09:34:58 [gpu_model_runner.py:2840] Starting to load model hmellor/tiny-random-Gemma2ForCausalLM... 2025-12-04T09:34:58.3406974Z (EngineCore_DP0 pid=2001) INFO 12-04 09:34:58 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:34:58.4047378Z (EngineCore_DP0 pid=2001) INFO 12-04 09:34:58 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:34:58.4798994Z (EngineCore_DP0 pid=2001) INFO 12-04 09:34:58 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-12-04T09:34:58.5371312Z (EngineCore_DP0 pid=2001) INFO 12-04 09:34:58 [weight_utils.py:480] No model.safetensors.index.json found in remote. 2025-12-04T09:34:58.5377307Z (EngineCore_DP0 pid=2001) 2025-12-04T09:34:58.5505468Z Loading safetensors checkpoint shards: 0% 0/1 [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-12-04T09:35:22.4746232Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:22.4775699Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:22.4782725Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:22.4790136Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:22.4797093Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:22.4804065Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:22.4806649Z (EngineCore_DP0 pid=2252) INFO 12-04 09:35:22 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:35:22.7443312Z (EngineCore_DP0 pid=2252) INFO 12-04 09:35:22 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:35:22.7937022Z (EngineCore_DP0 pid=2252) INFO 12-04 09:35:22 [gpu_model_runner.py:2840] Starting to load model TheBloke/TinyLlama-1.1B-Chat-v0.3-AWQ... 2025-12-04T09:35:23.0526273Z (EngineCore_DP0 pid=2252) INFO 12-04 09:35:23 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:35:23.1166010Z (EngineCore_DP0 pid=2252) INFO 12-04 09:35:23 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:35:23.2189763Z (EngineCore_DP0 pid=2252) INFO 12-04 09:35:23 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-12-04T09:35:23.3003870Z (EngineCore_DP0 pid=2252) 2025-12-04T09:35:24.2275079Z model.safetensors: 0% 0.00/766M [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-12-04T09:35:37.8367938Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:37.8397552Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:37.8404305Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:37.8411515Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:37.8418368Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:37.8425568Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:37.8428127Z (EngineCore_DP0 pid=2418) INFO 12-04 09:35:37 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:35:38.0960625Z (EngineCore_DP0 pid=2418) INFO 12-04 09:35:38 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:35:38.1667544Z (EngineCore_DP0 pid=2418) INFO 12-04 09:35:38 [gpu_model_runner.py:2840] Starting to load model TheBloke/TinyLlama-1.1B-Chat-v0.3-AWQ... 2025-12-04T09:35:38.4167205Z (EngineCore_DP0 pid=2418) INFO 12-04 09:35:38 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:35:38.4168549Z (EngineCore_DP0 pid=2418) INFO 12-04 09:35:38 [transformers.py:493] Using Transformers backend. 2025-12-04T09:35:38.4171394Z (EngineCore_DP0 pid=2418) `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:35:38.5942108Z (EngineCore_DP0 pid=2418) INFO 12-04 09:35:38 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:35:38.6623913Z (EngineCore_DP0 pid=2418) INFO 12-04 09:35:38 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-12-04T09:35:38.7195767Z (EngineCore_DP0 pid=2418) INFO 12-04 09:35:38 [weight_utils.py:480] No model.safetensors.index.json found in remote. 2025-12-04T09:35:38.7200995Z (EngineCore_DP0 pid=2418) 2025-12-04T09:35:38.8741059Z Loading safetensors checkpoint shards: 0% 0/1 [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-12-04T09:35:53.0119694Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:53.0148964Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:53.0155547Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:53.0162814Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:53.0169744Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:53.0176686Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:35:53.0179229Z (EngineCore_DP0 pid=2555) INFO 12-04 09:35:53 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:35:53.2764767Z (EngineCore_DP0 pid=2555) INFO 12-04 09:35:53 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:35:53.3222630Z (EngineCore_DP0 pid=2555) INFO 12-04 09:35:53 [gpu_model_runner.py:2840] Starting to load model TheBloke/TinyLlama-1.1B-Chat-v0.3-GPTQ... 2025-12-04T09:35:53.5674553Z (EngineCore_DP0 pid=2555) INFO 12-04 09:35:53 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:35:53.5689149Z (EngineCore_DP0 pid=2555) INFO 12-04 09:35:53 [gptq_marlin.py:357] Using MarlinLinearKernel for GPTQMarlinLinearMethod 2025-12-04T09:35:53.6322216Z (EngineCore_DP0 pid=2555) INFO 12-04 09:35:53 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:35:53.7737623Z (EngineCore_DP0 pid=2555) INFO 12-04 09:35:53 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-12-04T09:35:53.8567784Z (EngineCore_DP0 pid=2555) 2025-12-04T09:35:54.4657481Z model.safetensors: 0% 0.00/768M [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-12-04T09:36:08.6617970Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:08.6647351Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:08.6654401Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:08.6661567Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:08.6669144Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:08.6676609Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:08.6679166Z (EngineCore_DP0 pid=2744) INFO 12-04 09:36:08 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:36:08.9249421Z (EngineCore_DP0 pid=2744) INFO 12-04 09:36:08 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:36:08.9952372Z (EngineCore_DP0 pid=2744) INFO 12-04 09:36:08 [gpu_model_runner.py:2840] Starting to load model TheBloke/TinyLlama-1.1B-Chat-v0.3-GPTQ... 2025-12-04T09:36:09.2470777Z (EngineCore_DP0 pid=2744) INFO 12-04 09:36:09 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:36:09.2472035Z (EngineCore_DP0 pid=2744) INFO 12-04 09:36:09 [transformers.py:493] Using Transformers backend. 2025-12-04T09:36:09.2474192Z (EngineCore_DP0 pid=2744) `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:36:09.3781347Z (EngineCore_DP0 pid=2744) INFO 12-04 09:36:09 [gptq_marlin.py:357] Using MarlinLinearKernel for GPTQMarlinLinearMethod 2025-12-04T09:36:09.4454372Z (EngineCore_DP0 pid=2744) INFO 12-04 09:36:09 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:36:09.5188758Z (EngineCore_DP0 pid=2744) INFO 12-04 09:36:09 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-12-04T09:36:09.5824193Z (EngineCore_DP0 pid=2744) INFO 12-04 09:36:09 [weight_utils.py:480] No model.safetensors.index.json found in remote. 2025-12-04T09:36:09.5830289Z (EngineCore_DP0 pid=2744) 2025-12-04T09:36:09.7578739Z Loading safetensors checkpoint shards: 0% 0/1 [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-12-04T09:36:23.5757974Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:23.5788189Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:23.5794860Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:23.5802303Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:23.5809619Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:23.5816407Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:23.5818835Z (EngineCore_DP0 pid=2878) INFO 12-04 09:36:23 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:36:23.8416742Z (EngineCore_DP0 pid=2878) INFO 12-04 09:36:23 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:36:23.8867280Z (EngineCore_DP0 pid=2878) INFO 12-04 09:36:23 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-12-04T09:36:24.1318153Z (EngineCore_DP0 pid=2878) INFO 12-04 09:36:24 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:36:24.3497590Z (EngineCore_DP0 pid=2878) INFO 12-04 09:36:24 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:36:24.3728246Z (EngineCore_DP0 pid=2878) INFO 12-04 09:36:24 [bitsandbytes_loader.py:791] Loading weights with BitsAndBytes quantization. May take a while ... 2025-12-04T09:36:24.4620233Z (EngineCore_DP0 pid=2878) INFO 12-04 09:36:24 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-12-04T09:36:24.5212827Z (EngineCore_DP0 pid=2878) INFO 12-04 09:36:24 [weight_utils.py:480] No model.safetensors.index.json found in remote. 2025-12-04T09:36:24.5221902Z (EngineCore_DP0 pid=2878) 2025-12-04T09:36:24.8899920Z Loading safetensors checkpoint shards: 0% 0/1 [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-12-04T09:36:38.8436172Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:38.8465683Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:38.8472979Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:38.8479632Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:38.8486711Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:38.8493487Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:36:38.8496403Z (EngineCore_DP0 pid=3012) INFO 12-04 09:36:38 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:36:39.1035869Z (EngineCore_DP0 pid=3012) INFO 12-04 09:36:39 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:36:39.1761799Z (EngineCore_DP0 pid=3012) INFO 12-04 09:36:39 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-12-04T09:36:39.4216347Z (EngineCore_DP0 pid=3012) INFO 12-04 09:36:39 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:36:39.4217175Z (EngineCore_DP0 pid=3012) INFO 12-04 09:36:39 [transformers.py:493] Using Transformers backend. 2025-12-04T09:36:39.4219359Z (EngineCore_DP0 pid=3012) `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:36:39.6827349Z (EngineCore_DP0 pid=3012) INFO 12-04 09:36:39 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:36:39.6901213Z (EngineCore_DP0 pid=3012) INFO 12-04 09:36:39 [bitsandbytes_loader.py:791] Loading weights with BitsAndBytes quantization. May take a while ... 2025-12-04T09:36:39.7747767Z (EngineCore_DP0 pid=3012) INFO 12-04 09:36:39 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-12-04T09:36:39.8541749Z (EngineCore_DP0 pid=3012) INFO 12-04 09:36:39 [weight_utils.py:480] No model.safetensors.index.json found in remote. 2025-12-04T09:36:39.8551229Z (EngineCore_DP0 pid=3012) 2025-12-04T09:36:40.2304627Z Loading safetensors checkpoint shards: 0% 0/1 [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-12-04T09:37:01.6105551Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:37:01.6134846Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:37:01.6141782Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:37:01.6148756Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:37:01.6155740Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:37:01.6163121Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:37:01.6165568Z (EngineCore_DP0 pid=3200) INFO 12-04 09:37:01 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:37:01.8662828Z (EngineCore_DP0 pid=3200) INFO 12-04 09:37:01 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:37:01.9024957Z (EngineCore_DP0 pid=3200) INFO 12-04 09:37:01 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen3-Embedding-0.6B... 2025-12-04T09:37:02.1480794Z (EngineCore_DP0 pid=3200) INFO 12-04 09:37:02 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:37:02.1481632Z (EngineCore_DP0 pid=3200) INFO 12-04 09:37:02 [transformers.py:493] Using Transformers backend. 2025-12-04T09:37:02.1483737Z (EngineCore_DP0 pid=3200) `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:37:02.3558461Z (EngineCore_DP0 pid=3200) INFO 12-04 09:37:02 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:37:02.4853008Z (EngineCore_DP0 pid=3200) INFO 12-04 09:37:02 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-12-04T09:37:02.5579679Z (EngineCore_DP0 pid=3200) 2025-12-04T09:37:02.9137943Z model.safetensors: 0% 0.00/1.19G [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-12-04T09:37:15.6334565Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:37:15.6364213Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:37:15.6371375Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:37:15.6378703Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:37:15.6385779Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:37:15.6392950Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:37:15.6395481Z (EngineCore_DP0 pid=3364) INFO 12-04 09:37:15 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:37:15.8921426Z (EngineCore_DP0 pid=3364) INFO 12-04 09:37:15 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:37:15.9290694Z (EngineCore_DP0 pid=3364) INFO 12-04 09:37:15 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-12-04T09:37:16.1734894Z (EngineCore_DP0 pid=3364) INFO 12-04 09:37:16 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:37:16.1735994Z (EngineCore_DP0 pid=3364) INFO 12-04 09:37:16 [transformers.py:493] Using Transformers backend. 2025-12-04T09:37:16.1738290Z (EngineCore_DP0 pid=3364) `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:37:16.3576461Z (EngineCore_DP0 pid=3364) INFO 12-04 09:37:16 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:37:16.4415149Z (EngineCore_DP0 pid=3364) INFO 12-04 09:37:16 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-12-04T09:37:16.5010011Z (EngineCore_DP0 pid=3364) INFO 12-04 09:37:16 [weight_utils.py:480] No model.safetensors.index.json found in remote. 2025-12-04T09:37:16.5015578Z (EngineCore_DP0 pid=3364) 2025-12-04T09:37:16.8076916Z Loading safetensors checkpoint shards: 0% 0/1 [00:00:488 2025-12-04T09:37:20.9087238Z :488: DeprecationWarning: builtin type SwigPyPacked has no __module__ attribute 2025-12-04T09:37:20.9087984Z 2025-12-04T09:37:20.9088173Z :488 2025-12-04T09:37:20.9089108Z :488: DeprecationWarning: builtin type SwigPyObject has no __module__ attribute 2025-12-04T09:37:20.9089868Z 2025-12-04T09:37:20.9090213Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305 2025-12-04T09:37:20.9091928Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305: DeprecationWarning: jsonschema.exceptions.RefResolutionError is deprecated as of version 4.18.0. If you wish to catch potential reference resolution errors, directly catch referencing.exceptions.Unresolvable. 2025-12-04T09:37:20.9093404Z ref_error: type[Exception] = jsonschema.RefResolutionError, 2025-12-04T09:37:20.9093627Z 2025-12-04T09:37:20.9093876Z tests/models/test_transformers.py::test_models[meta-llama/Llama-3.2-1B-Instruct-transformers] 2025-12-04T09:37:20.9094687Z /opt/conda/envs/py_3.12/lib/python3.12/multiprocessing/popen_fork.py:66: DeprecationWarning: This process (pid=655) is multi-threaded, use of fork() may lead to deadlocks in the child. 2025-12-04T09:37:20.9095421Z self.pid = os.fork() 2025-12-04T09:37:20.9095552Z 2025-12-04T09:37:20.9095787Z tests/models/test_transformers.py::test_models[meta-llama/Llama-3.2-1B-Instruct-transformers] 2025-12-04T09:37:20.9096325Z /var/lib/jenkins/workspace/vllm/tests/models/test_transformers.py:54: UserWarning: Test0: 2025-12-04T09:37:20.9096748Z Matched tokens: [85, 4178, 44, 374, 6319, 311, 387, 7701] 2025-12-04T09:37:20.9097971Z ref: 'vLLM is designed to be highly parallelizable, allowing for efficient inference and serving of large-scale language models. It leverages a combination of techniques, including' {15638: -1.381659984588623, 69311: -1.381659984588623, 63174: -2.256659984588623, 34440: -2.756659984588623, 11297: -2.881659984588623} 2025-12-04T09:37:20.9100368Z test: 'vLLM is designed to be highly scalable, flexible, and adaptable to various use cases, including but not limited to, language translation, text summarization,' {69311: Logprob(logprob=-1.3481978178024292, rank=1, decoded_token=' scalable'), 15638: Logprob(logprob=-1.4731978178024292, rank=2, decoded_token=' parallel'), 63174: Logprob(logprob=-2.2231979370117188, rank=3, decoded_token=' customizable'), 34440: Logprob(logprob=-2.7231979370117188, rank=4, decoded_token=' optimized'), 11297: Logprob(logprob=-2.8481979370117188, rank=5, decoded_token=' efficient')} 2025-12-04T09:37:20.9101916Z check_logprobs_close( 2025-12-04T09:37:20.9102049Z 2025-12-04T09:37:20.9102290Z tests/models/test_transformers.py::test_models[meta-llama/Llama-3.2-1B-Instruct-transformers] 2025-12-04T09:37:20.9102844Z /var/lib/jenkins/workspace/vllm/tests/models/test_transformers.py:54: UserWarning: Test1: 2025-12-04T09:37:20.9103213Z Matched tokens: [] 2025-12-04T09:37:20.9103921Z ref: '**Answer:**\n1. **1950s:** The first AI programs were developed, including the Logical Theorist, which was a program that could reason' {334: -1.832015037536621, 8586: -1.832015037536621, 791: -2.082015037536621, 567: -2.082015037536621, 9470: -2.332015037536621} 2025-12-04T09:37:20.9105924Z test: 'Here are the major milestones in the development of artificial intelligence from 1950 to 2020:\n\n**1950s:**\n\n* Alan Turing publishes "Comput' {8586: Logprob(logprob=-1.7877479791641235, rank=1, decoded_token='Here'), 334: Logprob(logprob=-1.9127479791641235, rank=2, decoded_token='**'), 567: Logprob(logprob=-2.037747859954834, rank=3, decoded_token='##'), 791: Logprob(logprob=-2.162747859954834, rank=4, decoded_token='The'), 9470: Logprob(logprob=-2.287747859954834, rank=5, decoded_token='Art')} 2025-12-04T09:37:20.9107238Z check_logprobs_close( 2025-12-04T09:37:20.9107354Z 2025-12-04T09:37:20.9107595Z tests/models/test_transformers.py::test_models[meta-llama/Llama-3.2-1B-Instruct-transformers] 2025-12-04T09:37:20.9108128Z /var/lib/jenkins/workspace/vllm/tests/models/test_transformers.py:54: UserWarning: Test3: 2025-12-04T09:37:20.9108507Z Matched tokens: [32, 30828, 4009, 374, 264] 2025-12-04T09:37:20.9109349Z ref: 'A neural network is a complex system of interconnected nodes or "neurons" that process information. The basic components of a neural network include:\n1. **Art' {6485: -1.63894784450531, 955: -1.76394784450531, 6500: -2.0139479637145996, 55580: -2.3889479637145996, 1646: -2.7639479637145996} 2025-12-04T09:37:20.9111467Z test: 'A neural network is a type of machine learning model that is inspired by the structure and function of the human brain. It is composed of layers of interconnected nodes or' {955: Logprob(logprob=-1.7134931087493896, rank=2, decoded_token=' type'), 6485: Logprob(logprob=-1.7134931087493896, rank=1, decoded_token=' complex'), 6500: Logprob(logprob=-2.0884931087493896, rank=3, decoded_token=' computer'), 55580: Logprob(logprob=-2.3384931087493896, rank=4, decoded_token=' computational'), 1646: Logprob(logprob=-2.7134931087493896, rank=5, decoded_token=' model')} 2025-12-04T09:37:20.9112957Z check_logprobs_close( 2025-12-04T09:37:20.9113073Z 2025-12-04T09:37:20.9113298Z tests/models/test_transformers.py::test_models[meta-llama/Llama-3.2-1B-Instruct-transformers] 2025-12-04T09:37:20.9113813Z /var/lib/jenkins/workspace/vllm/tests/models/test_transformers.py:54: UserWarning: Test4: 2025-12-04T09:37:20.9114296Z Matched tokens: [2170, 279, 7160, 743, 927, 279, 3363, 11, 264, 47766, 12585, 7086, 1901, 1955, 14980] 2025-12-04T09:37:20.9115224Z ref: 'As the sun set over the city, a lone robot named Zeta stood on the rooftop of a high-rise building, gazing out at the twinkling lights' {389: -1.337048888206482, 304: -1.462048888206482, 520: -2.4620490074157715, 7636: -2.4620490074157715, 16615: -2.8370490074157715} 2025-12-04T09:37:20.9117369Z test: 'As the sun set over the city, a lone robot named Zeta stood in the empty lot where he had spent countless hours before. He had been built to' {304: Logprob(logprob=-1.4122865200042725, rank=2, decoded_token=' in'), 389: Logprob(logprob=-1.4122865200042725, rank=1, decoded_token=' on'), 520: Logprob(logprob=-2.4122865200042725, rank=3, decoded_token=' at'), 7636: Logprob(logprob=-2.5372865200042725, rank=4, decoded_token=' alone'), 16615: Logprob(logprob=-2.7872865200042725, rank=5, decoded_token=' tall')} 2025-12-04T09:37:20.9118693Z check_logprobs_close( 2025-12-04T09:37:20.9118809Z 2025-12-04T09:37:20.9119037Z tests/models/test_transformers.py::test_models[meta-llama/Llama-3.2-1B-Instruct-transformers] 2025-12-04T09:37:20.9119564Z /var/lib/jenkins/workspace/vllm/tests/models/test_transformers.py:54: UserWarning: Test5: 2025-12-04T09:37:20.9120312Z Matched tokens: [791, 20562, 12, 777, 28522, 706, 1047, 264, 28254, 5536, 389, 3728, 7100, 14726, 323, 3938, 2626, 4211, 13, 578, 28522, 706, 69627, 8312, 27271, 11, 9057, 24716] 2025-12-04T09:37:20.9121488Z ref: 'The COVID-19 pandemic has had a profound impact on global economic structures and future business models. The pandemic has disrupted supply chains, caused widespread lockdowns, and' {51235: -0.9368098974227905, 2683: -1.0618098974227905, 7100: -2.18681001663208, 61000: -3.06181001663208, 2626: -3.56181001663208} 2025-12-04T09:37:20.9124008Z test: 'The COVID-19 pandemic has had a profound impact on global economic structures and future business models. The pandemic has disrupted supply chains, caused widespread job losses, and' {2683: Logprob(logprob=-0.9869983196258545, rank=1, decoded_token=' job'), 51235: Logprob(logprob=-0.9869983196258545, rank=2, decoded_token=' lockdown'), 7100: Logprob(logprob=-2.2369983196258545, rank=3, decoded_token=' economic'), 61000: Logprob(logprob=-3.1119983196258545, rank=4, decoded_token=' closures'), 2626: Logprob(logprob=-3.4869983196258545, rank=5, decoded_token=' business')} 2025-12-04T09:37:20.9125716Z check_logprobs_close( 2025-12-04T09:37:20.9125838Z 2025-12-04T09:37:20.9126018Z tests/models/test_transformers.py::test_models[hmellor/Ilama-3.2-1B-auto] 2025-12-04T09:37:20.9126635Z /var/lib/jenkins/workspace/vllm/tests/models/test_transformers.py:54: UserWarning: Test2: 2025-12-04T09:37:20.9127140Z Matched tokens: [28474, 323, 13168, 21075, 11478, 449, 3823, 11478, 304, 3878, 315, 8863, 2038, 13] 2025-12-04T09:37:20.9128278Z ref: 'Compare and contrast artificial intelligence with human intelligence in terms of processing information. Artificial intelligence is a branch of computer science that deals with the design of intelligent machines. It' {59294: -2.893242359161377, 83017: -2.893242359161377, 763: -3.018242359161377, 66379: -3.143242359161377, 2650: -3.143242359161377} 2025-12-04T09:37:20.9130664Z test: 'Compare and contrast artificial intelligence with human intelligence in terms of processing information. Explain how the two differ.\nArtificial intelligence (AI) is a branch of computer science' {83017: Logprob(logprob=-2.7976644039154053, rank=1, decoded_token=' Explain'), 59294: Logprob(logprob=-2.9226644039154053, rank=2, decoded_token=' Artificial'), 66379: Logprob(logprob=-3.0476644039154053, rank=3, decoded_token=' Discuss'), 763: Logprob(logprob=-3.0476644039154053, rank=4, decoded_token=' In'), 2650: Logprob(logprob=-3.1726644039154053, rank=5, decoded_token=' How')} 2025-12-04T09:37:20.9132237Z check_logprobs_close( 2025-12-04T09:37:20.9132358Z 2025-12-04T09:37:20.9132671Z tests/models/test_transformers.py::test_quantization[5-32-TheBloke/TinyLlama-1.1B-Chat-v0.3-AWQ-quantization_kwargs0] 2025-12-04T09:37:20.9133291Z /var/lib/jenkins/workspace/vllm/tests/models/test_transformers.py:179: UserWarning: Test6: 2025-12-04T09:37:20.9133911Z Matched tokens: [5618, 338, 278, 16375, 26002, 310, 278, 2598, 29874, 29420, 20413, 29973, 13, 1576, 2598, 29874, 29420] 2025-12-04T09:37:20.9135478Z transformers: 'What is the cultural significance of the Mona Lisa painting?\nThe Mona Lisa painting is a famous painting by Leonardo da Vinci, which was created' {20413: Logprob(logprob=-0.8314066529273987, rank=1, decoded_token='painting'), 338: Logprob(logprob=-0.8392191529273987, rank=2, decoded_token='is'), 29892: Logprob(logprob=-3.057969093322754, rank=3, decoded_token=','), 756: Logprob(logprob=-3.761094093322754, rank=4, decoded_token='has'), 313: Logprob(logprob=-4.229844093322754, rank=5, decoded_token='(')} 2025-12-04T09:37:20.9138000Z vllm: 'What is the cultural significance of the Mona Lisa painting?\nThe Mona Lisa is a famous painting by Leonardo da Vinci, which was created for' {338: Logprob(logprob=-0.8350272178649902, rank=1, decoded_token='is'), 20413: Logprob(logprob=-0.8350272178649902, rank=2, decoded_token='painting'), 29892: Logprob(logprob=-3.0615897178649902, rank=3, decoded_token=','), 756: Logprob(logprob=-3.7647147178649902, rank=4, decoded_token='has'), 313: Logprob(logprob=-4.23346471786499, rank=5, decoded_token='(')} 2025-12-04T09:37:20.9139436Z check_logprobs_close( 2025-12-04T09:37:20.9139560Z 2025-12-04T09:37:20.9139747Z -- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html 2025-12-04T09:37:20.9140393Z ============ 8 passed, 4 skipped, 25 warnings in 346.78s (0:05:46) ============= 2025-12-04T09:37:21.2136317Z sys:1: DeprecationWarning: builtin type swigvarlink has no __module__ attribute 2025-12-04T09:37:22.6164397Z 2025-12-04 09:37:22,615 [INFO] cli.lib.core.vllm.lib: Finish running step: pytest -v -s models/test_transformers.py 2025-12-04T09:37:22.6165615Z 2025-12-04 09:37:22,616 [INFO] cli.lib.core.vllm.lib: Running step: pytest -v -s models/test_registry.py 2025-12-04T09:37:22.6166645Z 2025-12-04 09:37:22,616 [INFO] cli.lib.common.utils: [shell] pytest -v -s models/test_registry.py 2025-12-04T09:37:27.2736255Z INFO 12-04 09:37:27 [__init__.py:224] Automatically detected platform cuda. 2025-12-04T09:37:29.1367487Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_asyncio/plugin.py:208: PytestDeprecationWarning: The configuration option "asyncio_default_fixture_loop_scope" is unset. 2025-12-04T09:37:29.1369523Z The event loop scope for asynchronous fixtures will default to the fixture caching scope. Future versions of pytest-asyncio will default the loop scope for asynchronous fixtures to function scope. Set the default fixture loop scope explicitly in order to avoid unexpected behavior in the future. Valid fixture loop scopes are: "function", "class", "module", "package", "session" 2025-12-04T09:37:29.1370662Z 2025-12-04T09:37:29.1370864Z warnings.warn(PytestDeprecationWarning(_DEFAULT_FIXTURE_LOOP_SCOPE_UNSET)) 2025-12-04T09:37:29.1435353Z ============================= test session starts ============================== 2025-12-04T09:37:29.1435937Z platform linux -- Python 3.12.5, pytest-8.3.5, pluggy-1.5.0 -- /opt/conda/envs/py_3.12/bin/python 2025-12-04T09:37:29.1530230Z cachedir: .pytest_cache 2025-12-04T09:37:29.1530840Z hypothesis profile 'ci' -> database=None, deadline=None, print_blob=True, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:37:29.1531463Z rootdir: /var/lib/jenkins/workspace/vllm 2025-12-04T09:37:29.1531761Z configfile: pyproject.toml 2025-12-04T09:37:29.1532798Z plugins: cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, xdist-3.3.1, xdoctest-1.3.0, hydra-core-1.3.2, anyio-4.6.2.post1, buildkite-test-collector-0.1.9, forked-1.6.0, shard-0.1.2, mock-3.14.0, schemathesis-3.39.15, subtests-0.14.1, hypothesis-6.131.0, cov-6.3.0, asyncio-0.24.0, timeout-2.3.1 2025-12-04T09:37:29.1533916Z asyncio: mode=Mode.STRICT, default_loop_scope=None 2025-12-04T09:37:30.7006126Z collecting ... WARNING 12-04 09:37:30 [interface.py:514] Current platform cuda does not have '_pytestfixturefunction' attribute. 2025-12-04T09:37:31.0020373Z WARNING 12-04 09:37:31 [interface.py:514] Current platform cuda does not have '__test__' attribute. 2025-12-04T09:37:31.0021089Z WARNING 12-04 09:37:31 [interface.py:514] Current platform cuda does not have '__bases__' attribute. 2025-12-04T09:37:31.0021723Z WARNING 12-04 09:37:31 [interface.py:514] Current platform cuda does not have '__test__' attribute. 2025-12-04T09:37:31.0022370Z WARNING 12-04 09:37:31 [interface.py:514] Current platform cuda does not have '_schemathesis_test' attribute. 2025-12-04T09:37:31.0087627Z  2025-12-04T09:37:31.0112386Z collecting 229 items  2025-12-04T09:37:31.0113051Z collected 229 items  2025-12-04T09:37:31.0167673Z Running 229 items in this shard: tests/models/test_registry.py::test_registry_imports[ApertusForCausalLM], tests/models/test_registry.py::test_registry_imports[AquilaModel], tests/models/test_registry.py::test_registry_imports[AquilaForCausalLM], tests/models/test_registry.py::test_registry_imports[ArceeForCausalLM], tests/models/test_registry.py::test_registry_imports[ArcticForCausalLM], tests/models/test_registry.py::test_registry_imports[MiniMaxForCausalLM], tests/models/test_registry.py::test_registry_imports[MiniMaxText01ForCausalLM], tests/models/test_registry.py::test_registry_imports[MiniMaxM1ForCausalLM], tests/models/test_registry.py::test_registry_imports[BaiChuanForCausalLM], tests/models/test_registry.py::test_registry_imports[BaichuanForCausalLM], tests/models/test_registry.py::test_registry_imports[BailingMoeForCausalLM], tests/models/test_registry.py::test_registry_imports[BailingMoeV2ForCausalLM], tests/models/test_registry.py::test_registry_imports[BambaForCausalLM], tests/models/test_registry.py::test_registry_imports[BloomForCausalLM], tests/models/test_registry.py::test_registry_imports[ChatGLMModel], tests/models/test_registry.py::test_registry_imports[ChatGLMForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[CohereForCausalLM], tests/models/test_registry.py::test_registry_imports[Cohere2ForCausalLM], tests/models/test_registry.py::test_registry_imports[CwmForCausalLM], tests/models/test_registry.py::test_registry_imports[DbrxForCausalLM], tests/models/test_registry.py::test_registry_imports[DeciLMForCausalLM], tests/models/test_registry.py::test_registry_imports[DeepseekForCausalLM], tests/models/test_registry.py::test_registry_imports[DeepseekV2ForCausalLM], tests/models/test_registry.py::test_registry_imports[DeepseekV3ForCausalLM], tests/models/test_registry.py::test_registry_imports[DeepseekV32ForCausalLM], tests/models/test_registry.py::test_registry_imports[Dots1ForCausalLM], tests/models/test_registry.py::test_registry_imports[Ernie4_5ForCausalLM], tests/models/test_registry.py::test_registry_imports[Ernie4_5_MoeForCausalLM], tests/models/test_registry.py::test_registry_imports[ExaoneForCausalLM], tests/models/test_registry.py::test_registry_imports[Exaone4ForCausalLM], tests/models/test_registry.py::test_registry_imports[FalconForCausalLM], tests/models/test_registry.py::test_registry_imports[Fairseq2LlamaForCausalLM], tests/models/test_registry.py::test_registry_imports[GemmaForCausalLM], tests/models/test_registry.py::test_registry_imports[Gemma2ForCausalLM], tests/models/test_registry.py::test_registry_imports[Gemma3ForCausalLM], tests/models/test_registry.py::test_registry_imports[Gemma3nForCausalLM], tests/models/test_registry.py::test_registry_imports[Qwen3NextForCausalLM], tests/models/test_registry.py::test_registry_imports[GlmForCausalLM], tests/models/test_registry.py::test_registry_imports[Glm4ForCausalLM], tests/models/test_registry.py::test_registry_imports[Glm4MoeForCausalLM], tests/models/test_registry.py::test_registry_imports[GptOssForCausalLM], tests/models/test_registry.py::test_registry_imports[GPT2LMHeadModel], tests/models/test_registry.py::test_registry_imports[GPTBigCodeForCausalLM], tests/models/test_registry.py::test_registry_imports[GPTJForCausalLM], tests/models/test_registry.py::test_registry_imports[GPTNeoXForCausalLM], tests/models/test_registry.py::test_registry_imports[GraniteForCausalLM], tests/models/test_registry.py::test_registry_imports[GraniteMoeForCausalLM], tests/models/test_registry.py::test_registry_imports[GraniteMoeHybridForCausalLM], tests/models/test_registry.py::test_registry_imports[GraniteMoeSharedForCausalLM], tests/models/test_registry.py::test_registry_imports[GritLM], tests/models/test_registry.py::test_registry_imports[Grok1ModelForCausalLM], tests/models/test_registry.py::test_registry_imports[HunYuanMoEV1ForCausalLM], tests/models/test_registry.py::test_registry_imports[HunYuanDenseV1ForCausalLM], tests/models/test_registry.py::test_registry_imports[HCXVisionForCausalLM], tests/models/test_registry.py::test_registry_imports[InternLMForCausalLM], tests/models/test_registry.py::test_registry_imports[InternLM2ForCausalLM], tests/models/test_registry.py::test_registry_imports[InternLM2VEForCausalLM], tests/models/test_registry.py::test_registry_imports[InternLM3ForCausalLM], tests/models/test_registry.py::test_registry_imports[JAISLMHeadModel], tests/models/test_registry.py::test_registry_imports[JambaForCausalLM], tests/models/test_registry.py::test_registry_imports[Lfm2ForCausalLM], tests/models/test_registry.py::test_registry_imports[Lfm2MoeForCausalLM], tests/models/test_registry.py::test_registry_imports[LlamaForCausalLM], tests/models/test_registry.py::test_registry_imports[Llama4ForCausalLM], tests/models/test_registry.py::test_registry_imports[LLaMAForCausalLM], tests/models/test_registry.py::test_registry_imports[LongcatFlashForCausalLM], tests/models/test_registry.py::test_registry_imports[MambaForCausalLM], tests/models/test_registry.py::test_registry_imports[FalconMambaForCausalLM], tests/models/test_registry.py::test_registry_imports[FalconH1ForCausalLM], tests/models/test_registry.py::test_registry_imports[Mamba2ForCausalLM], tests/models/test_registry.py::test_registry_imports[MiniCPMForCausalLM], tests/models/test_registry.py::test_registry_imports[MiniCPM3ForCausalLM], tests/models/test_registry.py::test_registry_imports[MistralForCausalLM], tests/models/test_registry.py::test_registry_imports[MixtralForCausalLM], tests/models/test_registry.py::test_registry_imports[MptForCausalLM], tests/models/test_registry.py::test_registry_imports[MPTForCausalLM], tests/models/test_registry.py::test_registry_imports[MiMoForCausalLM], tests/models/test_registry.py::test_registry_imports[NemotronForCausalLM], tests/models/test_registry.py::test_registry_imports[NemotronHForCausalLM], tests/models/test_registry.py::test_registry_imports[OlmoForCausalLM], tests/models/test_registry.py::test_registry_imports[Olmo2ForCausalLM], tests/models/test_registry.py::test_registry_imports[Olmo3ForCausalLM], tests/models/test_registry.py::test_registry_imports[OlmoeForCausalLM], tests/models/test_registry.py::test_registry_imports[OPTForCausalLM], tests/models/test_registry.py::test_registry_imports[OrionForCausalLM], tests/models/test_registry.py::test_registry_imports[PersimmonForCausalLM], tests/models/test_registry.py::test_registry_imports[PhiForCausalLM], tests/models/test_registry.py::test_registry_imports[Phi3ForCausalLM], tests/models/test_registry.py::test_registry_imports[PhiMoEForCausalLM], tests/models/test_registry.py::test_registry_imports[Plamo2ForCausalLM], tests/models/test_registry.py::test_registry_imports[QWenLMHeadModel], tests/models/test_registry.py::test_registry_imports[Qwen2ForCausalLM], tests/models/test_registry.py::test_registry_imports[Qwen2MoeForCausalLM], tests/models/test_registry.py::test_registry_imports[Qwen3ForCausalLM], tests/models/test_registry.py::test_registry_imports[Qwen3MoeForCausalLM], tests/models/test_registry.py::test_registry_imports[RWForCausalLM], tests/models/test_registry.py::test_registry_imports[SeedOssForCausalLM], tests/models/test_registry.py::test_registry_imports[Step3TextForCausalLM], tests/models/test_registry.py::test_registry_imports[StableLMEpochForCausalLM], tests/models/test_registry.py::test_registry_imports[StableLmForCausalLM], tests/models/test_registry.py::test_registry_imports[Starcoder2ForCausalLM], tests/models/test_registry.py::test_registry_imports[SolarForCausalLM], tests/models/test_registry.py::test_registry_imports[TeleChat2ForCausalLM], tests/models/test_registry.py::test_registry_imports[TeleFLMForCausalLM], tests/models/test_registry.py::test_registry_imports[XverseForCausalLM], tests/models/test_registry.py::test_registry_imports[Zamba2ForCausalLM], tests/models/test_registry.py::test_registry_imports[BertModel], tests/models/test_registry.py::test_registry_imports[Gemma2Model], tests/models/test_registry.py::test_registry_imports[Gemma3TextModel], tests/models/test_registry.py::test_registry_imports[GPT2ForSequenceClassification], tests/models/test_registry.py::test_registry_imports[GteModel], tests/models/test_registry.py::test_registry_imports[GteNewModel], tests/models/test_registry.py::test_registry_imports[InternLM2ForRewardModel], tests/models/test_registry.py::test_registry_imports[JambaForSequenceClassification], tests/models/test_registry.py::test_registry_imports[LlamaModel], tests/models/test_registry.py::test_registry_imports[MistralModel], tests/models/test_registry.py::test_registry_imports[ModernBertModel], tests/models/test_registry.py::test_registry_imports[NomicBertModel], tests/models/test_registry.py::test_registry_imports[Qwen2Model], tests/models/test_registry.py::test_registry_imports[Qwen2ForRewardModel], tests/models/test_registry.py::test_registry_imports[Qwen2ForProcessRewardModel], tests/models/test_registry.py::test_registry_imports[RobertaForMaskedLM], tests/models/test_registry.py::test_registry_imports[RobertaModel], tests/models/test_registry.py::test_registry_imports[XLMRobertaModel], tests/models/test_registry.py::test_registry_imports[CLIPModel], tests/models/test_registry.py::test_registry_imports[LlavaNextForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Phi3VForCausalLM], tests/models/test_registry.py::test_registry_imports[Qwen2VLForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[PrithviGeoSpatialMAE], tests/models/test_registry.py::test_registry_imports[Terratorch], tests/models/test_registry.py::test_registry_imports[BertForSequenceClassification], tests/models/test_registry.py::test_registry_imports[BertForTokenClassification], tests/models/test_registry.py::test_registry_imports[GteNewForSequenceClassification], tests/models/test_registry.py::test_registry_imports[ModernBertForSequenceClassification], tests/models/test_registry.py::test_registry_imports[ModernBertForTokenClassification], tests/models/test_registry.py::test_registry_imports[RobertaForSequenceClassification], tests/models/test_registry.py::test_registry_imports[XLMRobertaForSequenceClassification], tests/models/test_registry.py::test_registry_imports[JinaVLForRanking], tests/models/test_registry.py::test_registry_imports[AriaForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[AyaVisionForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Blip2ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[ChameleonForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Cohere2VisionForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[DeepseekVLV2ForCausalLM], tests/models/test_registry.py::test_registry_imports[DotsOCRForCausalLM], tests/models/test_registry.py::test_registry_imports[Ernie4_5_VLMoeForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[FuyuForCausalLM], tests/models/test_registry.py::test_registry_imports[Gemma3ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Gemma3nForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[GLM4VForCausalLM], tests/models/test_registry.py::test_registry_imports[Glm4vForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Glm4vMoeForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[GraniteSpeechForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[H2OVLChatModel], tests/models/test_registry.py::test_registry_imports[InternVLChatModel], tests/models/test_registry.py::test_registry_imports[NemotronH_Nano_VL_V2], tests/models/test_registry.py::test_registry_imports[InternS1ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[InternVLForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Idefics3ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[SmolVLMForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[KeyeForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[KeyeVL1_5ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[RForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[KimiVLForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Llama_Nemotron_Nano_VL], tests/models/test_registry.py::test_registry_imports[Llama4ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[LlavaForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[LlavaNextVideoForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[LlavaOnevisionForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[MantisForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[MiDashengLMModel], tests/models/test_registry.py::test_registry_imports[MiniMaxVL01ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[MiniCPMO], tests/models/test_registry.py::test_registry_imports[MiniCPMV], tests/models/test_registry.py::test_registry_imports[Mistral3ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[MolmoForCausalLM], tests/models/test_registry.py::test_registry_imports[NVLM_D], tests/models/test_registry.py::test_registry_imports[Ovis], tests/models/test_registry.py::test_registry_imports[Ovis2_5], tests/models/test_registry.py::test_registry_imports[PaliGemmaForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Phi4MMForCausalLM], tests/models/test_registry.py::test_registry_imports[Phi4MultimodalForCausalLM], tests/models/test_registry.py::test_registry_imports[PixtralForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[QwenVLForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Qwen2_5_VLForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Qwen2AudioForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Qwen2_5OmniModel], tests/models/test_registry.py::test_registry_imports[Qwen2_5OmniForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Qwen3VLForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Qwen3VLMoeForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[SkyworkR1VChatModel], tests/models/test_registry.py::test_registry_imports[Step3VLForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[TarsierForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Tarsier2ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[UltravoxModel], tests/models/test_registry.py::test_registry_imports[VoxtralForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[WhisperForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[MiMoMTPModel], tests/models/test_registry.py::test_registry_imports[EagleLlamaForCausalLM], tests/models/test_registry.py::test_registry_imports[EagleLlama4ForCausalLM], tests/models/test_registry.py::test_registry_imports[EagleMiniCPMForCausalLM], tests/models/test_registry.py::test_registry_imports[Eagle3LlamaForCausalLM], tests/models/test_registry.py::test_registry_imports[LlamaForCausalLMEagle3], tests/models/test_registry.py::test_registry_imports[Eagle3Qwen2_5vlForCausalLM], tests/models/test_registry.py::test_registry_imports[EagleDeepSeekMTPModel], tests/models/test_registry.py::test_registry_imports[DeepSeekMTPModel], tests/models/test_registry.py::test_registry_imports[ErnieMTPModel], tests/models/test_registry.py::test_registry_imports[LongCatFlashMTPModel], tests/models/test_registry.py::test_registry_imports[Glm4MoeMTPModel], tests/models/test_registry.py::test_registry_imports[MedusaModel], tests/models/test_registry.py::test_registry_imports[Qwen3NextMTP], tests/models/test_registry.py::test_registry_imports[SmolLM3ForCausalLM], tests/models/test_registry.py::test_registry_imports[Emu3ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[TransformersForCausalLM], tests/models/test_registry.py::test_registry_imports[TransformersForMultimodalLM], tests/models/test_registry.py::test_registry_imports[TransformersMoEForCausalLM], tests/models/test_registry.py::test_registry_imports[TransformersMoEForMultimodalLM], tests/models/test_registry.py::test_registry_imports[TransformersEmbeddingModel], tests/models/test_registry.py::test_registry_imports[TransformersForSequenceClassification], tests/models/test_registry.py::test_registry_imports[TransformersMoEForSequenceClassification], tests/models/test_registry.py::test_registry_imports[TransformersMoEEmbeddingModel], tests/models/test_registry.py::test_registry_model_property[LlamaForCausalLM-False-False-False], tests/models/test_registry.py::test_registry_model_property[LlavaForConditionalGeneration-True-True-False], tests/models/test_registry.py::test_registry_model_property[BertForSequenceClassification-False-False-True], tests/models/test_registry.py::test_registry_model_property[RobertaForSequenceClassification-False-False-True], tests/models/test_registry.py::test_registry_model_property[XLMRobertaForSequenceClassification-False-False-True], tests/models/test_registry.py::test_registry_is_pp[DeepseekV2ForCausalLM-True-False], tests/models/test_registry.py::test_registry_is_pp[Qwen2VLForConditionalGeneration-True-True], tests/models/test_registry.py::test_hf_registry_coverage 2025-12-04T09:37:31.0219429Z 2025-12-04T09:37:31.2973726Z models/test_registry.py::test_registry_imports[ApertusForCausalLM] PASSED 2025-12-04T09:37:31.5211633Z models/test_registry.py::test_registry_imports[AquilaModel] PASSED 2025-12-04T09:37:31.7409386Z models/test_registry.py::test_registry_imports[AquilaForCausalLM] PASSED 2025-12-04T09:37:31.9736036Z models/test_registry.py::test_registry_imports[ArceeForCausalLM] PASSED 2025-12-04T09:37:32.3467893Z models/test_registry.py::test_registry_imports[ArcticForCausalLM] PASSED 2025-12-04T09:37:32.6442743Z models/test_registry.py::test_registry_imports[MiniMaxForCausalLM] PASSED 2025-12-04T09:37:32.8644652Z models/test_registry.py::test_registry_imports[MiniMaxText01ForCausalLM] PASSED 2025-12-04T09:37:33.0877261Z models/test_registry.py::test_registry_imports[MiniMaxM1ForCausalLM] PASSED 2025-12-04T09:37:33.3233912Z models/test_registry.py::test_registry_imports[BaiChuanForCausalLM] PASSED 2025-12-04T09:37:33.5452169Z models/test_registry.py::test_registry_imports[BaichuanForCausalLM] PASSED 2025-12-04T09:37:33.7796370Z models/test_registry.py::test_registry_imports[BailingMoeForCausalLM] PASSED 2025-12-04T09:37:34.0012051Z models/test_registry.py::test_registry_imports[BailingMoeV2ForCausalLM] PASSED 2025-12-04T09:37:34.3786312Z models/test_registry.py::test_registry_imports[BambaForCausalLM] PASSED 2025-12-04T09:37:34.6329013Z models/test_registry.py::test_registry_imports[BloomForCausalLM] PASSED 2025-12-04T09:37:34.8545357Z models/test_registry.py::test_registry_imports[ChatGLMModel] SKIPPED 2025-12-04T09:37:35.0934006Z models/test_registry.py::test_registry_imports[ChatGLMForConditionalGeneration] PASSED 2025-12-04T09:37:35.3415158Z models/test_registry.py::test_registry_imports[CohereForCausalLM] PASSED 2025-12-04T09:37:35.5638975Z models/test_registry.py::test_registry_imports[Cohere2ForCausalLM] PASSED 2025-12-04T09:37:35.7856562Z models/test_registry.py::test_registry_imports[CwmForCausalLM] PASSED 2025-12-04T09:37:36.0369798Z models/test_registry.py::test_registry_imports[DbrxForCausalLM] PASSED 2025-12-04T09:37:36.2713522Z models/test_registry.py::test_registry_imports[DeciLMForCausalLM] PASSED 2025-12-04T09:37:36.4997361Z models/test_registry.py::test_registry_imports[DeepseekForCausalLM] PASSED 2025-12-04T09:37:36.7323071Z models/test_registry.py::test_registry_imports[DeepseekV2ForCausalLM] PASSED 2025-12-04T09:37:36.9555402Z models/test_registry.py::test_registry_imports[DeepseekV3ForCausalLM] PASSED 2025-12-04T09:37:37.1776036Z models/test_registry.py::test_registry_imports[DeepseekV32ForCausalLM] PASSED 2025-12-04T09:37:37.4058716Z models/test_registry.py::test_registry_imports[Dots1ForCausalLM] PASSED 2025-12-04T09:37:37.6298795Z models/test_registry.py::test_registry_imports[Ernie4_5ForCausalLM] PASSED 2025-12-04T09:37:37.8548972Z models/test_registry.py::test_registry_imports[Ernie4_5_MoeForCausalLM] PASSED 2025-12-04T09:37:38.0805329Z models/test_registry.py::test_registry_imports[ExaoneForCausalLM] PASSED 2025-12-04T09:37:38.3091633Z models/test_registry.py::test_registry_imports[Exaone4ForCausalLM] PASSED 2025-12-04T09:37:38.5361954Z models/test_registry.py::test_registry_imports[FalconForCausalLM] PASSED 2025-12-04T09:37:38.7593370Z models/test_registry.py::test_registry_imports[Fairseq2LlamaForCausalLM] PASSED 2025-12-04T09:37:38.9872810Z models/test_registry.py::test_registry_imports[GemmaForCausalLM] PASSED 2025-12-04T09:37:39.2132251Z models/test_registry.py::test_registry_imports[Gemma2ForCausalLM] PASSED 2025-12-04T09:37:39.4443509Z models/test_registry.py::test_registry_imports[Gemma3ForCausalLM] PASSED 2025-12-04T09:37:40.4439548Z models/test_registry.py::test_registry_imports[Gemma3nForCausalLM] PASSED 2025-12-04T09:37:40.6793147Z models/test_registry.py::test_registry_imports[Qwen3NextForCausalLM] SKIPPED 2025-12-04T09:37:40.9180895Z models/test_registry.py::test_registry_imports[GlmForCausalLM] PASSED 2025-12-04T09:37:41.1574063Z models/test_registry.py::test_registry_imports[Glm4ForCausalLM] PASSED 2025-12-04T09:37:41.3961546Z models/test_registry.py::test_registry_imports[Glm4MoeForCausalLM] PASSED 2025-12-04T09:37:41.6361152Z models/test_registry.py::test_registry_imports[GptOssForCausalLM] PASSED 2025-12-04T09:37:41.8781749Z models/test_registry.py::test_registry_imports[GPT2LMHeadModel] PASSED 2025-12-04T09:37:42.1174546Z models/test_registry.py::test_registry_imports[GPTBigCodeForCausalLM] PASSED 2025-12-04T09:37:42.3572630Z models/test_registry.py::test_registry_imports[GPTJForCausalLM] PASSED 2025-12-04T09:37:42.5970323Z models/test_registry.py::test_registry_imports[GPTNeoXForCausalLM] PASSED 2025-12-04T09:37:42.8364061Z models/test_registry.py::test_registry_imports[GraniteForCausalLM] PASSED 2025-12-04T09:37:43.0764114Z models/test_registry.py::test_registry_imports[GraniteMoeForCausalLM] PASSED 2025-12-04T09:37:43.3241784Z models/test_registry.py::test_registry_imports[GraniteMoeHybridForCausalLM] PASSED 2025-12-04T09:37:43.5622690Z models/test_registry.py::test_registry_imports[GraniteMoeSharedForCausalLM] PASSED 2025-12-04T09:37:43.7993224Z models/test_registry.py::test_registry_imports[GritLM] PASSED 2025-12-04T09:37:44.0381684Z models/test_registry.py::test_registry_imports[Grok1ModelForCausalLM] PASSED 2025-12-04T09:37:44.2786088Z models/test_registry.py::test_registry_imports[HunYuanMoEV1ForCausalLM] PASSED 2025-12-04T09:37:44.5161836Z models/test_registry.py::test_registry_imports[HunYuanDenseV1ForCausalLM] PASSED 2025-12-04T09:37:44.7655727Z models/test_registry.py::test_registry_imports[HCXVisionForCausalLM] PASSED 2025-12-04T09:37:45.0023701Z models/test_registry.py::test_registry_imports[InternLMForCausalLM] PASSED 2025-12-04T09:37:45.2453652Z models/test_registry.py::test_registry_imports[InternLM2ForCausalLM] PASSED 2025-12-04T09:37:45.4846614Z models/test_registry.py::test_registry_imports[InternLM2VEForCausalLM] PASSED 2025-12-04T09:37:45.7211484Z models/test_registry.py::test_registry_imports[InternLM3ForCausalLM] PASSED 2025-12-04T09:37:45.9602969Z models/test_registry.py::test_registry_imports[JAISLMHeadModel] PASSED 2025-12-04T09:37:46.2078067Z models/test_registry.py::test_registry_imports[JambaForCausalLM] PASSED 2025-12-04T09:37:46.4549511Z models/test_registry.py::test_registry_imports[Lfm2ForCausalLM] PASSED 2025-12-04T09:37:46.6909825Z models/test_registry.py::test_registry_imports[Lfm2MoeForCausalLM] SKIPPED 2025-12-04T09:37:46.9319835Z models/test_registry.py::test_registry_imports[LlamaForCausalLM] PASSED 2025-12-04T09:37:47.1766289Z models/test_registry.py::test_registry_imports[Llama4ForCausalLM] PASSED 2025-12-04T09:37:47.4135593Z models/test_registry.py::test_registry_imports[LLaMAForCausalLM] PASSED 2025-12-04T09:37:47.6526389Z models/test_registry.py::test_registry_imports[LongcatFlashForCausalLM] PASSED 2025-12-04T09:37:47.8952433Z models/test_registry.py::test_registry_imports[MambaForCausalLM] PASSED 2025-12-04T09:37:48.1363328Z models/test_registry.py::test_registry_imports[FalconMambaForCausalLM] PASSED 2025-12-04T09:37:48.3801000Z models/test_registry.py::test_registry_imports[FalconH1ForCausalLM] PASSED 2025-12-04T09:37:48.6193293Z models/test_registry.py::test_registry_imports[Mamba2ForCausalLM] PASSED 2025-12-04T09:37:48.8574992Z models/test_registry.py::test_registry_imports[MiniCPMForCausalLM] PASSED 2025-12-04T09:37:49.0961084Z models/test_registry.py::test_registry_imports[MiniCPM3ForCausalLM] PASSED 2025-12-04T09:37:49.3334710Z models/test_registry.py::test_registry_imports[MistralForCausalLM] PASSED 2025-12-04T09:37:49.5755059Z models/test_registry.py::test_registry_imports[MixtralForCausalLM] PASSED 2025-12-04T09:37:49.8168889Z models/test_registry.py::test_registry_imports[MptForCausalLM] PASSED 2025-12-04T09:37:50.0545331Z models/test_registry.py::test_registry_imports[MPTForCausalLM] PASSED 2025-12-04T09:37:50.2943178Z models/test_registry.py::test_registry_imports[MiMoForCausalLM] PASSED 2025-12-04T09:37:50.5337062Z models/test_registry.py::test_registry_imports[NemotronForCausalLM] PASSED 2025-12-04T09:37:50.7727122Z models/test_registry.py::test_registry_imports[NemotronHForCausalLM] PASSED 2025-12-04T09:37:51.0162710Z models/test_registry.py::test_registry_imports[OlmoForCausalLM] PASSED 2025-12-04T09:37:51.2589231Z models/test_registry.py::test_registry_imports[Olmo2ForCausalLM] PASSED 2025-12-04T09:37:51.4969030Z models/test_registry.py::test_registry_imports[Olmo3ForCausalLM] PASSED 2025-12-04T09:37:51.7385321Z models/test_registry.py::test_registry_imports[OlmoeForCausalLM] PASSED 2025-12-04T09:37:51.9813676Z models/test_registry.py::test_registry_imports[OPTForCausalLM] PASSED 2025-12-04T09:37:52.2208708Z models/test_registry.py::test_registry_imports[OrionForCausalLM] PASSED 2025-12-04T09:37:52.4646519Z models/test_registry.py::test_registry_imports[PersimmonForCausalLM] PASSED 2025-12-04T09:37:52.7064466Z models/test_registry.py::test_registry_imports[PhiForCausalLM] PASSED 2025-12-04T09:37:52.9487000Z models/test_registry.py::test_registry_imports[Phi3ForCausalLM] PASSED 2025-12-04T09:37:53.1892177Z models/test_registry.py::test_registry_imports[PhiMoEForCausalLM] PASSED 2025-12-04T09:37:53.4275990Z models/test_registry.py::test_registry_imports[Plamo2ForCausalLM] SKIPPED 2025-12-04T09:37:53.6673856Z models/test_registry.py::test_registry_imports[QWenLMHeadModel] SKIPPED 2025-12-04T09:37:53.9061222Z models/test_registry.py::test_registry_imports[Qwen2ForCausalLM] PASSED 2025-12-04T09:37:54.1508153Z models/test_registry.py::test_registry_imports[Qwen2MoeForCausalLM] PASSED 2025-12-04T09:37:54.3915561Z models/test_registry.py::test_registry_imports[Qwen3ForCausalLM] PASSED 2025-12-04T09:37:54.6321774Z models/test_registry.py::test_registry_imports[Qwen3MoeForCausalLM] PASSED 2025-12-04T09:37:54.8745501Z models/test_registry.py::test_registry_imports[RWForCausalLM] PASSED 2025-12-04T09:37:55.1204681Z models/test_registry.py::test_registry_imports[SeedOssForCausalLM] PASSED 2025-12-04T09:37:55.3625704Z models/test_registry.py::test_registry_imports[Step3TextForCausalLM] PASSED 2025-12-04T09:37:55.6079727Z models/test_registry.py::test_registry_imports[StableLMEpochForCausalLM] PASSED 2025-12-04T09:37:55.8468747Z models/test_registry.py::test_registry_imports[StableLmForCausalLM] PASSED 2025-12-04T09:37:56.0919563Z models/test_registry.py::test_registry_imports[Starcoder2ForCausalLM] PASSED 2025-12-04T09:37:56.3365006Z models/test_registry.py::test_registry_imports[SolarForCausalLM] PASSED 2025-12-04T09:37:56.5775371Z models/test_registry.py::test_registry_imports[TeleChat2ForCausalLM] PASSED 2025-12-04T09:37:56.8205280Z models/test_registry.py::test_registry_imports[TeleFLMForCausalLM] PASSED 2025-12-04T09:37:57.0604929Z models/test_registry.py::test_registry_imports[XverseForCausalLM] PASSED 2025-12-04T09:37:57.3080128Z models/test_registry.py::test_registry_imports[Zamba2ForCausalLM] PASSED 2025-12-04T09:37:57.5544114Z models/test_registry.py::test_registry_imports[BertModel] PASSED 2025-12-04T09:37:57.7938360Z models/test_registry.py::test_registry_imports[Gemma2Model] PASSED 2025-12-04T09:37:58.0358701Z models/test_registry.py::test_registry_imports[Gemma3TextModel] PASSED 2025-12-04T09:37:58.0364677Z models/test_registry.py::test_registry_imports[GPT2ForSequenceClassification] WARNING 12-04 09:37:58 [interfaces_base.py:74] The model () is missing the `get_input_embeddings` method. 2025-12-04T09:37:58.0368995Z WARNING 12-04 09:37:58 [interfaces_base.py:74] The model () is missing the `get_input_embeddings` method. 2025-12-04T09:37:58.0370260Z WARNING 12-04 09:37:58 [interfaces_base.py:74] The model () is missing the `get_input_embeddings` method. 2025-12-04T09:37:58.2768899Z PASSED 2025-12-04T09:37:58.5194850Z models/test_registry.py::test_registry_imports[GteModel] PASSED 2025-12-04T09:37:58.7594647Z models/test_registry.py::test_registry_imports[GteNewModel] PASSED 2025-12-04T09:37:58.9995644Z models/test_registry.py::test_registry_imports[InternLM2ForRewardModel] PASSED 2025-12-04T09:37:59.2392387Z models/test_registry.py::test_registry_imports[JambaForSequenceClassification] PASSED 2025-12-04T09:37:59.4790175Z models/test_registry.py::test_registry_imports[LlamaModel] PASSED 2025-12-04T09:37:59.7180130Z models/test_registry.py::test_registry_imports[MistralModel] PASSED 2025-12-04T09:37:59.9647978Z models/test_registry.py::test_registry_imports[ModernBertModel] PASSED 2025-12-04T09:38:00.2034932Z models/test_registry.py::test_registry_imports[NomicBertModel] PASSED 2025-12-04T09:38:00.4433405Z models/test_registry.py::test_registry_imports[Qwen2Model] PASSED 2025-12-04T09:38:00.6830406Z models/test_registry.py::test_registry_imports[Qwen2ForRewardModel] SKIPPED 2025-12-04T09:38:00.9229115Z models/test_registry.py::test_registry_imports[Qwen2ForProcessRewardModel] SKIPPED 2025-12-04T09:38:01.1725545Z models/test_registry.py::test_registry_imports[RobertaForMaskedLM] PASSED 2025-12-04T09:38:01.4128834Z models/test_registry.py::test_registry_imports[RobertaModel] PASSED 2025-12-04T09:38:01.6519566Z models/test_registry.py::test_registry_imports[XLMRobertaModel] PASSED 2025-12-04T09:38:01.8910789Z models/test_registry.py::test_registry_imports[CLIPModel] PASSED 2025-12-04T09:38:02.7680785Z models/test_registry.py::test_registry_imports[LlavaNextForConditionalGeneration] PASSED 2025-12-04T09:38:03.0164050Z models/test_registry.py::test_registry_imports[Phi3VForCausalLM] SKIPPED 2025-12-04T09:38:03.2781531Z models/test_registry.py::test_registry_imports[Qwen2VLForConditionalGeneration] PASSED 2025-12-04T09:38:11.4940817Z models/test_registry.py::test_registry_imports[PrithviGeoSpatialMAE] PASSED 2025-12-04T09:38:11.8683076Z models/test_registry.py::test_registry_imports[Terratorch] PASSED 2025-12-04T09:38:12.2412660Z models/test_registry.py::test_registry_imports[BertForSequenceClassification] PASSED 2025-12-04T09:38:12.6085908Z models/test_registry.py::test_registry_imports[BertForTokenClassification] PASSED 2025-12-04T09:38:12.9742136Z models/test_registry.py::test_registry_imports[GteNewForSequenceClassification] PASSED 2025-12-04T09:38:13.3397475Z models/test_registry.py::test_registry_imports[ModernBertForSequenceClassification] PASSED 2025-12-04T09:38:13.7156087Z models/test_registry.py::test_registry_imports[ModernBertForTokenClassification] PASSED 2025-12-04T09:38:14.0970807Z models/test_registry.py::test_registry_imports[RobertaForSequenceClassification] PASSED 2025-12-04T09:38:14.4644598Z models/test_registry.py::test_registry_imports[XLMRobertaForSequenceClassification] PASSED 2025-12-04T09:38:14.8451808Z models/test_registry.py::test_registry_imports[JinaVLForRanking] PASSED 2025-12-04T09:38:15.2471667Z models/test_registry.py::test_registry_imports[AriaForConditionalGeneration] PASSED 2025-12-04T09:38:15.6404518Z models/test_registry.py::test_registry_imports[AyaVisionForConditionalGeneration] PASSED 2025-12-04T09:38:16.0361602Z models/test_registry.py::test_registry_imports[Blip2ForConditionalGeneration] PASSED 2025-12-04T09:38:16.4244908Z models/test_registry.py::test_registry_imports[ChameleonForConditionalGeneration] PASSED 2025-12-04T09:38:16.8196256Z models/test_registry.py::test_registry_imports[Cohere2VisionForConditionalGeneration] PASSED 2025-12-04T09:38:17.1959878Z models/test_registry.py::test_registry_imports[DeepseekVLV2ForCausalLM] SKIPPED 2025-12-04T09:38:17.5832929Z models/test_registry.py::test_registry_imports[DotsOCRForCausalLM] PASSED 2025-12-04T09:38:17.9635132Z models/test_registry.py::test_registry_imports[Ernie4_5_VLMoeForConditionalGeneration] PASSED 2025-12-04T09:38:18.3522332Z models/test_registry.py::test_registry_imports[FuyuForCausalLM] PASSED 2025-12-04T09:38:18.7401085Z models/test_registry.py::test_registry_imports[Gemma3ForConditionalGeneration] PASSED 2025-12-04T09:38:19.1683358Z models/test_registry.py::test_registry_imports[Gemma3nForConditionalGeneration] PASSED 2025-12-04T09:38:19.5535528Z models/test_registry.py::test_registry_imports[GLM4VForCausalLM] PASSED 2025-12-04T09:38:19.9502380Z models/test_registry.py::test_registry_imports[Glm4vForConditionalGeneration] PASSED 2025-12-04T09:38:20.3361045Z models/test_registry.py::test_registry_imports[Glm4vMoeForConditionalGeneration] PASSED 2025-12-04T09:38:20.7218598Z models/test_registry.py::test_registry_imports[GraniteSpeechForConditionalGeneration] PASSED 2025-12-04T09:38:21.0980229Z models/test_registry.py::test_registry_imports[H2OVLChatModel] SKIPPED 2025-12-04T09:38:21.4860453Z models/test_registry.py::test_registry_imports[InternVLChatModel] PASSED 2025-12-04T09:38:21.8750733Z models/test_registry.py::test_registry_imports[NemotronH_Nano_VL_V2] PASSED 2025-12-04T09:38:22.2705608Z models/test_registry.py::test_registry_imports[InternS1ForConditionalGeneration] PASSED 2025-12-04T09:38:22.6459573Z models/test_registry.py::test_registry_imports[InternVLForConditionalGeneration] PASSED 2025-12-04T09:38:23.0389612Z models/test_registry.py::test_registry_imports[Idefics3ForConditionalGeneration] PASSED 2025-12-04T09:38:23.5214731Z models/test_registry.py::test_registry_imports[SmolVLMForConditionalGeneration] PASSED 2025-12-04T09:38:23.9111382Z models/test_registry.py::test_registry_imports[KeyeForConditionalGeneration] PASSED 2025-12-04T09:38:24.2947305Z models/test_registry.py::test_registry_imports[KeyeVL1_5ForConditionalGeneration] PASSED 2025-12-04T09:38:24.7039545Z models/test_registry.py::test_registry_imports[RForConditionalGeneration] PASSED 2025-12-04T09:38:25.0980633Z models/test_registry.py::test_registry_imports[KimiVLForConditionalGeneration] PASSED 2025-12-04T09:38:25.4894472Z models/test_registry.py::test_registry_imports[Llama_Nemotron_Nano_VL] PASSED 2025-12-04T09:38:25.8859929Z models/test_registry.py::test_registry_imports[Llama4ForConditionalGeneration] PASSED 2025-12-04T09:38:26.2737671Z models/test_registry.py::test_registry_imports[LlavaForConditionalGeneration] PASSED 2025-12-04T09:38:26.6690033Z models/test_registry.py::test_registry_imports[LlavaNextVideoForConditionalGeneration] PASSED 2025-12-04T09:38:27.0549220Z models/test_registry.py::test_registry_imports[LlavaOnevisionForConditionalGeneration] PASSED 2025-12-04T09:38:27.4382009Z models/test_registry.py::test_registry_imports[MantisForConditionalGeneration] SKIPPED 2025-12-04T09:38:28.0322072Z models/test_registry.py::test_registry_imports[MiDashengLMModel] PASSED 2025-12-04T09:38:28.4184411Z models/test_registry.py::test_registry_imports[MiniMaxVL01ForConditionalGeneration] PASSED 2025-12-04T09:38:28.8173595Z models/test_registry.py::test_registry_imports[MiniCPMO] PASSED 2025-12-04T09:38:29.2069447Z models/test_registry.py::test_registry_imports[MiniCPMV] PASSED 2025-12-04T09:38:29.5964514Z models/test_registry.py::test_registry_imports[Mistral3ForConditionalGeneration] PASSED 2025-12-04T09:38:29.9880265Z models/test_registry.py::test_registry_imports[MolmoForCausalLM] SKIPPED 2025-12-04T09:38:30.3739842Z models/test_registry.py::test_registry_imports[NVLM_D] PASSED 2025-12-04T09:38:30.7633413Z models/test_registry.py::test_registry_imports[Ovis] SKIPPED (`trans...) 2025-12-04T09:38:31.1730065Z models/test_registry.py::test_registry_imports[Ovis2_5] PASSED 2025-12-04T09:38:31.5640068Z models/test_registry.py::test_registry_imports[PaliGemmaForConditionalGeneration] PASSED 2025-12-04T09:38:31.9543499Z models/test_registry.py::test_registry_imports[Phi4MMForCausalLM] PASSED 2025-12-04T09:38:32.3746024Z models/test_registry.py::test_registry_imports[Phi4MultimodalForCausalLM] PASSED 2025-12-04T09:38:32.7597743Z models/test_registry.py::test_registry_imports[PixtralForConditionalGeneration] PASSED 2025-12-04T09:38:33.1567477Z models/test_registry.py::test_registry_imports[QwenVLForConditionalGeneration] PASSED 2025-12-04T09:38:33.5501931Z models/test_registry.py::test_registry_imports[Qwen2_5_VLForConditionalGeneration] PASSED 2025-12-04T09:38:33.9527197Z models/test_registry.py::test_registry_imports[Qwen2AudioForConditionalGeneration] PASSED 2025-12-04T09:38:34.3873452Z models/test_registry.py::test_registry_imports[Qwen2_5OmniModel] PASSED 2025-12-04T09:38:34.7810636Z models/test_registry.py::test_registry_imports[Qwen2_5OmniForConditionalGeneration] PASSED 2025-12-04T09:38:35.1718025Z models/test_registry.py::test_registry_imports[Qwen3VLForConditionalGeneration] SKIPPED 2025-12-04T09:38:35.5653634Z models/test_registry.py::test_registry_imports[Qwen3VLMoeForConditionalGeneration] SKIPPED 2025-12-04T09:38:35.9628136Z models/test_registry.py::test_registry_imports[SkyworkR1VChatModel] PASSED 2025-12-04T09:38:36.3562699Z models/test_registry.py::test_registry_imports[Step3VLForConditionalGeneration] PASSED 2025-12-04T09:38:36.7439581Z models/test_registry.py::test_registry_imports[TarsierForConditionalGeneration] PASSED 2025-12-04T09:38:37.1343036Z models/test_registry.py::test_registry_imports[Tarsier2ForConditionalGeneration] PASSED 2025-12-04T09:38:37.5253925Z models/test_registry.py::test_registry_imports[UltravoxModel] PASSED 2025-12-04T09:38:37.9760004Z models/test_registry.py::test_registry_imports[VoxtralForConditionalGeneration] PASSED 2025-12-04T09:38:38.3728548Z models/test_registry.py::test_registry_imports[WhisperForConditionalGeneration] PASSED 2025-12-04T09:38:38.7706765Z models/test_registry.py::test_registry_imports[MiMoMTPModel] PASSED 2025-12-04T09:38:39.1702764Z models/test_registry.py::test_registry_imports[EagleLlamaForCausalLM] PASSED 2025-12-04T09:38:39.5624493Z models/test_registry.py::test_registry_imports[EagleLlama4ForCausalLM] PASSED 2025-12-04T09:38:39.9562891Z models/test_registry.py::test_registry_imports[EagleMiniCPMForCausalLM] PASSED 2025-12-04T09:38:40.3528353Z models/test_registry.py::test_registry_imports[Eagle3LlamaForCausalLM] PASSED 2025-12-04T09:38:40.7407781Z models/test_registry.py::test_registry_imports[LlamaForCausalLMEagle3] PASSED 2025-12-04T09:38:41.1408002Z models/test_registry.py::test_registry_imports[Eagle3Qwen2_5vlForCausalLM] PASSED 2025-12-04T09:38:41.5328920Z models/test_registry.py::test_registry_imports[EagleDeepSeekMTPModel] PASSED 2025-12-04T09:38:41.9310090Z models/test_registry.py::test_registry_imports[DeepSeekMTPModel] PASSED 2025-12-04T09:38:42.3319081Z models/test_registry.py::test_registry_imports[ErnieMTPModel] PASSED 2025-12-04T09:38:42.7313488Z models/test_registry.py::test_registry_imports[LongCatFlashMTPModel] PASSED 2025-12-04T09:38:43.1337414Z models/test_registry.py::test_registry_imports[Glm4MoeMTPModel] PASSED 2025-12-04T09:38:43.5355299Z models/test_registry.py::test_registry_imports[MedusaModel] PASSED 2025-12-04T09:38:43.9263434Z models/test_registry.py::test_registry_imports[Qwen3NextMTP] SKIPPED 2025-12-04T09:38:44.3235812Z models/test_registry.py::test_registry_imports[SmolLM3ForCausalLM] PASSED 2025-12-04T09:38:44.7203293Z models/test_registry.py::test_registry_imports[Emu3ForConditionalGeneration] PASSED 2025-12-04T09:38:45.1215071Z models/test_registry.py::test_registry_imports[TransformersForCausalLM] PASSED 2025-12-04T09:38:45.5208362Z models/test_registry.py::test_registry_imports[TransformersForMultimodalLM] PASSED 2025-12-04T09:38:45.9116326Z models/test_registry.py::test_registry_imports[TransformersMoEForCausalLM] SKIPPED 2025-12-04T09:38:46.3091123Z models/test_registry.py::test_registry_imports[TransformersMoEForMultimodalLM] SKIPPED 2025-12-04T09:38:46.7048167Z models/test_registry.py::test_registry_imports[TransformersEmbeddingModel] SKIPPED 2025-12-04T09:38:47.1018742Z models/test_registry.py::test_registry_imports[TransformersForSequenceClassification] SKIPPED 2025-12-04T09:38:47.4995942Z models/test_registry.py::test_registry_imports[TransformersMoEForSequenceClassification] SKIPPED 2025-12-04T09:38:47.8966285Z models/test_registry.py::test_registry_imports[TransformersMoEEmbeddingModel] SKIPPED 2025-12-04T09:38:47.9214447Z models/test_registry.py::test_registry_model_property[LlamaForCausalLM-False-False-False] Fork a new process to run a test 3546 2025-12-04T09:38:47.9226279Z Fork a new process to run a test 0 2025-12-04T09:38:48.4244295Z PASSED 2025-12-04T09:38:48.4485970Z models/test_registry.py::test_registry_model_property[LlavaForConditionalGeneration-True-True-False] Fork a new process to run a test 3547 2025-12-04T09:38:48.4497831Z Fork a new process to run a test 0 2025-12-04T09:38:55.7562131Z PASSED 2025-12-04T09:38:55.7802860Z models/test_registry.py::test_registry_model_property[BertForSequenceClassification-False-False-True] Fork a new process to run a test 3599 2025-12-04T09:38:55.7814434Z Fork a new process to run a test 0 2025-12-04T09:39:02.6407837Z PASSED 2025-12-04T09:39:02.6649100Z models/test_registry.py::test_registry_model_property[RobertaForSequenceClassification-False-False-True] Fork a new process to run a test 3651 2025-12-04T09:39:02.6661192Z Fork a new process to run a test 0 2025-12-04T09:39:09.5808114Z PASSED 2025-12-04T09:39:09.6048984Z models/test_registry.py::test_registry_model_property[XLMRobertaForSequenceClassification-False-False-True] Fork a new process to run a test 3703 2025-12-04T09:39:09.6060001Z Fork a new process to run a test 0 2025-12-04T09:39:10.1044309Z PASSED 2025-12-04T09:39:10.1287958Z models/test_registry.py::test_registry_is_pp[DeepseekV2ForCausalLM-True-False] Fork a new process to run a test 3704 2025-12-04T09:39:10.1299175Z Fork a new process to run a test 0 2025-12-04T09:39:16.9021173Z PASSED 2025-12-04T09:39:16.9261119Z models/test_registry.py::test_registry_is_pp[Qwen2VLForConditionalGeneration-True-True] Fork a new process to run a test 3755 2025-12-04T09:39:16.9273491Z Fork a new process to run a test 0 2025-12-04T09:39:23.7271861Z PASSED 2025-12-04T09:39:24.1237659Z models/test_registry.py::test_hf_registry_coverage PASSED 2025-12-04T09:39:24.1238010Z 2025-12-04T09:39:24.1238262Z =============================== warnings summary =============================== 2025-12-04T09:39:24.1255630Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/jit/_script.py:362: 14 warnings 2025-12-04T09:39:24.1256939Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/jit/_script.py:362: DeprecationWarning: `torch.jit.script_method` is deprecated. Please switch to `torch.compile` or `torch.export`. 2025-12-04T09:39:24.1257934Z warnings.warn( 2025-12-04T09:39:24.1258086Z 2025-12-04T09:39:24.1258219Z :488 2025-12-04T09:39:24.1258731Z :488: DeprecationWarning: builtin type SwigPyPacked has no __module__ attribute 2025-12-04T09:39:24.1259272Z 2025-12-04T09:39:24.1259371Z :488 2025-12-04T09:39:24.1259973Z :488: DeprecationWarning: builtin type SwigPyObject has no __module__ attribute 2025-12-04T09:39:24.1260682Z 2025-12-04T09:39:24.1261195Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305 2025-12-04T09:39:24.1263012Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305: DeprecationWarning: jsonschema.exceptions.RefResolutionError is deprecated as of version 4.18.0. If you wish to catch potential reference resolution errors, directly catch referencing.exceptions.Unresolvable. 2025-12-04T09:39:24.1264336Z ref_error: type[Exception] = jsonschema.RefResolutionError, 2025-12-04T09:39:24.1264552Z 2025-12-04T09:39:24.1264772Z tests/models/test_registry.py::test_registry_imports[LlavaNextForConditionalGeneration] 2025-12-04T09:39:24.1265272Z tests/models/test_registry.py::test_registry_imports[LlavaNextForConditionalGeneration] 2025-12-04T09:39:24.1265762Z tests/models/test_registry.py::test_registry_imports[LlavaNextForConditionalGeneration] 2025-12-04T09:39:24.1266239Z tests/models/test_registry.py::test_registry_imports[LlavaNextForConditionalGeneration] 2025-12-04T09:39:24.1267315Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/triton/runtime/autotuner.py:99: DeprecationWarning: warmup, rep, and use_cuda_graph parameters are deprecated. See https://github.com/triton-lang/triton/pull/4496 for details. 2025-12-04T09:39:24.1268451Z warnings.warn(("warmup, rep, and use_cuda_graph parameters are deprecated. See " 2025-12-04T09:39:24.1268717Z 2025-12-04T09:39:24.1268941Z tests/models/test_registry.py::test_registry_imports[LlavaNextForConditionalGeneration] 2025-12-04T09:39:24.1269648Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:357: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-12-04T09:39:24.1270350Z Overriding a previously registered kernel for the same operator and the same dispatch key 2025-12-04T09:39:24.1271509Z operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-12-04T09:39:24.1272488Z registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 2025-12-04T09:39:24.1272886Z dispatch key: ADInplaceOrView 2025-12-04T09:39:24.1273117Z previous kernel: no debug info 2025-12-04T09:39:24.1273842Z new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-12-04T09:39:24.1274564Z self.m.impl( 2025-12-04T09:39:24.1274666Z 2025-12-04T09:39:24.1274878Z tests/models/test_registry.py::test_registry_imports[PrithviGeoSpatialMAE] 2025-12-04T09:39:24.1275685Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pretrainedmodels/models/dpn.py:255: SyntaxWarning: "is" with 'str' literal. Did you mean "=="? 2025-12-04T09:39:24.1276219Z if block_type is 'proj': 2025-12-04T09:39:24.1276351Z 2025-12-04T09:39:24.1276540Z tests/models/test_registry.py::test_registry_imports[PrithviGeoSpatialMAE] 2025-12-04T09:39:24.1277187Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pretrainedmodels/models/dpn.py:258: SyntaxWarning: "is" with 'str' literal. Did you mean "=="? 2025-12-04T09:39:24.1277728Z elif block_type is 'down': 2025-12-04T09:39:24.1277858Z 2025-12-04T09:39:24.1278034Z tests/models/test_registry.py::test_registry_imports[PrithviGeoSpatialMAE] 2025-12-04T09:39:24.1278667Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pretrainedmodels/models/dpn.py:262: SyntaxWarning: "is" with 'str' literal. Did you mean "=="? 2025-12-04T09:39:24.1279206Z assert block_type is 'normal' 2025-12-04T09:39:24.1279345Z 2025-12-04T09:39:24.1279523Z tests/models/test_registry.py::test_registry_imports[PrithviGeoSpatialMAE] 2025-12-04T09:39:24.1280536Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/geopandas/_compat.py:7: DeprecationWarning: The 'shapely.geos' module is deprecated, and will be removed in a future version. All attributes of 'shapely.geos' are available directly from the top-level 'shapely' namespace (since shapely 2.0.0). 2025-12-04T09:39:24.1281444Z import shapely.geos 2025-12-04T09:39:24.1281564Z 2025-12-04T09:39:24.1281806Z tests/models/test_registry.py::test_registry_model_property[LlamaForCausalLM-False-False-False] 2025-12-04T09:39:24.1282396Z tests/models/test_registry.py::test_registry_model_property[LlavaForConditionalGeneration-True-True-False] 2025-12-04T09:39:24.1283024Z tests/models/test_registry.py::test_registry_model_property[BertForSequenceClassification-False-False-True] 2025-12-04T09:39:24.1283667Z tests/models/test_registry.py::test_registry_model_property[RobertaForSequenceClassification-False-False-True] 2025-12-04T09:39:24.1284346Z tests/models/test_registry.py::test_registry_model_property[XLMRobertaForSequenceClassification-False-False-True] 2025-12-04T09:39:24.1284930Z tests/models/test_registry.py::test_registry_is_pp[DeepseekV2ForCausalLM-True-False] 2025-12-04T09:39:24.1285433Z tests/models/test_registry.py::test_registry_is_pp[Qwen2VLForConditionalGeneration-True-True] 2025-12-04T09:39:24.1286379Z /var/lib/jenkins/workspace/vllm/tests/utils.py:872: DeprecationWarning: This process (pid=3489) is multi-threaded, use of fork() may lead to deadlocks in the child. 2025-12-04T09:39:24.1286974Z pid = os.fork() 2025-12-04T09:39:24.1287080Z 2025-12-04T09:39:24.1287254Z -- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html 2025-12-04T09:39:24.1287910Z =========== 207 passed, 22 skipped, 33 warnings in 114.98s (0:01:54) =========== 2025-12-04T09:39:24.5553411Z sys:1: DeprecationWarning: builtin type swigvarlink has no __module__ attribute 2025-12-04T09:39:26.3733692Z 2025-12-04 09:39:26,372 [INFO] cli.lib.core.vllm.lib: Finish running step: pytest -v -s models/test_registry.py 2025-12-04T09:39:26.3734451Z 2025-12-04 09:39:26,373 [INFO] cli.lib.core.vllm.lib: Running step: pytest -v -s models/test_utils.py 2025-12-04T09:39:26.3735006Z 2025-12-04 09:39:26,373 [INFO] cli.lib.common.utils: [shell] pytest -v -s models/test_utils.py 2025-12-04T09:39:30.9632292Z INFO 12-04 09:39:30 [__init__.py:224] Automatically detected platform cuda. 2025-12-04T09:39:32.5407335Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_asyncio/plugin.py:208: PytestDeprecationWarning: The configuration option "asyncio_default_fixture_loop_scope" is unset. 2025-12-04T09:39:32.5409355Z The event loop scope for asynchronous fixtures will default to the fixture caching scope. Future versions of pytest-asyncio will default the loop scope for asynchronous fixtures to function scope. Set the default fixture loop scope explicitly in order to avoid unexpected behavior in the future. Valid fixture loop scopes are: "function", "class", "module", "package", "session" 2025-12-04T09:39:32.5411003Z 2025-12-04T09:39:32.5411201Z warnings.warn(PytestDeprecationWarning(_DEFAULT_FIXTURE_LOOP_SCOPE_UNSET)) 2025-12-04T09:39:32.5474111Z ============================= test session starts ============================== 2025-12-04T09:39:32.5474678Z platform linux -- Python 3.12.5, pytest-8.3.5, pluggy-1.5.0 -- /opt/conda/envs/py_3.12/bin/python 2025-12-04T09:39:32.5566099Z cachedir: .pytest_cache 2025-12-04T09:39:32.5566680Z hypothesis profile 'ci' -> database=None, deadline=None, print_blob=True, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:39:32.5567313Z rootdir: /var/lib/jenkins/workspace/vllm 2025-12-04T09:39:32.5567609Z configfile: pyproject.toml 2025-12-04T09:39:32.5568642Z plugins: cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, xdist-3.3.1, xdoctest-1.3.0, hydra-core-1.3.2, anyio-4.6.2.post1, buildkite-test-collector-0.1.9, forked-1.6.0, shard-0.1.2, mock-3.14.0, schemathesis-3.39.15, subtests-0.14.1, hypothesis-6.131.0, cov-6.3.0, asyncio-0.24.0, timeout-2.3.1 2025-12-04T09:39:32.5569759Z asyncio: mode=Mode.STRICT, default_loop_scope=None 2025-12-04T09:39:33.9857690Z collecting ...  2025-12-04T09:39:33.9858409Z collecting 4 items  2025-12-04T09:39:33.9858938Z collected 4 items  2025-12-04T09:39:33.9860385Z Running 4 items in this shard: tests/models/test_utils.py::test_module_with_batchnorm_can_load, tests/models/test_utils.py::test_module_with_child_containing_batchnorm_can_autoload, tests/models/test_utils.py::test_module_skip_prefix, tests/models/test_utils.py::test_module_skip_substr 2025-12-04T09:39:33.9872049Z 2025-12-04T09:39:34.2142855Z models/test_utils.py::test_module_with_batchnorm_can_load PASSED 2025-12-04T09:39:34.4259284Z models/test_utils.py::test_module_with_child_containing_batchnorm_can_autoload PASSED 2025-12-04T09:39:34.6374170Z models/test_utils.py::test_module_skip_prefix PASSED 2025-12-04T09:39:34.8494559Z models/test_utils.py::test_module_skip_substr PASSED 2025-12-04T09:39:34.8494909Z 2025-12-04T09:39:34.8495157Z =============================== warnings summary =============================== 2025-12-04T09:39:34.8512559Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/jit/_script.py:362: 14 warnings 2025-12-04T09:39:34.8514928Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/jit/_script.py:362: DeprecationWarning: `torch.jit.script_method` is deprecated. Please switch to `torch.compile` or `torch.export`. 2025-12-04T09:39:34.8516052Z warnings.warn( 2025-12-04T09:39:34.8516188Z 2025-12-04T09:39:34.8516319Z :488 2025-12-04T09:39:34.8517069Z :488: DeprecationWarning: builtin type SwigPyPacked has no __module__ attribute 2025-12-04T09:39:34.8517661Z 2025-12-04T09:39:34.8517761Z :488 2025-12-04T09:39:34.8518341Z :488: DeprecationWarning: builtin type SwigPyObject has no __module__ attribute 2025-12-04T09:39:34.8518807Z 2025-12-04T09:39:34.8519175Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305 2025-12-04T09:39:34.8520858Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305: DeprecationWarning: jsonschema.exceptions.RefResolutionError is deprecated as of version 4.18.0. If you wish to catch potential reference resolution errors, directly catch referencing.exceptions.Unresolvable. 2025-12-04T09:39:34.8522386Z ref_error: type[Exception] = jsonschema.RefResolutionError, 2025-12-04T09:39:34.8522678Z 2025-12-04T09:39:34.8522877Z -- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html 2025-12-04T09:39:34.8523553Z ======================== 4 passed, 17 warnings in 2.30s ======================== 2025-12-04T09:39:35.0778299Z sys:1: DeprecationWarning: builtin type swigvarlink has no __module__ attribute 2025-12-04T09:39:36.0233559Z 2025-12-04 09:39:36,022 [INFO] cli.lib.core.vllm.lib: Finish running step: pytest -v -s models/test_utils.py 2025-12-04T09:39:36.0234402Z 2025-12-04 09:39:36,023 [INFO] cli.lib.core.vllm.lib: Running step: pytest -v -s models/test_vision.py 2025-12-04T09:39:36.0235191Z 2025-12-04 09:39:36,023 [INFO] cli.lib.common.utils: [shell] pytest -v -s models/test_vision.py 2025-12-04T09:39:40.5935606Z INFO 12-04 09:39:40 [__init__.py:224] Automatically detected platform cuda. 2025-12-04T09:39:42.1218554Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_asyncio/plugin.py:208: PytestDeprecationWarning: The configuration option "asyncio_default_fixture_loop_scope" is unset. 2025-12-04T09:39:42.1220701Z The event loop scope for asynchronous fixtures will default to the fixture caching scope. Future versions of pytest-asyncio will default the loop scope for asynchronous fixtures to function scope. Set the default fixture loop scope explicitly in order to avoid unexpected behavior in the future. Valid fixture loop scopes are: "function", "class", "module", "package", "session" 2025-12-04T09:39:42.1222088Z 2025-12-04T09:39:42.1222338Z warnings.warn(PytestDeprecationWarning(_DEFAULT_FIXTURE_LOOP_SCOPE_UNSET)) 2025-12-04T09:39:42.1284600Z ============================= test session starts ============================== 2025-12-04T09:39:42.1285177Z platform linux -- Python 3.12.5, pytest-8.3.5, pluggy-1.5.0 -- /opt/conda/envs/py_3.12/bin/python 2025-12-04T09:39:42.1377766Z cachedir: .pytest_cache 2025-12-04T09:39:42.1378414Z hypothesis profile 'ci' -> database=None, deadline=None, print_blob=True, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:39:42.1379062Z rootdir: /var/lib/jenkins/workspace/vllm 2025-12-04T09:39:42.1379469Z configfile: pyproject.toml 2025-12-04T09:39:42.1380517Z plugins: cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, xdist-3.3.1, xdoctest-1.3.0, hydra-core-1.3.2, anyio-4.6.2.post1, buildkite-test-collector-0.1.9, forked-1.6.0, shard-0.1.2, mock-3.14.0, schemathesis-3.39.15, subtests-0.14.1, hypothesis-6.131.0, cov-6.3.0, asyncio-0.24.0, timeout-2.3.1 2025-12-04T09:39:42.1381649Z asyncio: mode=Mode.STRICT, default_loop_scope=None 2025-12-04T09:39:43.4259480Z collecting ... WARNING 12-04 09:39:43 [interface.py:514] Current platform cuda does not have '_pytestfixturefunction' attribute. 2025-12-04T09:39:43.7280875Z WARNING 12-04 09:39:43 [interface.py:514] Current platform cuda does not have '__test__' attribute. 2025-12-04T09:39:43.7281558Z WARNING 12-04 09:39:43 [interface.py:514] Current platform cuda does not have '__bases__' attribute. 2025-12-04T09:39:43.7282169Z WARNING 12-04 09:39:43 [interface.py:514] Current platform cuda does not have '__test__' attribute. 2025-12-04T09:39:43.7282808Z WARNING 12-04 09:39:43 [interface.py:514] Current platform cuda does not have '_schemathesis_test' attribute. 2025-12-04T09:39:43.7314226Z  2025-12-04T09:39:43.7318920Z collecting 19 items  2025-12-04T09:39:43.7319451Z collected 19 items  2025-12-04T09:39:43.7325567Z Running 19 items in this shard: tests/models/test_vision.py::test_resolve_visual_encoder_outputs[select_layers0-10-10-expected_features0], tests/models/test_vision.py::test_resolve_visual_encoder_outputs[select_layers1-10-10-expected_features1], tests/models/test_vision.py::test_resolve_visual_encoder_outputs[select_layers2-10-20-expected_features2], tests/models/test_vision.py::test_resolve_visual_encoder_outputs[select_layers3-10-20-expected_features3], tests/models/test_vision.py::test_run_dp_sharded_vision_model[1], tests/models/test_vision.py::test_run_dp_sharded_vision_model[4], tests/models/test_vision.py::test_run_dp_sharded_vision_model[5], tests/models/test_vision.py::test_get_load_balance_assignment_cases[sizes0-2-expected_shuffle_indices0-expected_gpu_sample_counts0-expected_grouped_sizes_per_gpu0-empty input], tests/models/test_vision.py::test_get_load_balance_assignment_cases[sizes1-4-expected_shuffle_indices1-expected_gpu_sample_counts1-expected_grouped_sizes_per_gpu1-fewer samples than GPUs], tests/models/test_vision.py::test_get_load_balance_assignment_cases[sizes2-1-expected_shuffle_indices2-expected_gpu_sample_counts2-expected_grouped_sizes_per_gpu2-single GPU], tests/models/test_vision.py::test_get_load_balance_assignment_cases[sizes3-2-expected_shuffle_indices3-expected_gpu_sample_counts3-expected_grouped_sizes_per_gpu3-balanced assignment], tests/models/test_vision.py::test_get_load_balance_assignment_cases[sizes4-2-expected_shuffle_indices4-expected_gpu_sample_counts4-expected_grouped_sizes_per_gpu4-unbalanced sizes], tests/models/test_vision.py::test_run_dp_sharded_mrope_vision_model[1], tests/models/test_vision.py::test_run_dp_sharded_mrope_vision_model[3], tests/models/test_vision.py::test_run_dp_sharded_mrope_vision_model[5], tests/models/test_vision.py::test_run_dp_sharded_mrope_vision_model_empty_input, tests/models/test_vision.py::test_run_dp_sharded_mrope_vision_model_uneven_load, tests/models/test_vision.py::test_simple_mrope_vision_model_spatial_merge[2], tests/models/test_vision.py::test_simple_mrope_vision_model_spatial_merge[4] 2025-12-04T09:39:43.7332147Z 2025-12-04T09:39:43.9577766Z models/test_vision.py::test_resolve_visual_encoder_outputs[select_layers0-10-10-expected_features0] PASSED 2025-12-04T09:39:44.1769122Z models/test_vision.py::test_resolve_visual_encoder_outputs[select_layers1-10-10-expected_features1] PASSED 2025-12-04T09:39:44.3952712Z models/test_vision.py::test_resolve_visual_encoder_outputs[select_layers2-10-20-expected_features2] PASSED 2025-12-04T09:39:44.6131318Z models/test_vision.py::test_resolve_visual_encoder_outputs[select_layers3-10-20-expected_features3] PASSED 2025-12-04T09:39:44.6139365Z models/test_vision.py::test_run_dp_sharded_vision_model[1] SKIPPED (...) 2025-12-04T09:39:44.6146828Z models/test_vision.py::test_run_dp_sharded_vision_model[4] SKIPPED (...) 2025-12-04T09:39:44.6153415Z models/test_vision.py::test_run_dp_sharded_vision_model[5] SKIPPED (...) 2025-12-04T09:39:44.8330452Z models/test_vision.py::test_get_load_balance_assignment_cases[sizes0-2-expected_shuffle_indices0-expected_gpu_sample_counts0-expected_grouped_sizes_per_gpu0-empty input] PASSED 2025-12-04T09:39:45.0513194Z models/test_vision.py::test_get_load_balance_assignment_cases[sizes1-4-expected_shuffle_indices1-expected_gpu_sample_counts1-expected_grouped_sizes_per_gpu1-fewer samples than GPUs] PASSED 2025-12-04T09:39:45.2687864Z models/test_vision.py::test_get_load_balance_assignment_cases[sizes2-1-expected_shuffle_indices2-expected_gpu_sample_counts2-expected_grouped_sizes_per_gpu2-single GPU] PASSED 2025-12-04T09:39:45.4868741Z models/test_vision.py::test_get_load_balance_assignment_cases[sizes3-2-expected_shuffle_indices3-expected_gpu_sample_counts3-expected_grouped_sizes_per_gpu3-balanced assignment] PASSED 2025-12-04T09:39:45.7044933Z models/test_vision.py::test_get_load_balance_assignment_cases[sizes4-2-expected_shuffle_indices4-expected_gpu_sample_counts4-expected_grouped_sizes_per_gpu4-unbalanced sizes] PASSED 2025-12-04T09:39:45.7051558Z models/test_vision.py::test_run_dp_sharded_mrope_vision_model[1] SKIPPED 2025-12-04T09:39:45.7058284Z models/test_vision.py::test_run_dp_sharded_mrope_vision_model[3] SKIPPED 2025-12-04T09:39:45.7064547Z models/test_vision.py::test_run_dp_sharded_mrope_vision_model[5] SKIPPED 2025-12-04T09:39:45.7070869Z models/test_vision.py::test_run_dp_sharded_mrope_vision_model_empty_input SKIPPED 2025-12-04T09:39:45.7077543Z models/test_vision.py::test_run_dp_sharded_mrope_vision_model_uneven_load SKIPPED 2025-12-04T09:39:46.2668915Z models/test_vision.py::test_simple_mrope_vision_model_spatial_merge[2] PASSED 2025-12-04T09:39:46.4900679Z models/test_vision.py::test_simple_mrope_vision_model_spatial_merge[4] PASSED 2025-12-04T09:39:46.4901077Z 2025-12-04T09:39:46.4901318Z =============================== warnings summary =============================== 2025-12-04T09:39:46.4918231Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/jit/_script.py:362: 14 warnings 2025-12-04T09:39:46.4919667Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/jit/_script.py:362: DeprecationWarning: `torch.jit.script_method` is deprecated. Please switch to `torch.compile` or `torch.export`. 2025-12-04T09:39:46.4920664Z warnings.warn( 2025-12-04T09:39:46.4920813Z 2025-12-04T09:39:46.4920933Z :488 2025-12-04T09:39:46.4921570Z :488: DeprecationWarning: builtin type SwigPyPacked has no __module__ attribute 2025-12-04T09:39:46.4922020Z 2025-12-04T09:39:46.4922144Z :488 2025-12-04T09:39:46.4922742Z :488: DeprecationWarning: builtin type SwigPyObject has no __module__ attribute 2025-12-04T09:39:46.4923256Z 2025-12-04T09:39:46.4923596Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305 2025-12-04T09:39:46.4924968Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305: DeprecationWarning: jsonschema.exceptions.RefResolutionError is deprecated as of version 4.18.0. If you wish to catch potential reference resolution errors, directly catch referencing.exceptions.Unresolvable. 2025-12-04T09:39:46.4926136Z ref_error: type[Exception] = jsonschema.RefResolutionError, 2025-12-04T09:39:46.4926375Z 2025-12-04T09:39:46.4926622Z -- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html 2025-12-04T09:39:46.4927332Z ================== 11 passed, 8 skipped, 17 warnings in 4.36s ================== 2025-12-04T09:39:46.7250685Z sys:1: DeprecationWarning: builtin type swigvarlink has no __module__ attribute 2025-12-04T09:39:47.8258829Z 2025-12-04 09:39:47,825 [INFO] cli.lib.core.vllm.lib: Finish running step: pytest -v -s models/test_vision.py 2025-12-04T09:39:47.8259854Z 2025-12-04 09:39:47,825 [INFO] cli.lib.core.vllm.lib: Running step: pytest -v -s models/test_initialization.py 2025-12-04T09:39:47.8261038Z 2025-12-04 09:39:47,825 [INFO] cli.lib.common.utils: [shell] pytest -v -s models/test_initialization.py 2025-12-04T09:39:52.3600682Z INFO 12-04 09:39:52 [__init__.py:224] Automatically detected platform cuda. 2025-12-04T09:39:53.8910418Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_asyncio/plugin.py:208: PytestDeprecationWarning: The configuration option "asyncio_default_fixture_loop_scope" is unset. 2025-12-04T09:39:53.8912458Z The event loop scope for asynchronous fixtures will default to the fixture caching scope. Future versions of pytest-asyncio will default the loop scope for asynchronous fixtures to function scope. Set the default fixture loop scope explicitly in order to avoid unexpected behavior in the future. Valid fixture loop scopes are: "function", "class", "module", "package", "session" 2025-12-04T09:39:53.8913813Z 2025-12-04T09:39:53.8914058Z warnings.warn(PytestDeprecationWarning(_DEFAULT_FIXTURE_LOOP_SCOPE_UNSET)) 2025-12-04T09:39:53.8975784Z ============================= test session starts ============================== 2025-12-04T09:39:53.8976382Z platform linux -- Python 3.12.5, pytest-8.3.5, pluggy-1.5.0 -- /opt/conda/envs/py_3.12/bin/python 2025-12-04T09:39:53.9068620Z cachedir: .pytest_cache 2025-12-04T09:39:53.9069273Z hypothesis profile 'ci' -> database=None, deadline=None, print_blob=True, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:39:53.9069928Z rootdir: /var/lib/jenkins/workspace/vllm 2025-12-04T09:39:53.9070234Z configfile: pyproject.toml 2025-12-04T09:39:53.9071877Z plugins: cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, xdist-3.3.1, xdoctest-1.3.0, hydra-core-1.3.2, anyio-4.6.2.post1, buildkite-test-collector-0.1.9, forked-1.6.0, shard-0.1.2, mock-3.14.0, schemathesis-3.39.15, subtests-0.14.1, hypothesis-6.131.0, cov-6.3.0, asyncio-0.24.0, timeout-2.3.1 2025-12-04T09:39:53.9072969Z asyncio: mode=Mode.STRICT, default_loop_scope=None 2025-12-04T09:39:55.4865628Z collecting ...  2025-12-04T09:39:55.4888154Z collecting 225 items  2025-12-04T09:39:55.4888738Z collected 225 items  2025-12-04T09:39:55.4954124Z Running 225 items in this shard: tests/models/test_initialization.py::test_can_initialize_small_subset[LlavaForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_small_subset[Llama4ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_small_subset[BertForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_small_subset[Gemma3nForCausalLM], tests/models/test_initialization.py::test_can_initialize_small_subset[JinaVLForRanking], tests/models/test_initialization.py::test_can_initialize_small_subset[InternVLChatModel], tests/models/test_initialization.py::test_can_initialize_small_subset[InternLM2ForRewardModel], tests/models/test_initialization.py::test_can_initialize_small_subset[TransformersForMultimodalLM], tests/models/test_initialization.py::test_can_initialize_small_subset[PrithviGeoSpatialMAE], tests/models/test_initialization.py::test_can_initialize_small_subset[UltravoxModel], tests/models/test_initialization.py::test_can_initialize_small_subset[DeepSeekMTPModel], tests/models/test_initialization.py::test_can_initialize_small_subset[XLMRobertaModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Starcoder2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Gemma3TextModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Step3VLForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[TeleFLMForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[DeepseekV3ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Phi4MultimodalForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[InternLMForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[StableLMEpochForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[CohereForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[BambaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2_5_VLForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[TransformersEmbeddingModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Gemma3ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[InternLM2VEForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[FalconForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[AriaForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Phi3VForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[TransformersMoEForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[DbrxForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[ChameleonForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[GraniteSpeechForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[NemotronH_Nano_VL_V2], tests/models/test_initialization.py::test_can_initialize_large_subset[ModernBertForTokenClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2AudioForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[DeepseekV2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GritLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GteNewForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[DeepseekForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Gemma3ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Llama4ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Exaone4ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GPT2LMHeadModel], tests/models/test_initialization.py::test_can_initialize_large_subset[KeyeVL1_5ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[WhisperForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[MiDashengLMModel], tests/models/test_initialization.py::test_can_initialize_large_subset[ApertusForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2ForProcessRewardModel], tests/models/test_initialization.py::test_can_initialize_large_subset[MiniMaxM1ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Glm4MoeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[JambaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[ExaoneForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MistralModel], tests/models/test_initialization.py::test_can_initialize_large_subset[EagleDeepSeekMTPModel], tests/models/test_initialization.py::test_can_initialize_large_subset[QWenLMHeadModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Dots1ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[ModernBertModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Ernie4_5_MoeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[AquilaModel], tests/models/test_initialization.py::test_can_initialize_large_subset[XLMRobertaForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[MiniCPMV], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen3NextMTP], tests/models/test_initialization.py::test_can_initialize_large_subset[JAISLMHeadModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Gemma3nForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Glm4MoeMTPModel], tests/models/test_initialization.py::test_can_initialize_large_subset[GteModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Mamba2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GraniteMoeHybridForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2VLForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[GraniteForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[BailingMoeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[BertModel], tests/models/test_initialization.py::test_can_initialize_large_subset[MolmoForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[DeepseekVLV2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Plamo2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[DotsOCRForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Gemma2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[LlamaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[H2OVLChatModel], tests/models/test_initialization.py::test_can_initialize_large_subset[EagleMiniCPMForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2_5OmniForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[ArcticForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[KeyeForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Glm4ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[OlmoeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Glm4vMoeForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[PixtralForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[StableLmForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[DeepseekV32ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[InternLM3ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[ChatGLMForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[MiniMaxForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Zamba2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[JambaForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[Cohere2VisionForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[LongCatFlashMTPModel], tests/models/test_initialization.py::test_can_initialize_large_subset[HunYuanMoEV1ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[NomicBertModel], tests/models/test_initialization.py::test_can_initialize_large_subset[MptForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[TeleChat2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GteNewModel], tests/models/test_initialization.py::test_can_initialize_large_subset[SmolVLMForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[GPTBigCodeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[HunYuanDenseV1ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MedusaModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2ForRewardModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen3VLMoeForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[RWForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GraniteMoeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[ModernBertForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[Lfm2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Cohere2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[FuyuForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[TransformersMoEForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[BloomForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GlmForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[SmolLM3ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[ChatGLMModel], tests/models/test_initialization.py::test_can_initialize_large_subset[TransformersMoEForMultimodalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Lfm2MoeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Eagle3LlamaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[OrionForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[PhiMoEForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Emu3ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[SeedOssForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[HCXVisionForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Eagle3Qwen2_5vlForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Phi3ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[InternVLForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[AquilaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GLM4VForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Idefics3ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[CLIPModel], tests/models/test_initialization.py::test_can_initialize_large_subset[LlavaOnevisionForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[PhiForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[LlamaForCausalLMEagle3], tests/models/test_initialization.py::test_can_initialize_large_subset[NemotronForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[CwmForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[RobertaForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[MambaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MiMoForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Glm4vForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[ErnieMTPModel], tests/models/test_initialization.py::test_can_initialize_large_subset[GemmaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[TarsierForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen3ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[BailingMoeV2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[SolarForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GPTNeoXForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[EagleLlamaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[RForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[LongcatFlashForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MantisForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[PersimmonForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Step3TextForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[InternLM2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[LlamaModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen3MoeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[ArceeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[NVLM_D], tests/models/test_initialization.py::test_can_initialize_large_subset[Ovis2_5], tests/models/test_initialization.py::test_can_initialize_large_subset[FalconH1ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MistralForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[QwenVLForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Mistral3ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen3VLForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Olmo2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MiMoMTPModel], tests/models/test_initialization.py::test_can_initialize_large_subset[GptOssForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[FalconMambaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GPTJForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen3NextForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Olmo3ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2Model], tests/models/test_initialization.py::test_can_initialize_large_subset[EagleLlama4ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2_5OmniModel], tests/models/test_initialization.py::test_can_initialize_large_subset[SkyworkR1VChatModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Phi4MMForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Gemma2Model], tests/models/test_initialization.py::test_can_initialize_large_subset[TransformersForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[AyaVisionForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Ernie4_5_VLMoeForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[MixtralForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[KimiVLForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[MiniMaxVL01ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[MiniCPMForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[BaichuanForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MPTForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GraniteMoeSharedForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[DeciLMForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[BaiChuanForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Terratorch], tests/models/test_initialization.py::test_can_initialize_large_subset[TransformersForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[LlavaNextForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Tarsier2ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[OlmoForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Fairseq2LlamaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GPT2ForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[Llama_Nemotron_Nano_VL], tests/models/test_initialization.py::test_can_initialize_large_subset[TransformersMoEEmbeddingModel], tests/models/test_initialization.py::test_can_initialize_large_subset[RobertaForMaskedLM], tests/models/test_initialization.py::test_can_initialize_large_subset[LLaMAForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[InternS1ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[XverseForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[OPTForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Ernie4_5ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[VoxtralForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Blip2ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[PaliGemmaForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[MiniMaxText01ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[LlavaNextVideoForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[NemotronHForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2MoeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Grok1ModelForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MiniCPMO], tests/models/test_initialization.py::test_can_initialize_large_subset[MiniCPM3ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[RobertaModel], tests/models/test_initialization.py::test_can_initialize_large_subset[BertForTokenClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[Ovis], tests/models/test_initialization.py::test_implicit_converted_models[GemmaForSequenceClassification], tests/models/test_initialization.py::test_implicit_converted_models[LlamaForSequenceClassification], tests/models/test_initialization.py::test_implicit_converted_models[Qwen2ForSequenceClassification], tests/models/test_initialization.py::test_implicit_converted_models[Qwen3ForSequenceClassification] 2025-12-04T09:39:55.5015469Z 2025-12-04T09:39:55.5061945Z models/test_initialization.py::test_can_initialize_small_subset[LlavaForConditionalGeneration] Fork a new process to run a test 3973 2025-12-04T09:39:55.5070248Z Fork a new process to run a test 0 2025-12-04T09:39:55.5337447Z INFO 12-04 09:39:55 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='LlavaForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'llava-hf/llava-1.5-7b-hf'} 2025-12-04T09:39:55.6133391Z 2025-12-04T09:39:55.6134549Z config.json: 0% 0.00/950 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:40:04.6422612Z (EngineCore_DP0 pid=4002) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:357: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-12-04T09:40:04.6423938Z (EngineCore_DP0 pid=4002) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-12-04T09:40:04.6428201Z (EngineCore_DP0 pid=4002) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-12-04T09:40:04.6429662Z (EngineCore_DP0 pid=4002) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 2025-12-04T09:40:04.6430271Z (EngineCore_DP0 pid=4002) dispatch key: ADInplaceOrView 2025-12-04T09:40:04.6431031Z (EngineCore_DP0 pid=4002) previous kernel: no debug info 2025-12-04T09:40:04.6432077Z (EngineCore_DP0 pid=4002) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-12-04T09:40:04.6432998Z (EngineCore_DP0 pid=4002) self.m.impl( 2025-12-04T09:40:05.0307982Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:05.0335823Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:05.0343085Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:05.0350872Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:05.0357767Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:05.0364825Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:05.0367349Z (EngineCore_DP0 pid=4002) INFO 12-04 09:40:05 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:40:05.2860835Z (EngineCore_DP0 pid=4002) INFO 12-04 09:40:05 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:40:05.4058526Z (EngineCore_DP0 pid=4002) 2025-12-04T09:40:05.4058952Z processor_config.json: 0% 0.00/173 [00:00, model_arch='Llama4ForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'meta-llama/Llama-4-Scout-17B-16E-Instruct'} 2025-12-04T09:40:15.5620442Z 2025-12-04T09:40:15.5621630Z config.json: 0% 0.00/2.18k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:40:26.2799046Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:26.2827223Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:26.2834542Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:26.2842232Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:26.2850081Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:26.2858003Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:26.2861105Z (EngineCore_DP0 pid=4348) INFO 12-04 09:40:26 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:40:26.5644276Z (EngineCore_DP0 pid=4348) INFO 12-04 09:40:26 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:40:26.8143185Z (EngineCore_DP0 pid=4348) 2025-12-04T09:40:26.8143592Z processor_config.json: 0% 0.00/128 [00:00, model_arch='BertForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'cross-encoder/ms-marco-MiniLM-L-6-v2'} 2025-12-04T09:40:34.1186767Z 2025-12-04T09:40:34.1187709Z config.json: 0% 0.00/794 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:40:36.6979843Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:36.7010302Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:36.7017587Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:36.7024941Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:36.7032342Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:36.7040100Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:36.7042980Z (EngineCore_DP0 pid=4442) INFO 12-04 09:40:36 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:40:36.9872487Z (EngineCore_DP0 pid=4442) INFO 12-04 09:40:36 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:40:36.9973697Z (EngineCore_DP0 pid=4442) INFO 12-04 09:40:36 [gpu_model_runner.py:2840] Starting to load model cross-encoder/ms-marco-MiniLM-L-6-v2... 2025-12-04T09:40:37.2644015Z (EngineCore_DP0 pid=4442) INFO 12-04 09:40:37 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:40:37.2688211Z (EngineCore_DP0 pid=4442) INFO 12-04 09:40:37 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:40:37.8948063Z (EngineCore_DP0 pid=4442) INFO 12-04 09:40:37 [gpu_model_runner.py:2902] Model loading took 0.0263 GiB and 0.065808 seconds 2025-12-04T09:40:38.0464083Z (EngineCore_DP0 pid=4442) INFO 12-04 09:40:38 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T09:40:38.1377805Z (EngineCore_DP0 pid=4442) INFO 12-04 09:40:38 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T09:40:38.4249218Z (EngineCore_DP0 pid=4442) INFO 12-04 09:40:38 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:40:38.4269078Z INFO 12-04 09:40:38 [llm.py:337] Supported tasks: ['encode', 'classify', 'score'] 2025-12-04T09:40:38.8417679Z PASSED 2025-12-04T09:40:38.8565712Z models/test_initialization.py::test_can_initialize_small_subset[Gemma3nForCausalLM] Fork a new process to run a test 4512 2025-12-04T09:40:38.8575280Z Fork a new process to run a test 0 2025-12-04T09:40:38.8852411Z INFO 12-04 09:40:38 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Gemma3nForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'google/gemma-3n-E2B-it'} 2025-12-04T09:40:39.0028047Z 2025-12-04T09:40:39.0029706Z config.json: 0% 0.00/4.25k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:40:51.6155247Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:51.6193610Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:51.6200719Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:51.6211870Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:51.6220741Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:51.6228695Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:40:51.6231353Z (EngineCore_DP0 pid=4591) INFO 12-04 09:40:51 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:40:51.9172110Z (EngineCore_DP0 pid=4591) INFO 12-04 09:40:51 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:40:52.1228991Z (EngineCore_DP0 pid=4591) 2025-12-04T09:40:52.1231052Z processor_config.json: 0% 0.00/98.0 [00:00, model_arch='JinaVLForRanking', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'jinaai/jina-reranker-m0'} 2025-12-04T09:41:07.5136423Z 2025-12-04T09:41:07.5138489Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:41:07.5138810Z config.json: 1.16kB [00:00, 7.56MB/s] 2025-12-04T09:41:07.6144998Z 2025-12-04T09:41:07.6147783Z preprocessor_config.json: 0% 0.00/316 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:41:16.8441967Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:16.8479982Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:16.8487431Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:16.8495036Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:16.8503276Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:16.8510922Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:16.8513570Z (EngineCore_DP0 pid=4772) INFO 12-04 09:41:16 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:41:17.1365739Z (EngineCore_DP0 pid=4772) INFO 12-04 09:41:17 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:41:18.7284262Z (EngineCore_DP0 pid=4772) 2025-12-04T09:41:18.7285193Z chat_template.json: 0.00B [00:00, ?B/s] 2025-12-04T09:41:18.7285587Z chat_template.json: 1.05kB [00:00, 3.56MB/s] 2025-12-04T09:41:19.2919984Z (EngineCore_DP0 pid=4772) INFO 12-04 09:41:19 [gpu_model_runner.py:2840] Starting to load model jinaai/jina-reranker-m0... 2025-12-04T09:41:19.6564241Z (EngineCore_DP0 pid=4772) INFO 12-04 09:41:19 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:41:19.7267745Z (EngineCore_DP0 pid=4772) INFO 12-04 09:41:19 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:41:20.3463717Z (EngineCore_DP0 pid=4772) INFO 12-04 09:41:20 [gpu_model_runner.py:2902] Model loading took 1.8029 GiB and 0.137159 seconds 2025-12-04T09:41:20.3467465Z (EngineCore_DP0 pid=4772) INFO 12-04 09:41:20 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T09:41:20.3468469Z (EngineCore_DP0 pid=4772) INFO 12-04 09:41:20 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 320.00x 2025-12-04T09:41:22.0737907Z (EngineCore_DP0 pid=4772) INFO 12-04 09:41:22 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:41:22.0752173Z INFO 12-04 09:41:22 [llm.py:337] Supported tasks: ['classify', 'score'] 2025-12-04T09:41:22.5594761Z PASSED 2025-12-04T09:41:22.5743130Z models/test_initialization.py::test_can_initialize_small_subset[InternVLChatModel] Fork a new process to run a test 4842 2025-12-04T09:41:22.5752153Z Fork a new process to run a test 0 2025-12-04T09:41:22.6026717Z INFO 12-04 09:41:22 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='InternVLChatModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'OpenGVLab/InternVL2-1B'} 2025-12-04T09:41:22.6706779Z 2025-12-04T09:41:22.6708711Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:41:22.6709008Z config.json: 3.73kB [00:00, 22.0MB/s] 2025-12-04T09:41:22.7310701Z 2025-12-04T09:41:22.7311758Z configuration_internvl_chat.py: 0.00B [00:00, ?B/s] 2025-12-04T09:41:22.7312219Z configuration_internvl_chat.py: 4.04kB [00:00, 33.6MB/s] 2025-12-04T09:41:22.7835950Z 2025-12-04T09:41:22.7836978Z configuration_intern_vit.py: 0.00B [00:00, ?B/s] 2025-12-04T09:41:22.7837573Z configuration_intern_vit.py: 5.55kB [00:00, 45.8MB/s] 2025-12-04T09:41:22.7868016Z A new version of the following files was downloaded from https://huggingface.co/OpenGVLab/InternVL2-1B: 2025-12-04T09:41:22.7868843Z - configuration_intern_vit.py 2025-12-04T09:41:22.7869718Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:41:22.7871280Z A new version of the following files was downloaded from https://huggingface.co/OpenGVLab/InternVL2-1B: 2025-12-04T09:41:22.7871966Z - configuration_internvl_chat.py 2025-12-04T09:41:22.7872338Z - configuration_intern_vit.py 2025-12-04T09:41:22.7873242Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:41:23.0755449Z 2025-12-04T09:41:23.0755870Z preprocessor_config.json: 0% 0.00/287 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:41:31.9472806Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:31.9501703Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:31.9508634Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:31.9515475Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:31.9522892Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:31.9530251Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:31.9533388Z (EngineCore_DP0 pid=4900) INFO 12-04 09:41:31 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:41:32.2319319Z (EngineCore_DP0 pid=4900) INFO 12-04 09:41:32 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:41:32.2915049Z (EngineCore_DP0 pid=4900) WARNING 12-04 09:41:32 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T09:41:32.2944733Z (EngineCore_DP0 pid=4900) WARNING 12-04 09:41:32 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T09:41:32.3627523Z (EngineCore_DP0 pid=4900) INFO 12-04 09:41:32 [gpu_model_runner.py:2840] Starting to load model OpenGVLab/InternVL2-1B... 2025-12-04T09:41:32.6364541Z (EngineCore_DP0 pid=4900) INFO 12-04 09:41:32 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:41:32.6408053Z (EngineCore_DP0 pid=4900) INFO 12-04 09:41:32 [layer.py:497] MultiHeadAttention attn_backend: _Backend.FLASH_ATTN, use_upstream_fa: False 2025-12-04T09:41:32.7059870Z (EngineCore_DP0 pid=4900) INFO 12-04 09:41:32 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:41:33.3109612Z (EngineCore_DP0 pid=4900) INFO 12-04 09:41:33 [gpu_model_runner.py:2902] Model loading took 0.5767 GiB and 0.129604 seconds 2025-12-04T09:41:33.3113960Z (EngineCore_DP0 pid=4900) INFO 12-04 09:41:33 [kv_cache_utils.py:1199] GPU KV cache size: 20,971,520 tokens 2025-12-04T09:41:33.3114699Z (EngineCore_DP0 pid=4900) INFO 12-04 09:41:33 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 640.00x 2025-12-04T09:41:33.7477565Z (EngineCore_DP0 pid=4900) WARNING 12-04 09:41:33 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T09:41:36.0843910Z (EngineCore_DP0 pid=4900) INFO 12-04 09:41:36 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:41:36.0858375Z INFO 12-04 09:41:36 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:41:36.5600675Z PASSED 2025-12-04T09:41:36.5749729Z models/test_initialization.py::test_can_initialize_small_subset[InternLM2ForRewardModel] Fork a new process to run a test 4986 2025-12-04T09:41:36.5757996Z Fork a new process to run a test 0 2025-12-04T09:41:36.6035193Z INFO 12-04 09:41:36 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='InternLM2ForRewardModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'internlm/internlm2-1_8b-reward'} 2025-12-04T09:41:36.6880303Z 2025-12-04T09:41:36.6881603Z config.json: 0% 0.00/813 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:41:46.1317968Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:46.1345925Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:46.1352803Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:46.1360270Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:46.1367891Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:46.1375532Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:41:46.1378291Z (EngineCore_DP0 pid=5065) INFO 12-04 09:41:46 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:41:46.4243389Z (EngineCore_DP0 pid=5065) INFO 12-04 09:41:46 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:41:46.5039264Z (EngineCore_DP0 pid=5065) INFO 12-04 09:41:46 [gpu_model_runner.py:2840] Starting to load model internlm/internlm2-1_8b-reward... 2025-12-04T09:41:46.7834062Z (EngineCore_DP0 pid=5065) INFO 12-04 09:41:46 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:41:46.8482056Z (EngineCore_DP0 pid=5065) INFO 12-04 09:41:46 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:41:47.4595326Z (EngineCore_DP0 pid=5065) INFO 12-04 09:41:47 [gpu_model_runner.py:2902] Model loading took 0.4785 GiB and 0.126880 seconds 2025-12-04T09:41:47.4599299Z (EngineCore_DP0 pid=5065) INFO 12-04 09:41:47 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T09:41:47.4600245Z (EngineCore_DP0 pid=5065) INFO 12-04 09:41:47 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 80.00x 2025-12-04T09:41:48.7187634Z (EngineCore_DP0 pid=5065) INFO 12-04 09:41:48 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T09:41:49.0034591Z (EngineCore_DP0 pid=5065) INFO 12-04 09:41:49 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:41:49.0048619Z INFO 12-04 09:41:49 [llm.py:337] Supported tasks: ['encode'] 2025-12-04T09:41:49.4800296Z PASSED 2025-12-04T09:41:49.4948990Z models/test_initialization.py::test_can_initialize_small_subset[TransformersForMultimodalLM] Fork a new process to run a test 5135 2025-12-04T09:41:49.4958153Z Fork a new process to run a test 0 2025-12-04T09:41:49.5228271Z INFO 12-04 09:41:49 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='TransformersForMultimodalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'transformers', 'model': 'BAAI/Emu3-Chat-hf'} 2025-12-04T09:41:49.6041287Z 2025-12-04T09:41:49.6280583Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:41:49.6280894Z config.json: 6.55MB [00:00, 274MB/s] 2025-12-04T09:41:50.0854129Z 2025-12-04T09:41:50.0854956Z preprocessor_config.json: 0% 0.00/611 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:42:20.2319008Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:42:20.2347079Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:42:20.2353780Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:42:20.2361757Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:42:20.2369479Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:42:20.2377950Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:42:20.2380846Z (EngineCore_DP0 pid=5215) INFO 12-04 09:42:20 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:42:20.5247156Z (EngineCore_DP0 pid=5215) INFO 12-04 09:42:20 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:42:20.6613560Z (EngineCore_DP0 pid=5215) 2025-12-04T09:42:20.6614557Z processor_config.json: 0% 0.00/69.0 [00:00, model_arch='PrithviGeoSpatialMAE', exist_overrides={}, use_original_num_layers=False), 'enforce_eager': True, 'model_impl': 'vllm', 'model': 'ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11'} 2025-12-04T09:43:06.9876203Z 2025-12-04T09:43:06.9878305Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:43:06.9878616Z config.json: 3.45kB [00:00, 18.3MB/s] 2025-12-04T09:43:17.7579198Z INFO 12-04 09:43:17 [model.py:915] Resolved `--runner auto` to `--runner pooling`. Pass the value explicitly to silence this message. 2025-12-04T09:43:17.7579908Z INFO 12-04 09:43:17 [model.py:653] Resolved architecture: Terratorch 2025-12-04T09:43:17.7580357Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:43:17.8154396Z ERROR 12-04 09:43:17 [config.py:298] Error retrieving safetensors: 'ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11' is not a safetensors repo. Couldn't find 'model.safetensors.index.json' or 'model.safetensors' files., retrying 1 of 2 2025-12-04T09:43:19.8687103Z ERROR 12-04 09:43:19 [config.py:296] Error retrieving safetensors: 'ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11' is not a safetensors repo. Couldn't find 'model.safetensors.index.json' or 'model.safetensors' files. 2025-12-04T09:43:19.8688164Z INFO 12-04 09:43:19 [model.py:1939] Downcasting torch.float32 to torch.float16. 2025-12-04T09:43:19.8689548Z WARNING 12-04 09:43:19 [model.py:2042] The model's config.json does not contain any of the following keys to determine the original maximum length of the model: ['max_position_embeddings', 'n_positions', 'max_seq_len', 'seq_length', 'model_max_length', 'max_target_positions', 'max_sequence_length', 'max_seq_length', 'seq_len']. Assuming the model's maximum length is 2048. 2025-12-04T09:43:19.8690846Z INFO 12-04 09:43:19 [model.py:1714] Using max model len 2048 2025-12-04T09:43:19.8982532Z INFO 12-04 09:43:19 [arg_utils.py:1725] (Disabling) chunked prefill by default 2025-12-04T09:43:19.8983025Z INFO 12-04 09:43:19 [arg_utils.py:1728] (Disabling) prefix caching by default 2025-12-04T09:43:20.0452832Z INFO 12-04 09:43:20 [vllm.py:354] Cudagraph is disabled under eager mode 2025-12-04T09:43:20.0453794Z INFO 12-04 09:43:20 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T09:43:23.6382018Z WARNING 12-04 09:43:23 [__init__.py:2879] We must use the `spawn` multiprocessing start method. Overriding VLLM_WORKER_MULTIPROC_METHOD to 'spawn'. See https://docs.vllm.ai/en/latest/usage/troubleshooting.html#python-multiprocessing for more information. Reasons: CUDA is initialized 2025-12-04T09:43:26.4602884Z INFO 12-04 09:43:26 [__init__.py:224] Automatically detected platform cuda. 2025-12-04T09:43:29.0239067Z (EngineCore_DP0 pid=5347) INFO 12-04 09:43:29 [core.py:727] Waiting for init message from front-end. 2025-12-04T09:43:29.0498704Z (EngineCore_DP0 pid=5347) INFO 12-04 09:43:29 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11', speculative_config=None, tokenizer='ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11', skip_tokenizer_init=True, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11, enable_prefix_caching=False, chunked_prefill_enabled=False, pooler_config=PoolerConfig(pooling_type='All', normalize=None, dimensions=None, enable_chunked_processing=None, max_embed_len=None, activation=None, logit_bias=None, softmax=None, step_tag_id=None, returned_token_ids=None), compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-12-04T09:43:33.2779016Z (EngineCore_DP0 pid=5347) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:357: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-12-04T09:43:33.2780424Z (EngineCore_DP0 pid=5347) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-12-04T09:43:33.2782186Z (EngineCore_DP0 pid=5347) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-12-04T09:43:33.2783720Z (EngineCore_DP0 pid=5347) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 2025-12-04T09:43:33.2784471Z (EngineCore_DP0 pid=5347) dispatch key: ADInplaceOrView 2025-12-04T09:43:33.2784997Z (EngineCore_DP0 pid=5347) previous kernel: no debug info 2025-12-04T09:43:33.2786247Z (EngineCore_DP0 pid=5347) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-12-04T09:43:33.2787490Z (EngineCore_DP0 pid=5347) self.m.impl( 2025-12-04T09:43:33.5709064Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:43:33.5735198Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:43:33.5742711Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:43:33.5750402Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:43:33.5757737Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:43:33.5765400Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:43:33.5768300Z (EngineCore_DP0 pid=5347) INFO 12-04 09:43:33 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:43:33.9113625Z (EngineCore_DP0 pid=5347) INFO 12-04 09:43:33 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:43:33.9204743Z (EngineCore_DP0 pid=5347) INFO 12-04 09:43:33 [gpu_model_runner.py:2840] Starting to load model ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11... 2025-12-04T09:43:34.2304840Z (EngineCore_DP0 pid=5347) INFO 12-04 09:43:34 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:43:34.3143497Z (EngineCore_DP0 pid=5347) 2025-12-04T09:43:34.3144201Z config.json: 0% 0.00/776 [00:00, model_arch='UltravoxModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'fixie-ai/ultravox-v0_5-llama-3_2-1b'} 2025-12-04T09:43:40.1621367Z 2025-12-04T09:43:40.1622544Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:43:40.1622877Z config.json: 2.15kB [00:00, 10.3MB/s] 2025-12-04T09:43:40.4154382Z 2025-12-04T09:43:40.4155137Z preprocessor_config.json: 0% 0.00/339 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:43:50.2813381Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:43:50.2842091Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:43:50.2849270Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:43:50.2856559Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:43:50.2865141Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:43:50.2872996Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:43:50.2875766Z (EngineCore_DP0 pid=5583) INFO 12-04 09:43:50 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:43:50.5736425Z (EngineCore_DP0 pid=5583) INFO 12-04 09:43:50 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:43:50.6906875Z (EngineCore_DP0 pid=5583) 2025-12-04T09:43:50.6907322Z processor_config.json: 0% 0.00/270 [00:00, model_arch='DeepSeekMTPModel', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'luccafong/deepseek_mtp_draft_random', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'luccafong/deepseek_mtp_main_random'} 2025-12-04T09:43:54.8407592Z 2025-12-04T09:43:54.8409641Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:43:54.8409964Z config.json: 1.95kB [00:00, 10.7MB/s] 2025-12-04T09:43:54.8870493Z INFO 12-04 09:43:54 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T09:44:01.2734586Z INFO 12-04 09:44:01 [model.py:653] Resolved architecture: DeepseekV3ForCausalLM 2025-12-04T09:44:01.2735106Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:44:01.2986596Z INFO 12-04 09:44:01 [model.py:1714] Using max model len 163840 2025-12-04T09:44:01.3929744Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T09:44:01.6596902Z 2025-12-04T09:44:01.6597696Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:44:01.6598080Z config.json: 1.95kB [00:00, 15.1MB/s] 2025-12-04T09:44:01.6888144Z INFO 12-04 09:44:01 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T09:44:08.0150612Z INFO 12-04 09:44:08 [model.py:653] Resolved architecture: DeepSeekMTPModel 2025-12-04T09:44:08.0152574Z INFO 12-04 09:44:08 [model.py:1714] Using max model len 163840 2025-12-04T09:44:08.0159996Z INFO 12-04 09:44:08 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:44:08.0883079Z 2025-12-04T09:44:08.0884334Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:44:08.0884790Z tokenizer_config.json: 3.13kB [00:00, 28.3MB/s] 2025-12-04T09:44:08.2136264Z 2025-12-04T09:44:08.2429992Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T09:44:08.2430341Z tokenizer.json: 7.85MB [00:00, 266MB/s] 2025-12-04T09:44:08.8202257Z INFO 12-04 09:44:08 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T09:44:08.8428514Z (EngineCore_DP0 pid=5740) INFO 12-04 09:44:08 [core.py:727] Waiting for init message from front-end. 2025-12-04T09:44:08.8470035Z (EngineCore_DP0 pid=5740) INFO 12-04 09:44:08 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='luccafong/deepseek_mtp_main_random', speculative_config=SpeculativeConfig(method='mtp', model='luccafong/deepseek_mtp_draft_random', num_spec_tokens=1), tokenizer='luccafong/deepseek_mtp_main_random', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=163840, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=fp8, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=luccafong/deepseek_mtp_main_random, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+quant_fp8'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:44:10.2222605Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:10.2249393Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:10.2256269Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:10.2263810Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:10.2271458Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:10.2279106Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:10.2281916Z (EngineCore_DP0 pid=5740) INFO 12-04 09:44:10 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:44:10.5085027Z (EngineCore_DP0 pid=5740) INFO 12-04 09:44:10 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:44:10.5231218Z (EngineCore_DP0 pid=5740) WARNING 12-04 09:44:10 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-12-04T09:44:10.5854735Z (EngineCore_DP0 pid=5740) INFO 12-04 09:44:10 [gpu_model_runner.py:2840] Starting to load model luccafong/deepseek_mtp_main_random... 2025-12-04T09:44:10.8607189Z (EngineCore_DP0 pid=5740) INFO 12-04 09:44:10 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:44:10.9399582Z (EngineCore_DP0 pid=5740) INFO 12-04 09:44:10 [cuda.py:328] Using Triton MLA backend on V1 engine. 2025-12-04T09:44:10.9609212Z (EngineCore_DP0 pid=5740) WARNING 12-04 09:44:10 [fp8.py:150] DeepGEMM backend requested but not available. 2025-12-04T09:44:10.9610288Z (EngineCore_DP0 pid=5740) INFO 12-04 09:44:10 [fp8.py:165] Using Triton backend for FP8 MoE 2025-12-04T09:44:11.0358453Z (EngineCore_DP0 pid=5740) WARNING 12-04 09:44:11 [fp8_utils.py:785] Using default W8A8 Block FP8 kernel config. Performance might be sub-optimal! Config file not found at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/model_executor/layers/quantization/utils/configs/N=8192,K=512,device_name=NVIDIA_L4,dtype=fp8_w8a8,block_shape=[128,128].json 2025-12-04T09:44:12.0724729Z (EngineCore_DP0 pid=5740) INFO 12-04 09:44:12 [gpu_model_runner.py:2879] Loading drafter model... 2025-12-04T09:44:12.0823539Z (EngineCore_DP0 pid=5740) INFO 12-04 09:44:12 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-12-04T09:44:12.0824523Z (EngineCore_DP0 pid=5740) INFO 12-04 09:44:12 [eagle.py:1027] Loading EAGLE LM head weights from the target model. 2025-12-04T09:44:12.6343135Z (EngineCore_DP0 pid=5740) INFO 12-04 09:44:12 [gpu_model_runner.py:2902] Model loading took 2.8305 GiB and 1.221949 seconds 2025-12-04T09:44:12.6346863Z (EngineCore_DP0 pid=5740) INFO 12-04 09:44:12 [kv_cache_utils.py:1199] GPU KV cache size: 4,660,336 tokens 2025-12-04T09:44:12.6347628Z (EngineCore_DP0 pid=5740) INFO 12-04 09:44:12 [kv_cache_utils.py:1204] Maximum concurrency for 163,840 tokens per request: 28.44x 2025-12-04T09:44:13.7929234Z (EngineCore_DP0 pid=5740) INFO 12-04 09:44:13 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:44:13.7947784Z INFO 12-04 09:44:13 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:44:14.2857044Z PASSED 2025-12-04T09:44:14.3008406Z models/test_initialization.py::test_can_initialize_small_subset[XLMRobertaModel] Fork a new process to run a test 5846 2025-12-04T09:44:14.3017335Z Fork a new process to run a test 0 2025-12-04T09:44:14.3287407Z INFO 12-04 09:44:14 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='XLMRobertaModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'intfloat/multilingual-e5-small'} 2025-12-04T09:44:14.4276263Z 2025-12-04T09:44:14.4277223Z config.json: 0% 0.00/655 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:44:24.2747967Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:24.2778363Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:24.2785558Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:24.2793012Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:24.2800400Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:24.2808012Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:24.2810798Z (EngineCore_DP0 pid=5926) INFO 12-04 09:44:24 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:44:24.5603980Z (EngineCore_DP0 pid=5926) INFO 12-04 09:44:24 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:44:24.5718255Z (EngineCore_DP0 pid=5926) INFO 12-04 09:44:24 [gpu_model_runner.py:2840] Starting to load model intfloat/multilingual-e5-small... 2025-12-04T09:44:24.8338091Z (EngineCore_DP0 pid=5926) INFO 12-04 09:44:24 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:44:24.8383027Z (EngineCore_DP0 pid=5926) INFO 12-04 09:44:24 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:44:25.4333774Z (EngineCore_DP0 pid=5926) INFO 12-04 09:44:25 [gpu_model_runner.py:2902] Model loading took 0.1834 GiB and 0.065493 seconds 2025-12-04T09:44:26.5215958Z (EngineCore_DP0 pid=5926) INFO 12-04 09:44:26 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T09:44:26.5270392Z (EngineCore_DP0 pid=5926) INFO 12-04 09:44:26 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T09:44:26.8053224Z (EngineCore_DP0 pid=5926) INFO 12-04 09:44:26 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:44:26.8068000Z INFO 12-04 09:44:26 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-12-04T09:44:27.2845156Z PASSED 2025-12-04T09:44:27.2995379Z models/test_initialization.py::test_can_initialize_large_subset[Starcoder2ForCausalLM] Fork a new process to run a test 5996 2025-12-04T09:44:27.3004242Z Fork a new process to run a test 0 2025-12-04T09:44:27.3281697Z INFO 12-04 09:44:27 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Starcoder2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'bigcode/starcoder2-3b'} 2025-12-04T09:44:27.4047406Z 2025-12-04T09:44:27.4048421Z config.json: 0% 0.00/700 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:44:35.9466668Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:35.9493752Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:35.9501157Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:35.9509198Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:35.9516382Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:35.9524280Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:35.9526911Z (EngineCore_DP0 pid=6054) INFO 12-04 09:44:35 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:44:36.2307253Z (EngineCore_DP0 pid=6054) INFO 12-04 09:44:36 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:44:36.2958578Z (EngineCore_DP0 pid=6054) INFO 12-04 09:44:36 [gpu_model_runner.py:2840] Starting to load model bigcode/starcoder2-3b... 2025-12-04T09:44:36.5689226Z (EngineCore_DP0 pid=6054) INFO 12-04 09:44:36 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:44:36.6336983Z (EngineCore_DP0 pid=6054) INFO 12-04 09:44:36 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:44:37.2243440Z (EngineCore_DP0 pid=6054) INFO 12-04 09:44:37 [gpu_model_runner.py:2902] Model loading took 0.4649 GiB and 0.124059 seconds 2025-12-04T09:44:37.2247993Z (EngineCore_DP0 pid=6054) INFO 12-04 09:44:37 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T09:44:37.2248916Z (EngineCore_DP0 pid=6054) INFO 12-04 09:44:37 [kv_cache_utils.py:1204] Maximum concurrency for 16,384 tokens per request: 852.22x 2025-12-04T09:44:38.6111320Z (EngineCore_DP0 pid=6054) INFO 12-04 09:44:38 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:44:38.6125630Z INFO 12-04 09:44:38 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:44:39.0689263Z PASSED 2025-12-04T09:44:39.0839087Z models/test_initialization.py::test_can_initialize_large_subset[Gemma3TextModel] Fork a new process to run a test 6124 2025-12-04T09:44:39.0847178Z Fork a new process to run a test 0 2025-12-04T09:44:39.1118577Z INFO 12-04 09:44:39 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Gemma3TextModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'google/embeddinggemma-300m'} 2025-12-04T09:44:39.2143359Z 2025-12-04T09:44:39.2144284Z config.json: 0% 0.00/1.49k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:44:49.7416420Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:49.7447195Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:49.7454206Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:49.7462077Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:49.7469827Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:49.7477478Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:44:49.7480224Z (EngineCore_DP0 pid=6203) INFO 12-04 09:44:49 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:44:50.0274251Z (EngineCore_DP0 pid=6203) INFO 12-04 09:44:50 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:44:50.0472072Z (EngineCore_DP0 pid=6203) INFO 12-04 09:44:50 [gpu_model_runner.py:2840] Starting to load model google/embeddinggemma-300m... 2025-12-04T09:44:50.3118933Z (EngineCore_DP0 pid=6203) INFO 12-04 09:44:50 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:44:50.3989626Z (EngineCore_DP0 pid=6203) INFO 12-04 09:44:50 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:44:50.5378525Z (EngineCore_DP0 pid=6203) 2025-12-04T09:44:50.5378891Z config.json: 0% 0.00/134 [00:00, model_arch='Step3VLForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'stepfun-ai/step3'} 2025-12-04T09:44:54.1066558Z 2025-12-04T09:44:54.1068990Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:44:54.1069413Z config.json: 1.54kB [00:00, 8.07MB/s] 2025-12-04T09:45:00.4661587Z INFO 12-04 09:45:00 [model.py:653] Resolved architecture: Step3VLForConditionalGeneration 2025-12-04T09:45:00.4662191Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:45:00.4913133Z INFO 12-04 09:45:00 [model.py:1714] Using max model len 65536 2025-12-04T09:45:00.4916133Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T09:45:00.6592212Z INFO 12-04 09:45:00 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:45:00.6897322Z 2025-12-04T09:45:00.6898588Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:45:00.6898989Z tokenizer_config.json: 3.94kB [00:00, 33.7MB/s] 2025-12-04T09:45:00.7932355Z 2025-12-04T09:45:00.8214198Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T09:45:00.8214643Z tokenizer.json: 7.85MB [00:00, 277MB/s] 2025-12-04T09:45:01.3228747Z 2025-12-04T09:45:01.3229227Z generation_config.json: 0% 0.00/158 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:45:03.1048224Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:03.1076376Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:03.1083797Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:03.1091901Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:03.1100354Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:03.1108845Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:03.1111082Z (EngineCore_DP0 pid=6352) INFO 12-04 09:45:03 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:45:03.3949878Z (EngineCore_DP0 pid=6352) INFO 12-04 09:45:03 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:45:03.5009467Z (EngineCore_DP0 pid=6352) INFO 12-04 09:45:03 [gpu_model_runner.py:2840] Starting to load model stepfun-ai/step3... 2025-12-04T09:45:03.7876893Z (EngineCore_DP0 pid=6352) INFO 12-04 09:45:03 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:45:03.8049373Z (EngineCore_DP0 pid=6352) INFO 12-04 09:45:03 [layer.py:497] MultiHeadAttention attn_backend: _Backend.XFORMERS, use_upstream_fa: False 2025-12-04T09:45:03.8717405Z (EngineCore_DP0 pid=6352) INFO 12-04 09:45:03 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:45:04.4838309Z (EngineCore_DP0 pid=6352) INFO 12-04 09:45:04 [gpu_model_runner.py:2902] Model loading took 5.3915 GiB and 0.143263 seconds 2025-12-04T09:45:04.4843366Z (EngineCore_DP0 pid=6352) INFO 12-04 09:45:04 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T09:45:04.4844543Z (EngineCore_DP0 pid=6352) INFO 12-04 09:45:04 [kv_cache_utils.py:1204] Maximum concurrency for 65,536 tokens per request: 160.00x 2025-12-04T09:45:06.1867159Z (EngineCore_DP0 pid=6352) INFO 12-04 09:45:06 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:45:06.1881689Z INFO 12-04 09:45:06 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:45:06.6836369Z PASSED 2025-12-04T09:45:06.6986861Z models/test_initialization.py::test_can_initialize_large_subset[TeleFLMForCausalLM] Fork a new process to run a test 6422 2025-12-04T09:45:06.6995561Z Fork a new process to run a test 0 2025-12-04T09:45:06.7266941Z INFO 12-04 09:45:06 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='TeleFLMForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'CofeAI/FLM-2-52B-Instruct-2407'} 2025-12-04T09:45:06.8219470Z 2025-12-04T09:45:06.8220454Z config.json: 0% 0.00/940 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:45:16.2275257Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:16.2302758Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:16.2309629Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:16.2316959Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:16.2324602Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:16.2332395Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:16.2335088Z (EngineCore_DP0 pid=6500) INFO 12-04 09:45:16 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:45:16.5182209Z (EngineCore_DP0 pid=6500) INFO 12-04 09:45:16 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:45:16.5938008Z (EngineCore_DP0 pid=6500) INFO 12-04 09:45:16 [gpu_model_runner.py:2840] Starting to load model CofeAI/FLM-2-52B-Instruct-2407... 2025-12-04T09:45:16.8737054Z (EngineCore_DP0 pid=6500) INFO 12-04 09:45:16 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:45:16.9389638Z (EngineCore_DP0 pid=6500) INFO 12-04 09:45:16 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:45:17.6009207Z (EngineCore_DP0 pid=6500) INFO 12-04 09:45:17 [gpu_model_runner.py:2902] Model loading took 2.7349 GiB and 0.173831 seconds 2025-12-04T09:45:17.6013020Z (EngineCore_DP0 pid=6500) INFO 12-04 09:45:17 [kv_cache_utils.py:1199] GPU KV cache size: 327,680 tokens 2025-12-04T09:45:17.6013920Z (EngineCore_DP0 pid=6500) INFO 12-04 09:45:17 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 160.00x 2025-12-04T09:45:17.8534267Z (EngineCore_DP0 pid=6500) WARNING 12-04 09:45:17 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-12-04T09:45:18.3538271Z (EngineCore_DP0 pid=6500) INFO 12-04 09:45:18 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:45:18.3555881Z INFO 12-04 09:45:18 [llm.py:337] Supported tasks: ['embed'] 2025-12-04T09:45:18.8269328Z PASSED 2025-12-04T09:45:18.8420035Z models/test_initialization.py::test_can_initialize_large_subset[DeepseekV3ForCausalLM] Fork a new process to run a test 6570 2025-12-04T09:45:18.8429187Z Fork a new process to run a test 0 2025-12-04T09:45:18.8708652Z INFO 12-04 09:45:18 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='DeepseekV3ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'deepseek-ai/DeepSeek-V3'} 2025-12-04T09:45:18.9409021Z 2025-12-04T09:45:18.9410843Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:45:18.9411159Z config.json: 1.66kB [00:00, 10.3MB/s] 2025-12-04T09:45:18.9805907Z INFO 12-04 09:45:18 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T09:45:19.0358655Z INFO 12-04 09:45:19 [model.py:653] Resolved architecture: DeepseekV3ForCausalLM 2025-12-04T09:45:19.0359185Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:45:19.0605985Z INFO 12-04 09:45:19 [model.py:1714] Using max model len 163840 2025-12-04T09:45:19.1548378Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T09:45:19.3148081Z INFO 12-04 09:45:19 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:45:19.3533818Z 2025-12-04T09:45:19.3535211Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:45:19.3535600Z tokenizer_config.json: 3.13kB [00:00, 26.7MB/s] 2025-12-04T09:45:19.4576233Z 2025-12-04T09:45:19.4856853Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T09:45:19.4857196Z tokenizer.json: 7.85MB [00:00, 280MB/s] 2025-12-04T09:45:20.0236490Z INFO 12-04 09:45:20 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T09:45:20.0463453Z (EngineCore_DP0 pid=6578) INFO 12-04 09:45:20 [core.py:727] Waiting for init message from front-end. 2025-12-04T09:45:20.0522606Z (EngineCore_DP0 pid=6578) INFO 12-04 09:45:20 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='deepseek-ai/DeepSeek-V3', speculative_config=None, tokenizer='deepseek-ai/DeepSeek-V3', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=163840, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=fp8, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=deepseek-ai/DeepSeek-V3, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+quant_fp8'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:45:21.3430645Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:21.3457930Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:21.3465156Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:21.3473158Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:21.3481230Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:21.3489388Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:21.3492638Z (EngineCore_DP0 pid=6578) INFO 12-04 09:45:21 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:45:21.6346818Z (EngineCore_DP0 pid=6578) INFO 12-04 09:45:21 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:45:21.7532540Z (EngineCore_DP0 pid=6578) INFO 12-04 09:45:21 [gpu_model_runner.py:2840] Starting to load model deepseek-ai/DeepSeek-V3... 2025-12-04T09:45:22.0327539Z (EngineCore_DP0 pid=6578) INFO 12-04 09:45:22 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:45:22.1222988Z (EngineCore_DP0 pid=6578) INFO 12-04 09:45:22 [cuda.py:328] Using Triton MLA backend on V1 engine. 2025-12-04T09:45:22.1377124Z (EngineCore_DP0 pid=6578) WARNING 12-04 09:45:22 [fp8.py:150] DeepGEMM backend requested but not available. 2025-12-04T09:45:22.1377873Z (EngineCore_DP0 pid=6578) INFO 12-04 09:45:22 [fp8.py:165] Using Triton backend for FP8 MoE 2025-12-04T09:45:22.2588396Z (EngineCore_DP0 pid=6578) WARNING 12-04 09:45:22 [fp8_utils.py:785] Using default W8A8 Block FP8 kernel config. Performance might be sub-optimal! Config file not found at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/model_executor/layers/quantization/utils/configs/N=32768,K=512,device_name=NVIDIA_L4,dtype=fp8_w8a8,block_shape=[128,128].json 2025-12-04T09:45:23.3349319Z (EngineCore_DP0 pid=6578) INFO 12-04 09:45:23 [gpu_model_runner.py:2902] Model loading took 4.3759 GiB and 0.724398 seconds 2025-12-04T09:45:23.3353797Z (EngineCore_DP0 pid=6578) INFO 12-04 09:45:23 [kv_cache_utils.py:1199] GPU KV cache size: 9,320,672 tokens 2025-12-04T09:45:23.3355131Z (EngineCore_DP0 pid=6578) INFO 12-04 09:45:23 [kv_cache_utils.py:1204] Maximum concurrency for 163,840 tokens per request: 56.89x 2025-12-04T09:45:24.9890457Z (EngineCore_DP0 pid=6578) INFO 12-04 09:45:24 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:45:24.9904609Z INFO 12-04 09:45:24 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:45:25.5022302Z PASSED 2025-12-04T09:45:25.5184942Z models/test_initialization.py::test_can_initialize_large_subset[Phi4MultimodalForCausalLM] Fork a new process to run a test 6652 2025-12-04T09:45:25.5193909Z Fork a new process to run a test 0 2025-12-04T09:45:25.5461228Z INFO 12-04 09:45:25 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'revision': 'refs/pr/70', 'hf_overrides': functools.partial(, model_arch='Phi4MultimodalForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'microsoft/Phi-4-multimodal-instruct'} 2025-12-04T09:45:25.6889974Z 2025-12-04T09:45:25.6892185Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:45:25.6892517Z config.json: 3.71kB [00:00, 22.3MB/s] 2025-12-04T09:45:25.7210089Z INFO 12-04 09:45:25 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T09:45:25.7998246Z 2025-12-04T09:45:25.7999391Z preprocessor_config.json: 0% 0.00/705 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:45:35.7407712Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:35.7436368Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:35.7443967Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:35.7452034Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:35.7459540Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:35.7467093Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:35.7470573Z (EngineCore_DP0 pid=6730) INFO 12-04 09:45:35 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:45:36.0280222Z (EngineCore_DP0 pid=6730) INFO 12-04 09:45:36 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:45:39.3215809Z (EngineCore_DP0 pid=6730) It is strongly recommended to pass the `sampling_rate` argument to `Phi4MultimodalFeatureExtractor()`. Failing to do so can result in silent errors that might be hard to debug. 2025-12-04T09:45:39.5300849Z (EngineCore_DP0 pid=6730) INFO 12-04 09:45:39 [gpu_model_runner.py:2840] Starting to load model microsoft/Phi-4-multimodal-instruct... 2025-12-04T09:45:39.8041820Z (EngineCore_DP0 pid=6730) INFO 12-04 09:45:39 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:45:39.8942715Z (EngineCore_DP0 pid=6730) INFO 12-04 09:45:39 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:45:40.5142973Z (EngineCore_DP0 pid=6730) INFO 12-04 09:45:40 [gpu_model_runner.py:2902] Model loading took 2.2617 GiB and 0.162682 seconds 2025-12-04T09:45:40.5147305Z (EngineCore_DP0 pid=6730) INFO 12-04 09:45:40 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T09:45:40.5148652Z (EngineCore_DP0 pid=6730) INFO 12-04 09:45:40 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 20.00x 2025-12-04T09:45:41.2886784Z (EngineCore_DP0 pid=6730) It is strongly recommended to pass the `sampling_rate` argument to `Phi4MultimodalFeatureExtractor()`. Failing to do so can result in silent errors that might be hard to debug. 2025-12-04T09:45:41.7036046Z (EngineCore_DP0 pid=6730) INFO 12-04 09:45:41 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:45:41.7084038Z INFO 12-04 09:45:41 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:45:42.2070867Z PASSED 2025-12-04T09:45:42.2218607Z models/test_initialization.py::test_can_initialize_large_subset[InternLMForCausalLM] Fork a new process to run a test 6816 2025-12-04T09:45:42.2228604Z Fork a new process to run a test 0 2025-12-04T09:45:42.2508736Z INFO 12-04 09:45:42 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='InternLMForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'internlm/internlm-chat-7b'} 2025-12-04T09:45:42.3372473Z 2025-12-04T09:45:42.3374193Z config.json: 0% 0.00/731 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:45:45.0693458Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:45.0720762Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:45.0728044Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:45.0735089Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:45.0742581Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:45.0749809Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:45.0752542Z (EngineCore_DP0 pid=6845) INFO 12-04 09:45:45 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:45:45.3589736Z (EngineCore_DP0 pid=6845) INFO 12-04 09:45:45 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:45:45.4245624Z (EngineCore_DP0 pid=6845) INFO 12-04 09:45:45 [gpu_model_runner.py:2840] Starting to load model internlm/internlm-chat-7b... 2025-12-04T09:45:45.7103383Z (EngineCore_DP0 pid=6845) INFO 12-04 09:45:45 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:45:45.7755766Z (EngineCore_DP0 pid=6845) INFO 12-04 09:45:45 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:45:46.3902791Z (EngineCore_DP0 pid=6845) INFO 12-04 09:45:46 [gpu_model_runner.py:2902] Model loading took 1.9517 GiB and 0.123587 seconds 2025-12-04T09:45:46.3907068Z (EngineCore_DP0 pid=6845) INFO 12-04 09:45:46 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-12-04T09:45:46.3908233Z (EngineCore_DP0 pid=6845) INFO 12-04 09:45:46 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 320.00x 2025-12-04T09:45:46.7038345Z (EngineCore_DP0 pid=6845) WARNING 12-04 09:45:46 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-12-04T09:45:47.1779409Z (EngineCore_DP0 pid=6845) INFO 12-04 09:45:47 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:45:47.1799053Z INFO 12-04 09:45:47 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:45:47.6210539Z PASSED 2025-12-04T09:45:47.6358288Z models/test_initialization.py::test_can_initialize_large_subset[StableLMEpochForCausalLM] Fork a new process to run a test 6915 2025-12-04T09:45:47.6366978Z Fork a new process to run a test 0 2025-12-04T09:45:47.6645062Z INFO 12-04 09:45:47 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='StableLMEpochForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'stabilityai/stablelm-zephyr-3b'} 2025-12-04T09:45:47.8066720Z 2025-12-04T09:45:47.8068440Z config.json: 0% 0.00/599 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:45:56.4460949Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:56.4488374Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:56.4495176Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:56.4502513Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:56.4509505Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:56.4516635Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:45:56.4519320Z (EngineCore_DP0 pid=6973) INFO 12-04 09:45:56 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:45:56.7361592Z (EngineCore_DP0 pid=6973) INFO 12-04 09:45:56 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:45:56.7993329Z (EngineCore_DP0 pid=6973) INFO 12-04 09:45:56 [gpu_model_runner.py:2840] Starting to load model stabilityai/stablelm-zephyr-3b... 2025-12-04T09:45:57.0804310Z (EngineCore_DP0 pid=6973) INFO 12-04 09:45:57 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:45:57.1449339Z (EngineCore_DP0 pid=6973) INFO 12-04 09:45:57 [cuda.py:420] Using FlexAttention backend for head_size=80 on V1 engine. 2025-12-04T09:45:57.2488466Z (EngineCore_DP0 pid=6973) WARNING 12-04 09:45:57 [vllm.py:821] `torch.compile` is turned on, but the model stabilityai/stablelm-zephyr-3b does not support it. Please open an issue on GitHub if you want it to be supported. 2025-12-04T09:45:57.7956734Z (EngineCore_DP0 pid=6973) INFO 12-04 09:45:57 [gpu_model_runner.py:2902] Model loading took 0.6296 GiB and 0.170869 seconds 2025-12-04T09:45:57.7961132Z (EngineCore_DP0 pid=6973) INFO 12-04 09:45:57 [kv_cache_utils.py:1199] GPU KV cache size: 1,048,576 tokens 2025-12-04T09:45:57.7962073Z (EngineCore_DP0 pid=6973) INFO 12-04 09:45:57 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 256.00x 2025-12-04T09:45:58.2663659Z (EngineCore_DP0 pid=6973) INFO 12-04 09:45:58 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:45:58.2684675Z INFO 12-04 09:45:58 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:45:58.7043386Z PASSED 2025-12-04T09:45:58.7191064Z models/test_initialization.py::test_can_initialize_large_subset[CohereForCausalLM] Fork a new process to run a test 7043 2025-12-04T09:45:58.7200362Z Fork a new process to run a test 0 2025-12-04T09:45:58.7474894Z INFO 12-04 09:45:58 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='CohereForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'CohereForAI/c4ai-command-r-v01'} 2025-12-04T09:45:58.9277987Z 2025-12-04T09:45:58.9278636Z config.json: 0% 0.00/765 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:46:08.5083621Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:08.5110433Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:08.5117774Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:08.5125202Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:08.5132382Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:08.5140257Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:08.5143139Z (EngineCore_DP0 pid=7121) INFO 12-04 09:46:08 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:46:08.7932255Z (EngineCore_DP0 pid=7121) INFO 12-04 09:46:08 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:46:08.9234029Z (EngineCore_DP0 pid=7121) INFO 12-04 09:46:08 [gpu_model_runner.py:2840] Starting to load model CohereForAI/c4ai-command-r-v01... 2025-12-04T09:46:09.1961511Z (EngineCore_DP0 pid=7121) INFO 12-04 09:46:09 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:46:09.2672342Z (EngineCore_DP0 pid=7121) INFO 12-04 09:46:09 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:46:09.8525117Z (EngineCore_DP0 pid=7121) INFO 12-04 09:46:09 [gpu_model_runner.py:2902] Model loading took 5.4688 GiB and 0.118708 seconds 2025-12-04T09:46:09.8529739Z (EngineCore_DP0 pid=7121) INFO 12-04 09:46:09 [kv_cache_utils.py:1199] GPU KV cache size: 327,680 tokens 2025-12-04T09:46:09.8530642Z (EngineCore_DP0 pid=7121) INFO 12-04 09:46:09 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 2.50x 2025-12-04T09:46:10.9298994Z (EngineCore_DP0 pid=7121) INFO 12-04 09:46:10 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:46:10.9314309Z INFO 12-04 09:46:10 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:46:11.4314541Z PASSED 2025-12-04T09:46:11.4462625Z models/test_initialization.py::test_can_initialize_large_subset[BambaForCausalLM] Fork a new process to run a test 7191 2025-12-04T09:46:11.4472514Z Fork a new process to run a test 0 2025-12-04T09:46:11.4743267Z INFO 12-04 09:46:11 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='BambaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ibm-ai-platform/Bamba-9B-v1'} 2025-12-04T09:46:11.5667293Z 2025-12-04T09:46:11.5668317Z config.json: 0% 0.00/921 [00:00= mamba page size. 2025-12-04T09:46:18.4722858Z INFO 12-04 09:46:18 [config.py:453] Padding mamba page size by 0.69% to ensure that mamba page size and attention page size are exactly equal. 2025-12-04T09:46:18.5118959Z 2025-12-04T09:46:18.5122234Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:46:18.5122625Z tokenizer_config.json: 50.6kB [00:00, 178MB/s] 2025-12-04T09:46:18.5928009Z 2025-12-04T09:46:18.6303640Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T09:46:18.6303973Z tokenizer.json: 9.09MB [00:00, 242MB/s] 2025-12-04T09:46:18.8200690Z 2025-12-04T09:46:18.8202185Z special_tokens_map.json: 0% 0.00/301 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:46:20.6336730Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:20.6364080Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:20.6371693Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:20.6379409Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:20.6387554Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:20.6394858Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:20.6397647Z (EngineCore_DP0 pid=7253) INFO 12-04 09:46:20 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:46:20.9175431Z (EngineCore_DP0 pid=7253) INFO 12-04 09:46:20 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:46:20.9812247Z (EngineCore_DP0 pid=7253) INFO 12-04 09:46:20 [gpu_model_runner.py:2840] Starting to load model ibm-ai-platform/Bamba-9B-v1... 2025-12-04T09:46:21.2530518Z (EngineCore_DP0 pid=7253) INFO 12-04 09:46:21 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:46:21.8519448Z (EngineCore_DP0 pid=7253) INFO 12-04 09:46:21 [gpu_model_runner.py:2902] Model loading took 2.4767 GiB and 0.061686 seconds 2025-12-04T09:46:21.8524983Z (EngineCore_DP0 pid=7253) INFO 12-04 09:46:21 [kv_cache_utils.py:1199] GPU KV cache size: 20,332,544 tokens 2025-12-04T09:46:21.8525919Z (EngineCore_DP0 pid=7253) INFO 12-04 09:46:21 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 4964.00x 2025-12-04T09:46:22.6269072Z (EngineCore_DP0 pid=7253) INFO 12-04 09:46:22 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:46:22.6289614Z INFO 12-04 09:46:22 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:46:23.0717926Z PASSED 2025-12-04T09:46:23.0866219Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2_5_VLForConditionalGeneration] Fork a new process to run a test 7323 2025-12-04T09:46:23.0875564Z Fork a new process to run a test 0 2025-12-04T09:46:23.1143489Z INFO 12-04 09:46:23 [utils.py:239] non-default args: {'load_format': 'dummy', 'max_model_len': 4096, 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2_5_VLForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen2.5-VL-3B-Instruct'} 2025-12-04T09:46:23.1939415Z 2025-12-04T09:46:23.1941374Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:46:23.1941681Z config.json: 1.37kB [00:00, 7.92MB/s] 2025-12-04T09:46:23.2829900Z 2025-12-04T09:46:23.2831251Z preprocessor_config.json: 0% 0.00/350 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:46:32.1376153Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:32.1415719Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:32.1422988Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:32.1430188Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:32.1438132Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:32.1445484Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:32.1448293Z (EngineCore_DP0 pid=7381) INFO 12-04 09:46:32 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:46:32.4237975Z (EngineCore_DP0 pid=7381) INFO 12-04 09:46:32 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:46:33.8446006Z (EngineCore_DP0 pid=7381) 2025-12-04T09:46:33.8447072Z chat_template.json: 0.00B [00:00, ?B/s] 2025-12-04T09:46:33.8447399Z chat_template.json: 1.05kB [00:00, 3.63MB/s] 2025-12-04T09:46:34.3912740Z (EngineCore_DP0 pid=7381) INFO 12-04 09:46:34 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen2.5-VL-3B-Instruct... 2025-12-04T09:46:34.7393411Z (EngineCore_DP0 pid=7381) INFO 12-04 09:46:34 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:46:34.8053452Z (EngineCore_DP0 pid=7381) INFO 12-04 09:46:34 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:46:35.4122154Z (EngineCore_DP0 pid=7381) INFO 12-04 09:46:35 [gpu_model_runner.py:2902] Model loading took 2.0974 GiB and 0.131455 seconds 2025-12-04T09:46:35.4126296Z (EngineCore_DP0 pid=7381) INFO 12-04 09:46:35 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T09:46:35.4127676Z (EngineCore_DP0 pid=7381) INFO 12-04 09:46:35 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 2560.00x 2025-12-04T09:46:37.0014764Z (EngineCore_DP0 pid=7381) INFO 12-04 09:46:37 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:46:37.0029535Z INFO 12-04 09:46:37 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:46:37.4699370Z PASSED 2025-12-04T09:46:37.4846124Z models/test_initialization.py::test_can_initialize_large_subset[TransformersEmbeddingModel] Fork a new process to run a test 7451 2025-12-04T09:46:37.4855501Z Fork a new process to run a test 0 2025-12-04T09:46:37.4859386Z `transformers==4.56.2` installed, but `transformers>=4.57.0.dev0` is required to run this model. 2025-12-04T09:46:37.7644902Z PASSED 2025-12-04T09:46:37.7793649Z models/test_initialization.py::test_can_initialize_large_subset[Gemma3ForConditionalGeneration] Fork a new process to run a test 7452 2025-12-04T09:46:37.7802541Z Fork a new process to run a test 0 2025-12-04T09:46:37.8080367Z INFO 12-04 09:46:37 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Gemma3ForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'google/gemma-3-4b-it'} 2025-12-04T09:46:37.9300068Z 2025-12-04T09:46:37.9302881Z config.json: 0% 0.00/855 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:46:49.8903648Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:49.8932399Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:49.8939836Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:49.8948559Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:49.8956075Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:49.8963181Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:46:49.8965962Z (EngineCore_DP0 pid=7531) INFO 12-04 09:46:49 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:46:50.1819453Z (EngineCore_DP0 pid=7531) INFO 12-04 09:46:50 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:46:50.3705976Z (EngineCore_DP0 pid=7531) 2025-12-04T09:46:50.3706779Z processor_config.json: 0% 0.00/70.0 [00:00, model_arch='InternLM2VEForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'OpenGVLab/Mono-InternVL-2B'} 2025-12-04T09:47:00.2074322Z 2025-12-04T09:47:00.2076294Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:47:00.2700707Z config.json: 3.75kB [00:00, 21.2MB/s] 2025-12-04T09:47:00.2701167Z 2025-12-04T09:47:00.2703186Z configuration_internvl_chat.py: 0.00B [00:00, ?B/s] 2025-12-04T09:47:00.2704013Z configuration_internvl_chat.py: 4.44kB [00:00, 18.9MB/s] 2025-12-04T09:47:00.3269910Z 2025-12-04T09:47:00.3272634Z configuration_intern_patch.py: 0.00B [00:00, ?B/s] 2025-12-04T09:47:00.3272971Z configuration_intern_patch.py: 4.40kB [00:00, 22.9MB/s] 2025-12-04T09:47:00.3291948Z A new version of the following files was downloaded from https://huggingface.co/OpenGVLab/Mono-InternVL-2B: 2025-12-04T09:47:00.3292533Z - configuration_intern_patch.py 2025-12-04T09:47:00.3293181Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:47:00.3665859Z 2025-12-04T09:47:00.3667324Z configuration_internlm2.py: 0.00B [00:00, ?B/s] 2025-12-04T09:47:00.3689293Z configuration_internlm2.py: 7.00kB [00:00, 66.9MB/s] 2025-12-04T09:47:00.3690051Z A new version of the following files was downloaded from https://huggingface.co/OpenGVLab/Mono-InternVL-2B: 2025-12-04T09:47:00.3690594Z - configuration_internlm2.py 2025-12-04T09:47:00.3691256Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:47:00.3692124Z A new version of the following files was downloaded from https://huggingface.co/OpenGVLab/Mono-InternVL-2B: 2025-12-04T09:47:00.3692656Z - configuration_internvl_chat.py 2025-12-04T09:47:00.3692924Z - configuration_intern_patch.py 2025-12-04T09:47:00.3693187Z - configuration_internlm2.py 2025-12-04T09:47:00.3693800Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:47:00.5747294Z INFO 12-04 09:47:00 [model.py:653] Resolved architecture: InternVLChatModel 2025-12-04T09:47:00.5748105Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:47:00.5988044Z INFO 12-04 09:47:00 [model.py:1714] Using max model len 32768 2025-12-04T09:47:00.5990472Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T09:47:00.6545075Z INFO 12-04 09:47:00 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:47:00.6906713Z 2025-12-04T09:47:00.6908292Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:47:00.6908694Z tokenizer_config.json: 4.00kB [00:00, 37.3MB/s] 2025-12-04T09:47:00.7324343Z 2025-12-04T09:47:00.7326277Z tokenization_internlm2.py: 0.00B [00:00, ?B/s] 2025-12-04T09:47:00.7326745Z tokenization_internlm2.py: 8.79kB [00:00, 63.1MB/s] 2025-12-04T09:47:00.7376687Z A new version of the following files was downloaded from https://huggingface.co/OpenGVLab/Mono-InternVL-2B: 2025-12-04T09:47:00.7377235Z - tokenization_internlm2.py 2025-12-04T09:47:00.7377856Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:47:00.8184285Z 2025-12-04T09:47:01.0463440Z ./tokenizer.model: 0% 0.00/1.48M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:47:03.3163156Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:03.3191859Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:03.3198768Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:03.3206014Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:03.3213356Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:03.3220707Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:03.3223518Z (EngineCore_DP0 pid=7645) INFO 12-04 09:47:03 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:47:03.6015232Z (EngineCore_DP0 pid=7645) INFO 12-04 09:47:03 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:47:03.7019930Z (EngineCore_DP0 pid=7645) WARNING 12-04 09:47:03 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T09:47:03.7027326Z (EngineCore_DP0 pid=7645) WARNING 12-04 09:47:03 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T09:47:03.8908735Z (EngineCore_DP0 pid=7645) INFO 12-04 09:47:03 [gpu_model_runner.py:2840] Starting to load model OpenGVLab/Mono-InternVL-2B... 2025-12-04T09:47:04.1653121Z (EngineCore_DP0 pid=7645) INFO 12-04 09:47:04 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:47:04.2230007Z (EngineCore_DP0 pid=7645) INFO 12-04 09:47:04 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:47:04.8196876Z (EngineCore_DP0 pid=7645) INFO 12-04 09:47:04 [gpu_model_runner.py:2902] Model loading took 0.9523 GiB and 0.117021 seconds 2025-12-04T09:47:04.8201147Z (EngineCore_DP0 pid=7645) INFO 12-04 09:47:04 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T09:47:04.8202373Z (EngineCore_DP0 pid=7645) INFO 12-04 09:47:04 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 80.00x 2025-12-04T09:47:05.1002251Z (EngineCore_DP0 pid=7645) WARNING 12-04 09:47:05 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-12-04T09:47:05.1998015Z (EngineCore_DP0 pid=7645) WARNING 12-04 09:47:05 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T09:47:05.7436091Z (EngineCore_DP0 pid=7645) INFO 12-04 09:47:05 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:47:05.7449883Z INFO 12-04 09:47:05 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:47:06.1949202Z PASSED 2025-12-04T09:47:06.2102051Z models/test_initialization.py::test_can_initialize_large_subset[FalconForCausalLM] Fork a new process to run a test 7715 2025-12-04T09:47:06.2111640Z Fork a new process to run a test 0 2025-12-04T09:47:06.2375647Z INFO 12-04 09:47:06 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='FalconForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'tiiuae/falcon-7b'} 2025-12-04T09:47:06.3096629Z 2025-12-04T09:47:06.3098882Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:47:06.3099297Z config.json: 1.05kB [00:00, 5.04MB/s] 2025-12-04T09:47:12.7235184Z INFO 12-04 09:47:12 [model.py:653] Resolved architecture: FalconForCausalLM 2025-12-04T09:47:12.7235591Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:47:12.7487007Z INFO 12-04 09:47:12 [model.py:1714] Using max model len 2048 2025-12-04T09:47:12.9350626Z INFO 12-04 09:47:12 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:47:12.9655094Z 2025-12-04T09:47:12.9656180Z tokenizer_config.json: 0% 0.00/287 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:47:14.7058039Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:14.7085448Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:14.7092513Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:14.7099591Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:14.7107205Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:14.7114577Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:14.7117421Z (EngineCore_DP0 pid=7773) INFO 12-04 09:47:14 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:47:14.9904642Z (EngineCore_DP0 pid=7773) INFO 12-04 09:47:14 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:47:15.0978295Z (EngineCore_DP0 pid=7773) INFO 12-04 09:47:15 [gpu_model_runner.py:2840] Starting to load model tiiuae/falcon-7b... 2025-12-04T09:47:15.3723747Z (EngineCore_DP0 pid=7773) INFO 12-04 09:47:15 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:47:15.4363208Z (EngineCore_DP0 pid=7773) INFO 12-04 09:47:15 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:47:16.0294815Z (EngineCore_DP0 pid=7773) INFO 12-04 09:47:16 [gpu_model_runner.py:2902] Model loading took 0.9383 GiB and 0.123074 seconds 2025-12-04T09:47:16.0299275Z (EngineCore_DP0 pid=7773) INFO 12-04 09:47:16 [kv_cache_utils.py:1199] GPU KV cache size: 41,943,040 tokens 2025-12-04T09:47:16.0300606Z (EngineCore_DP0 pid=7773) INFO 12-04 09:47:16 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 20480.00x 2025-12-04T09:47:20.0987920Z (EngineCore_DP0 pid=7773) INFO 12-04 09:47:20 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:47:20.1002784Z INFO 12-04 09:47:20 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:47:20.5904709Z PASSED 2025-12-04T09:47:20.6054150Z models/test_initialization.py::test_can_initialize_large_subset[AriaForConditionalGeneration] Fork a new process to run a test 7843 2025-12-04T09:47:20.6064215Z Fork a new process to run a test 0 2025-12-04T09:47:20.6342955Z INFO 12-04 09:47:20 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='AriaForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'rhymes-ai/Aria'} 2025-12-04T09:47:20.7071732Z 2025-12-04T09:47:20.7073577Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:47:20.7073972Z config.json: 1.08kB [00:00, 5.98MB/s] 2025-12-04T09:47:20.8132340Z 2025-12-04T09:47:20.8133448Z preprocessor_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:47:20.8133896Z preprocessor_config.json: 1.21kB [00:00, 10.9MB/s] 2025-12-04T09:47:27.1864778Z INFO 12-04 09:47:27 [model.py:653] Resolved architecture: AriaForConditionalGeneration 2025-12-04T09:47:27.1865338Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:47:27.2118334Z INFO 12-04 09:47:27 [model.py:1714] Using max model len 65536 2025-12-04T09:47:27.3863035Z INFO 12-04 09:47:27 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:47:27.4181797Z 2025-12-04T09:47:27.4186130Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:47:27.4186515Z tokenizer_config.json: 110kB [00:00, 266MB/s] 2025-12-04T09:47:27.5170769Z 2025-12-04T09:47:27.7626562Z tokenizer.model: 0% 0.00/1.70M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:47:31.5386722Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:31.5413151Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:31.5420552Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:31.5427823Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:31.5434937Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:31.5442860Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:31.5445795Z (EngineCore_DP0 pid=7922) INFO 12-04 09:47:31 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:47:31.8230971Z (EngineCore_DP0 pid=7922) INFO 12-04 09:47:31 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:47:32.0048076Z (EngineCore_DP0 pid=7922) 2025-12-04T09:47:32.0049189Z processor_config.json: 0% 0.00/100 [00:00=4.57.0.dev0` is required to run this model. 2025-12-04T09:47:38.6890277Z PASSED 2025-12-04T09:47:38.7038862Z models/test_initialization.py::test_can_initialize_large_subset[DbrxForCausalLM] Fork a new process to run a test 8010 2025-12-04T09:47:38.7048765Z Fork a new process to run a test 0 2025-12-04T09:47:38.7319776Z INFO 12-04 09:47:38 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='DbrxForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'databricks/dbrx-instruct'} 2025-12-04T09:47:38.8648715Z 2025-12-04T09:47:38.8649731Z config.json: 0% 0.00/733 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:47:48.0893688Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:48.0920978Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:48.0927943Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:48.0935228Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:48.0942823Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:48.0950146Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:47:48.0952957Z (EngineCore_DP0 pid=8068) INFO 12-04 09:47:48 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:47:48.3749176Z (EngineCore_DP0 pid=8068) INFO 12-04 09:47:48 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:47:48.4750083Z (EngineCore_DP0 pid=8068) INFO 12-04 09:47:48 [gpu_model_runner.py:2840] Starting to load model databricks/dbrx-instruct... 2025-12-04T09:47:48.7468097Z (EngineCore_DP0 pid=8068) INFO 12-04 09:47:48 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:47:48.8119389Z (EngineCore_DP0 pid=8068) INFO 12-04 09:47:48 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:47:48.8145258Z (EngineCore_DP0 pid=8068) WARNING 12-04 09:47:48 [vllm.py:821] `torch.compile` is turned on, but the model databricks/dbrx-instruct does not support it. Please open an issue on GitHub if you want it to be supported. 2025-12-04T09:47:49.3524542Z (EngineCore_DP0 pid=8068) INFO 12-04 09:47:49 [gpu_model_runner.py:2902] Model loading took 8.3752 GiB and 0.070421 seconds 2025-12-04T09:47:49.3528790Z (EngineCore_DP0 pid=8068) INFO 12-04 09:47:49 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T09:47:49.3529576Z (EngineCore_DP0 pid=8068) INFO 12-04 09:47:49 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 80.00x 2025-12-04T09:47:49.9829371Z (EngineCore_DP0 pid=8068) INFO 12-04 09:47:49 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:47:49.9845972Z INFO 12-04 09:47:49 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:47:50.4534562Z PASSED 2025-12-04T09:47:50.4684318Z models/test_initialization.py::test_can_initialize_large_subset[ChameleonForConditionalGeneration] Fork a new process to run a test 8138 2025-12-04T09:47:50.4693804Z Fork a new process to run a test 0 2025-12-04T09:47:50.4972101Z INFO 12-04 09:47:50 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='ChameleonForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'facebook/chameleon-7b'} 2025-12-04T09:47:50.6048262Z 2025-12-04T09:47:50.7197075Z config.json: 0% 0.00/1.71M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:48:03.4873825Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:03.4901726Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:03.4909156Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:03.4916753Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:03.4924123Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:03.4931998Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:03.4934800Z (EngineCore_DP0 pid=8196) INFO 12-04 09:48:03 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:48:03.7721724Z (EngineCore_DP0 pid=8196) INFO 12-04 09:48:03 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:48:03.9403905Z (EngineCore_DP0 pid=8196) 2025-12-04T09:48:03.9406323Z processor_config.json: 0% 0.00/102 [00:00, model_arch='GraniteSpeechForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ibm-granite/granite-speech-3.3-2b'} 2025-12-04T09:48:12.8189450Z 2025-12-04T09:48:12.8191768Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:48:12.8192091Z config.json: 2.41kB [00:00, 12.7MB/s] 2025-12-04T09:48:12.9230884Z 2025-12-04T09:48:12.9231657Z preprocessor_config.json: 0% 0.00/2.00 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:48:21.4291801Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:21.4329165Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:21.4336105Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:21.4343922Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:21.4352076Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:21.4359132Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:21.4361795Z (EngineCore_DP0 pid=8340) INFO 12-04 09:48:21 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:48:21.7163262Z (EngineCore_DP0 pid=8340) INFO 12-04 09:48:21 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:48:22.9985575Z (EngineCore_DP0 pid=8340) INFO 12-04 09:48:22 [gpu_model_runner.py:2840] Starting to load model ibm-granite/granite-speech-3.3-2b... 2025-12-04T09:48:23.2766495Z (EngineCore_DP0 pid=8340) INFO 12-04 09:48:23 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:48:23.3248310Z (EngineCore_DP0 pid=8340) INFO 12-04 09:48:23 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:48:23.9649970Z (EngineCore_DP0 pid=8340) INFO 12-04 09:48:23 [gpu_model_runner.py:2902] Model loading took 0.4359 GiB and 0.142064 seconds 2025-12-04T09:48:23.9653927Z (EngineCore_DP0 pid=8340) INFO 12-04 09:48:23 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T09:48:23.9654879Z (EngineCore_DP0 pid=8340) INFO 12-04 09:48:23 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 40.00x 2025-12-04T09:48:25.0270428Z (EngineCore_DP0 pid=8340) INFO 12-04 09:48:25 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:48:25.0291483Z INFO 12-04 09:48:25 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:48:25.4907780Z PASSED 2025-12-04T09:48:25.5056315Z models/test_initialization.py::test_can_initialize_large_subset[NemotronH_Nano_VL_V2] Fork a new process to run a test 8426 2025-12-04T09:48:25.5066197Z Fork a new process to run a test 0 2025-12-04T09:48:25.5067673Z Model is not available online 2025-12-04T09:48:25.7892510Z PASSED 2025-12-04T09:48:25.8042308Z models/test_initialization.py::test_can_initialize_large_subset[ModernBertForTokenClassification] Fork a new process to run a test 8427 2025-12-04T09:48:25.8052639Z Fork a new process to run a test 0 2025-12-04T09:48:25.8323351Z INFO 12-04 09:48:25 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='ModernBertForTokenClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'disham993/electrical-ner-ModernBERT-base'} 2025-12-04T09:48:25.9928710Z 2025-12-04T09:48:25.9930896Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:48:25.9931202Z config.json: 1.96kB [00:00, 9.79MB/s] 2025-12-04T09:48:32.4892997Z INFO 12-04 09:48:32 [model.py:915] Resolved `--runner auto` to `--runner pooling`. Pass the value explicitly to silence this message. 2025-12-04T09:48:32.4893727Z INFO 12-04 09:48:32 [model.py:653] Resolved architecture: ModernBertForTokenClassification 2025-12-04T09:48:32.4894583Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:48:32.5149568Z INFO 12-04 09:48:32 [model.py:1939] Downcasting torch.float32 to torch.float16. 2025-12-04T09:48:32.5568993Z 2025-12-04T09:48:32.5571146Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:48:32.5571500Z tokenizer_config.json: 20.9kB [00:00, 146MB/s] 2025-12-04T09:48:32.5579926Z INFO 12-04 09:48:32 [model.py:1714] Using max model len 8192 2025-12-04T09:48:32.5892343Z INFO 12-04 09:48:32 [arg_utils.py:1725] (Disabling) chunked prefill by default 2025-12-04T09:48:32.5892865Z INFO 12-04 09:48:32 [arg_utils.py:1728] (Disabling) prefix caching by default 2025-12-04T09:48:32.7284365Z INFO 12-04 09:48:32 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T09:48:32.9016629Z 2025-12-04T09:48:32.9183353Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T09:48:32.9183785Z tokenizer.json: 3.58MB [00:00, 215MB/s] 2025-12-04T09:48:32.9558896Z 2025-12-04T09:48:32.9559569Z added_tokens.json: 0.00B [00:00, ?B/s] 2025-12-04T09:48:32.9559900Z added_tokens.json: 1.99kB [00:00, 31.2MB/s] 2025-12-04T09:48:32.9937256Z 2025-12-04T09:48:32.9937967Z special_tokens_map.json: 0% 0.00/694 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:48:34.6706880Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:34.6738371Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:34.6745445Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:34.6753048Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:34.6760837Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:34.6768904Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:34.6772251Z (EngineCore_DP0 pid=8486) INFO 12-04 09:48:34 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:48:34.9603529Z (EngineCore_DP0 pid=8486) INFO 12-04 09:48:34 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:48:34.9806764Z (EngineCore_DP0 pid=8486) INFO 12-04 09:48:34 [gpu_model_runner.py:2840] Starting to load model disham993/electrical-ner-ModernBERT-base... 2025-12-04T09:48:35.2511620Z (EngineCore_DP0 pid=8486) INFO 12-04 09:48:35 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:48:35.3166542Z (EngineCore_DP0 pid=8486) INFO 12-04 09:48:35 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:48:35.9110767Z (EngineCore_DP0 pid=8486) INFO 12-04 09:48:35 [gpu_model_runner.py:2902] Model loading took 0.0837 GiB and 0.126668 seconds 2025-12-04T09:48:36.0908699Z (EngineCore_DP0 pid=8486) INFO 12-04 09:48:36 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T09:48:36.0937010Z (EngineCore_DP0 pid=8486) INFO 12-04 09:48:36 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T09:48:36.3802233Z (EngineCore_DP0 pid=8486) INFO 12-04 09:48:36 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:48:36.3823905Z INFO 12-04 09:48:36 [llm.py:337] Supported tasks: ['encode'] 2025-12-04T09:48:36.8122782Z PASSED 2025-12-04T09:48:36.8273384Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2AudioForConditionalGeneration] Fork a new process to run a test 8556 2025-12-04T09:48:36.8282837Z Fork a new process to run a test 0 2025-12-04T09:48:36.8558357Z INFO 12-04 09:48:36 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2AudioForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen2-Audio-7B-Instruct'} 2025-12-04T09:48:36.9258164Z 2025-12-04T09:48:36.9259554Z config.json: 0% 0.00/853 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:48:46.1072616Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:46.1098822Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:46.1106460Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:46.1113855Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:46.1121436Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:46.1128993Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:46.1131578Z (EngineCore_DP0 pid=8614) INFO 12-04 09:48:46 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:48:46.3938308Z (EngineCore_DP0 pid=8614) INFO 12-04 09:48:46 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:48:50.5525815Z (EngineCore_DP0 pid=8614) INFO 12-04 09:48:50 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen2-Audio-7B-Instruct... 2025-12-04T09:48:50.9066956Z (EngineCore_DP0 pid=8614) INFO 12-04 09:48:50 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:48:50.9377268Z (EngineCore_DP0 pid=8614) INFO 12-04 09:48:50 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:48:51.5365074Z (EngineCore_DP0 pid=8614) INFO 12-04 09:48:51 [gpu_model_runner.py:2902] Model loading took 2.8225 GiB and 0.090528 seconds 2025-12-04T09:48:51.5369462Z (EngineCore_DP0 pid=8614) INFO 12-04 09:48:51 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-12-04T09:48:51.5370387Z (EngineCore_DP0 pid=8614) INFO 12-04 09:48:51 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 80.00x 2025-12-04T09:48:52.3869924Z (EngineCore_DP0 pid=8614) INFO 12-04 09:48:52 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:48:52.3889985Z INFO 12-04 09:48:52 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:48:52.8609519Z PASSED 2025-12-04T09:48:52.8758012Z models/test_initialization.py::test_can_initialize_large_subset[DeepseekV2ForCausalLM] Fork a new process to run a test 8700 2025-12-04T09:48:52.8767755Z Fork a new process to run a test 0 2025-12-04T09:48:52.9034442Z INFO 12-04 09:48:52 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='DeepseekV2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'deepseek-ai/DeepSeek-V2-Lite-Chat'} 2025-12-04T09:48:52.9824739Z 2025-12-04T09:48:52.9827109Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:48:52.9827453Z config.json: 1.52kB [00:00, 9.23MB/s] 2025-12-04T09:48:53.0605286Z 2025-12-04T09:48:53.0607310Z configuration_deepseek.py: 0.00B [00:00, ?B/s] 2025-12-04T09:48:53.0607733Z configuration_deepseek.py: 10.3kB [00:00, 64.1MB/s] 2025-12-04T09:48:53.0707235Z A new version of the following files was downloaded from https://huggingface.co/deepseek-ai/DeepSeek-V2-Lite-Chat: 2025-12-04T09:48:53.0707844Z - configuration_deepseek.py 2025-12-04T09:48:53.0708463Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:48:53.1029907Z INFO 12-04 09:48:53 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T09:48:53.2723315Z INFO 12-04 09:48:53 [model.py:653] Resolved architecture: DeepseekV2ForCausalLM 2025-12-04T09:48:53.2723838Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:48:53.2967828Z INFO 12-04 09:48:53 [model.py:1714] Using max model len 163840 2025-12-04T09:48:53.2969747Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T09:48:53.3487988Z INFO 12-04 09:48:53 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:48:53.3797163Z 2025-12-04T09:48:53.3798406Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:48:53.3798770Z tokenizer_config.json: 1.28kB [00:00, 14.0MB/s] 2025-12-04T09:48:53.5435781Z 2025-12-04T09:48:53.5652641Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T09:48:53.5652974Z tokenizer.json: 4.61MB [00:00, 213MB/s] 2025-12-04T09:48:53.9297079Z 2025-12-04T09:48:53.9297587Z generation_config.json: 0% 0.00/181 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:48:55.2382648Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:55.2409971Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:55.2417360Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:55.2424558Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:55.2432324Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:55.2439709Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:48:55.2442601Z (EngineCore_DP0 pid=8708) INFO 12-04 09:48:55 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:48:55.5292901Z (EngineCore_DP0 pid=8708) INFO 12-04 09:48:55 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:48:55.6018491Z (EngineCore_DP0 pid=8708) INFO 12-04 09:48:55 [gpu_model_runner.py:2840] Starting to load model deepseek-ai/DeepSeek-V2-Lite-Chat... 2025-12-04T09:48:55.8854416Z (EngineCore_DP0 pid=8708) INFO 12-04 09:48:55 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:48:55.9729970Z (EngineCore_DP0 pid=8708) INFO 12-04 09:48:55 [cuda.py:328] Using Triton MLA backend on V1 engine. 2025-12-04T09:48:56.5939617Z (EngineCore_DP0 pid=8708) INFO 12-04 09:48:56 [gpu_model_runner.py:2902] Model loading took 0.8919 GiB and 0.161054 seconds 2025-12-04T09:48:56.5944082Z (EngineCore_DP0 pid=8708) INFO 12-04 09:48:56 [kv_cache_utils.py:1199] GPU KV cache size: 9,320,672 tokens 2025-12-04T09:48:56.5944976Z (EngineCore_DP0 pid=8708) INFO 12-04 09:48:56 [kv_cache_utils.py:1204] Maximum concurrency for 163,840 tokens per request: 56.89x 2025-12-04T09:48:58.1213160Z (EngineCore_DP0 pid=8708) INFO 12-04 09:48:58 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:48:58.1227700Z INFO 12-04 09:48:58 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:48:58.5922094Z PASSED 2025-12-04T09:48:58.6071370Z models/test_initialization.py::test_can_initialize_large_subset[GritLM] Fork a new process to run a test 8778 2025-12-04T09:48:58.6081246Z Fork a new process to run a test 0 2025-12-04T09:48:58.6353311Z INFO 12-04 09:48:58 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GritLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'parasail-ai/GritLM-7B-vllm'} 2025-12-04T09:48:58.7966798Z 2025-12-04T09:48:58.7967787Z config.json: 0% 0.00/934 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:49:07.5935685Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:07.5965782Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:07.5972674Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:07.5979980Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:07.5987966Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:07.5995409Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:07.5997698Z (EngineCore_DP0 pid=8857) INFO 12-04 09:49:07 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:49:07.8778635Z (EngineCore_DP0 pid=8857) INFO 12-04 09:49:07 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:49:08.0214771Z (EngineCore_DP0 pid=8857) INFO 12-04 09:49:08 [gpu_model_runner.py:2840] Starting to load model parasail-ai/GritLM-7B-vllm... 2025-12-04T09:49:08.3017563Z (EngineCore_DP0 pid=8857) INFO 12-04 09:49:08 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:49:08.3696749Z (EngineCore_DP0 pid=8857) INFO 12-04 09:49:08 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:49:09.1258884Z (EngineCore_DP0 pid=8857) INFO 12-04 09:49:09 [gpu_model_runner.py:2902] Model loading took 0.9024 GiB and 0.259909 seconds 2025-12-04T09:49:09.2438190Z (EngineCore_DP0 pid=8857) INFO 12-04 09:49:09 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T09:49:09.2781923Z (EngineCore_DP0 pid=8857) INFO 12-04 09:49:09 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T09:49:09.2783052Z (EngineCore_DP0 pid=8857) INFO 12-04 09:49:09 [vllm.py:433] Only models using causal attention supports chunked prefill and prefix caching; disabling both. 2025-12-04T09:49:09.5624684Z (EngineCore_DP0 pid=8857) INFO 12-04 09:49:09 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:49:09.5646818Z INFO 12-04 09:49:09 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-12-04T09:49:10.0545582Z PASSED 2025-12-04T09:49:10.0694956Z models/test_initialization.py::test_can_initialize_large_subset[GteNewForSequenceClassification] Fork a new process to run a test 8927 2025-12-04T09:49:10.0704509Z Fork a new process to run a test 0 2025-12-04T09:49:10.0975286Z INFO 12-04 09:49:10 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GteNewForSequenceClassification', exist_overrides={'architectures': ['GteNewForSequenceClassification']}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Alibaba-NLP/gte-multilingual-reranker-base'} 2025-12-04T09:49:10.1705535Z 2025-12-04T09:49:10.1707492Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:49:10.1707808Z config.json: 1.49kB [00:00, 8.66MB/s] 2025-12-04T09:49:10.2383161Z 2025-12-04T09:49:10.2384750Z configuration.py: 0.00B [00:00, ?B/s] 2025-12-04T09:49:10.2385090Z configuration.py: 7.13kB [00:00, 54.7MB/s] 2025-12-04T09:49:10.2483503Z A new version of the following files was downloaded from https://huggingface.co/Alibaba-NLP/new-impl: 2025-12-04T09:49:10.2484044Z - configuration.py 2025-12-04T09:49:10.2484646Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:49:10.2805769Z INFO 12-04 09:49:10 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T09:49:16.7195764Z INFO 12-04 09:49:16 [model.py:915] Resolved `--runner auto` to `--runner pooling`. Pass the value explicitly to silence this message. 2025-12-04T09:49:16.7197088Z INFO 12-04 09:49:16 [model.py:653] Resolved architecture: GteNewForSequenceClassification 2025-12-04T09:49:16.7197700Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:49:16.7448894Z INFO 12-04 09:49:16 [model.py:1714] Using max model len 65536 2025-12-04T09:49:16.7451080Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T09:49:16.7812795Z INFO 12-04 09:49:16 [arg_utils.py:1725] (Disabling) chunked prefill by default 2025-12-04T09:49:16.7813326Z INFO 12-04 09:49:16 [arg_utils.py:1728] (Disabling) prefix caching by default 2025-12-04T09:49:16.8087945Z INFO 12-04 09:49:16 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T09:49:16.8436977Z 2025-12-04T09:49:16.8437797Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:49:16.8438172Z tokenizer_config.json: 1.34kB [00:00, 15.0MB/s] 2025-12-04T09:49:16.9632900Z 2025-12-04T09:49:17.2682843Z tokenizer.json: 0% 0.00/17.1M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:49:19.5309412Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:19.5337109Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:19.5344479Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:19.5351648Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:19.5359096Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:19.5366773Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:19.5369718Z (EngineCore_DP0 pid=9006) INFO 12-04 09:49:19 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:49:19.8152862Z (EngineCore_DP0 pid=9006) INFO 12-04 09:49:19 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:49:19.9048172Z (EngineCore_DP0 pid=9006) INFO 12-04 09:49:19 [gpu_model_runner.py:2840] Starting to load model Alibaba-NLP/gte-multilingual-reranker-base... 2025-12-04T09:49:20.1780055Z (EngineCore_DP0 pid=9006) INFO 12-04 09:49:20 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:49:20.2419636Z (EngineCore_DP0 pid=9006) INFO 12-04 09:49:20 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:49:20.8396593Z (EngineCore_DP0 pid=9006) INFO 12-04 09:49:20 [gpu_model_runner.py:2902] Model loading took 0.3780 GiB and 0.126863 seconds 2025-12-04T09:49:21.6174008Z (EngineCore_DP0 pid=9006) INFO 12-04 09:49:21 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T09:49:21.6222559Z (EngineCore_DP0 pid=9006) INFO 12-04 09:49:21 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T09:49:21.8980965Z (EngineCore_DP0 pid=9006) INFO 12-04 09:49:21 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:49:21.8995189Z INFO 12-04 09:49:21 [llm.py:337] Supported tasks: ['encode', 'classify', 'score'] 2025-12-04T09:49:22.4220189Z PASSED 2025-12-04T09:49:22.4369515Z models/test_initialization.py::test_can_initialize_large_subset[DeepseekForCausalLM] Fork a new process to run a test 9076 2025-12-04T09:49:22.4377895Z Fork a new process to run a test 0 2025-12-04T09:49:22.4648587Z INFO 12-04 09:49:22 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='DeepseekForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'deepseek-ai/deepseek-llm-7b-chat'} 2025-12-04T09:49:22.5456000Z 2025-12-04T09:49:22.5456628Z config.json: 0% 0.00/594 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:49:24.8601259Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:24.8627517Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:24.8634413Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:24.8641398Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:24.8649006Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:24.8656330Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:24.8659006Z (EngineCore_DP0 pid=9084) INFO 12-04 09:49:24 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:49:25.1482019Z (EngineCore_DP0 pid=9084) INFO 12-04 09:49:25 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:49:25.2123099Z (EngineCore_DP0 pid=9084) INFO 12-04 09:49:25 [gpu_model_runner.py:2840] Starting to load model deepseek-ai/deepseek-llm-7b-chat... 2025-12-04T09:49:25.4864503Z (EngineCore_DP0 pid=9084) INFO 12-04 09:49:25 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:49:25.5511733Z (EngineCore_DP0 pid=9084) INFO 12-04 09:49:25 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:49:26.1589271Z (EngineCore_DP0 pid=9084) INFO 12-04 09:49:26 [gpu_model_runner.py:2902] Model loading took 1.9405 GiB and 0.138413 seconds 2025-12-04T09:49:26.1593342Z (EngineCore_DP0 pid=9084) INFO 12-04 09:49:26 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-12-04T09:49:26.1594648Z (EngineCore_DP0 pid=9084) INFO 12-04 09:49:26 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 160.00x 2025-12-04T09:49:26.7716268Z (EngineCore_DP0 pid=9084) INFO 12-04 09:49:26 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:49:26.7738131Z INFO 12-04 09:49:26 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:49:27.2216273Z PASSED 2025-12-04T09:49:27.2366050Z models/test_initialization.py::test_can_initialize_large_subset[Gemma3ForCausalLM] Fork a new process to run a test 9154 2025-12-04T09:49:27.2375988Z Fork a new process to run a test 0 2025-12-04T09:49:27.2647271Z INFO 12-04 09:49:27 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Gemma3ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'google/gemma-3-1b-it'} 2025-12-04T09:49:27.3651781Z 2025-12-04T09:49:27.3653367Z config.json: 0% 0.00/899 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:49:37.6828505Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:37.6855415Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:37.6865676Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:37.6875110Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:37.6883188Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:37.6890805Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:37.6893562Z (EngineCore_DP0 pid=9233) INFO 12-04 09:49:37 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:49:37.9704537Z (EngineCore_DP0 pid=9233) INFO 12-04 09:49:37 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:49:38.0431154Z (EngineCore_DP0 pid=9233) INFO 12-04 09:49:38 [gpu_model_runner.py:2840] Starting to load model google/gemma-3-1b-it... 2025-12-04T09:49:38.3163527Z (EngineCore_DP0 pid=9233) INFO 12-04 09:49:38 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:49:38.4031854Z (EngineCore_DP0 pid=9233) INFO 12-04 09:49:38 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:49:38.9981518Z (EngineCore_DP0 pid=9233) INFO 12-04 09:49:38 [gpu_model_runner.py:2902] Model loading took 0.6289 GiB and 0.145606 seconds 2025-12-04T09:49:38.9986848Z (EngineCore_DP0 pid=9233) INFO 12-04 09:49:38 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T09:49:41.8181097Z (EngineCore_DP0 pid=9233) INFO 12-04 09:49:38 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 1202.50x 2025-12-04T09:49:41.8181888Z (EngineCore_DP0 pid=9233) INFO 12-04 09:49:41 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:49:41.8195755Z INFO 12-04 09:49:41 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:49:42.3290724Z PASSED 2025-12-04T09:49:42.3440468Z models/test_initialization.py::test_can_initialize_large_subset[Llama4ForCausalLM] Fork a new process to run a test 9303 2025-12-04T09:49:42.3450694Z Fork a new process to run a test 0 2025-12-04T09:49:42.3451879Z Model is not available online 2025-12-04T09:49:42.6286749Z PASSED 2025-12-04T09:49:42.6435914Z models/test_initialization.py::test_can_initialize_large_subset[Exaone4ForCausalLM] Fork a new process to run a test 9304 2025-12-04T09:49:42.6445164Z Fork a new process to run a test 0 2025-12-04T09:49:42.6724265Z INFO 12-04 09:49:42 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Exaone4ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'LGAI-EXAONE/EXAONE-4.0-32B'} 2025-12-04T09:49:42.7469278Z 2025-12-04T09:49:42.7471387Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:49:42.7471687Z config.json: 2.42kB [00:00, 15.1MB/s] 2025-12-04T09:49:49.2204059Z INFO 12-04 09:49:49 [model.py:653] Resolved architecture: Exaone4ForCausalLM 2025-12-04T09:49:49.2204851Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:49:49.2455050Z INFO 12-04 09:49:49 [model.py:1714] Using max model len 131072 2025-12-04T09:49:49.4159784Z INFO 12-04 09:49:49 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:49:49.4477466Z 2025-12-04T09:49:49.4481409Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:49:49.4481769Z tokenizer_config.json: 70.3kB [00:00, 194MB/s] 2025-12-04T09:49:49.5506130Z 2025-12-04T09:49:49.5630576Z vocab.json: 0.00B [00:00, ?B/s] 2025-12-04T09:49:49.5630990Z vocab.json: 1.93MB [00:00, 156MB/s] 2025-12-04T09:49:49.6138395Z 2025-12-04T09:49:49.6196681Z merges.txt: 0.00B [00:00, ?B/s] 2025-12-04T09:49:49.6197079Z merges.txt: 1.22MB [00:00, 210MB/s] 2025-12-04T09:49:49.6570373Z 2025-12-04T09:49:49.6805262Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T09:49:49.6805676Z tokenizer.json: 7.91MB [00:00, 337MB/s] 2025-12-04T09:49:49.7927036Z 2025-12-04T09:49:49.7928361Z special_tokens_map.json: 0.00B [00:00, ?B/s] 2025-12-04T09:49:49.7928726Z special_tokens_map.json: 6.70kB [00:00, 67.9MB/s] 2025-12-04T09:49:49.8307503Z 2025-12-04T09:49:49.8308877Z chat_template.jinja: 0.00B [00:00, ?B/s] 2025-12-04T09:49:49.8309213Z chat_template.jinja: 5.49kB [00:00, 51.3MB/s] 2025-12-04T09:49:50.1215507Z 2025-12-04T09:49:50.1216187Z generation_config.json: 0% 0.00/170 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:49:51.4818521Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:51.4845908Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:51.4853153Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:51.4861204Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:51.4868778Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:51.4877018Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:49:51.4879749Z (EngineCore_DP0 pid=9362) INFO 12-04 09:49:51 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:49:51.7678562Z (EngineCore_DP0 pid=9362) INFO 12-04 09:49:51 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:49:51.8780718Z (EngineCore_DP0 pid=9362) INFO 12-04 09:49:51 [gpu_model_runner.py:2840] Starting to load model LGAI-EXAONE/EXAONE-4.0-32B... 2025-12-04T09:49:52.1557914Z (EngineCore_DP0 pid=9362) INFO 12-04 09:49:52 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:49:52.2496398Z (EngineCore_DP0 pid=9362) INFO 12-04 09:49:52 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:49:52.8510619Z (EngineCore_DP0 pid=9362) INFO 12-04 09:49:52 [gpu_model_runner.py:2902] Model loading took 2.8868 GiB and 0.151615 seconds 2025-12-04T09:49:52.8514798Z (EngineCore_DP0 pid=9362) INFO 12-04 09:49:52 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T09:49:52.8515742Z (EngineCore_DP0 pid=9362) INFO 12-04 09:49:52 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 213.06x 2025-12-04T09:49:53.5370511Z (EngineCore_DP0 pid=9362) INFO 12-04 09:49:53 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:49:53.5391637Z INFO 12-04 09:49:53 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:49:54.0248600Z PASSED 2025-12-04T09:49:54.0398975Z models/test_initialization.py::test_can_initialize_large_subset[GPT2LMHeadModel] Fork a new process to run a test 9432 2025-12-04T09:49:54.0407622Z Fork a new process to run a test 0 2025-12-04T09:49:54.0679405Z INFO 12-04 09:49:54 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GPT2LMHeadModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'openai-community/gpt2'} 2025-12-04T09:49:54.1428915Z 2025-12-04T09:49:54.1430127Z config.json: 0% 0.00/665 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:50:02.8385451Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:02.8411848Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:02.8419297Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:02.8427236Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:02.8434784Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:02.8442497Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:02.8445598Z (EngineCore_DP0 pid=9491) INFO 12-04 09:50:02 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:50:03.1221938Z (EngineCore_DP0 pid=9491) INFO 12-04 09:50:03 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:50:03.1637802Z (EngineCore_DP0 pid=9491) INFO 12-04 09:50:03 [gpu_model_runner.py:2840] Starting to load model openai-community/gpt2... 2025-12-04T09:50:03.4289550Z (EngineCore_DP0 pid=9491) INFO 12-04 09:50:03 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:50:03.4327590Z (EngineCore_DP0 pid=9491) INFO 12-04 09:50:03 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:50:04.0260084Z (EngineCore_DP0 pid=9491) INFO 12-04 09:50:04 [gpu_model_runner.py:2902] Model loading took 0.0869 GiB and 0.063086 seconds 2025-12-04T09:50:04.0263963Z (EngineCore_DP0 pid=9491) INFO 12-04 09:50:04 [kv_cache_utils.py:1199] GPU KV cache size: 3,495,248 tokens 2025-12-04T09:50:04.0264895Z (EngineCore_DP0 pid=9491) INFO 12-04 09:50:04 [kv_cache_utils.py:1204] Maximum concurrency for 1,024 tokens per request: 3413.33x 2025-12-04T09:50:04.6036775Z (EngineCore_DP0 pid=9491) INFO 12-04 09:50:04 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:50:04.6057364Z INFO 12-04 09:50:04 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:50:05.0276281Z PASSED 2025-12-04T09:50:05.0425548Z models/test_initialization.py::test_can_initialize_large_subset[KeyeVL1_5ForConditionalGeneration] Fork a new process to run a test 9561 2025-12-04T09:50:05.0434429Z Fork a new process to run a test 0 2025-12-04T09:50:05.0698005Z INFO 12-04 09:50:05 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='KeyeVL1_5ForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Kwai-Keye/Keye-VL-1_5-8B'} 2025-12-04T09:50:05.1496732Z 2025-12-04T09:50:05.1499326Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:50:05.1499658Z config.json: 2.03kB [00:00, 10.4MB/s] 2025-12-04T09:50:05.2149830Z 2025-12-04T09:50:05.2151944Z configuration_keye_vl_1_5.py: 0.00B [00:00, ?B/s] 2025-12-04T09:50:05.2152504Z configuration_keye_vl_1_5.py: 12.9kB [00:00, 70.8MB/s] 2025-12-04T09:50:05.2257297Z A new version of the following files was downloaded from https://huggingface.co/Kwai-Keye/Keye-VL-1_5-8B: 2025-12-04T09:50:05.2257885Z - configuration_keye_vl_1_5.py 2025-12-04T09:50:05.2258528Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:50:05.4402299Z 2025-12-04T09:50:05.4403147Z preprocessor_config.json: 0% 0.00/458 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:50:14.7246926Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:14.7279935Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:14.7287088Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:14.7294997Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:14.7302848Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:14.7310116Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:14.7312810Z (EngineCore_DP0 pid=9640) INFO 12-04 09:50:14 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:50:15.0195234Z (EngineCore_DP0 pid=9640) INFO 12-04 09:50:15 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:50:15.1633339Z (EngineCore_DP0 pid=9640) 2025-12-04T09:50:15.1634300Z processor_config.json: 0% 0.00/132 [00:00, model_arch='WhisperForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'openai/whisper-large-v3'} 2025-12-04T09:50:20.5163457Z 2025-12-04T09:50:20.5165750Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:50:20.5166081Z config.json: 1.27kB [00:00, 7.19MB/s] 2025-12-04T09:50:20.6053456Z 2025-12-04T09:50:20.6054158Z preprocessor_config.json: 0% 0.00/340 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:50:36.9306434Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:36.9333831Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:36.9340523Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:36.9347559Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:36.9354003Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:36.9360785Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:36.9362975Z (EngineCore_DP0 pid=9792) INFO 12-04 09:50:36 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:50:37.1839836Z (EngineCore_DP0 pid=9792) INFO 12-04 09:50:37 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:50:38.8063848Z (EngineCore_DP0 pid=9792) INFO 12-04 09:50:38 [gpu_model_runner.py:2840] Starting to load model openai/whisper-large-v3... 2025-12-04T09:50:39.0802503Z (EngineCore_DP0 pid=9792) INFO 12-04 09:50:39 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:50:39.0832689Z (EngineCore_DP0 pid=9792) INFO 12-04 09:50:39 [layer.py:497] MultiHeadAttention attn_backend: _Backend.FLASH_ATTN, use_upstream_fa: False 2025-12-04T09:50:39.1400918Z (EngineCore_DP0 pid=9792) INFO 12-04 09:50:39 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:50:39.2152396Z (EngineCore_DP0 pid=9792) WARNING 12-04 09:50:39 [vllm.py:821] `torch.compile` is turned on, but the model openai/whisper-large-v3 does not support it. Please open an issue on GitHub if you want it to be supported. 2025-12-04T09:50:39.7202551Z (EngineCore_DP0 pid=9792) INFO 12-04 09:50:39 [gpu_model_runner.py:2902] Model loading took 2.8763 GiB and 0.155040 seconds 2025-12-04T09:50:39.9558357Z (EngineCore_DP0 pid=9792) INFO 12-04 09:50:39 [gpu_model_runner.py:3647] Encoder cache will be initialized with a budget of 1500 tokens, and profiled with 1 audio items of the maximum feature size. 2025-12-04T09:50:39.9576679Z (EngineCore_DP0 pid=9792) WARNING 12-04 09:50:39 [processing.py:1091] WhisperProcessor did not return `BatchFeature`. Make sure to match the behaviour of `ProcessorMixin` when implementing custom processors. 2025-12-04T09:50:41.1205336Z (EngineCore_DP0 pid=9792) INFO 12-04 09:50:41 [gpu_worker.py:314] Available KV cache memory: 14.45 GiB 2025-12-04T09:50:41.3551614Z (EngineCore_DP0 pid=9792) INFO 12-04 09:50:41 [kv_cache_utils.py:1199] GPU KV cache size: 47,328 tokens 2025-12-04T09:50:41.3552578Z (EngineCore_DP0 pid=9792) INFO 12-04 09:50:41 [kv_cache_utils.py:1204] Maximum concurrency for 448 tokens per request: 48.50x 2025-12-04T09:50:41.6043444Z (EngineCore_DP0 pid=9792) 2025-12-04T09:50:41.7352864Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/67 [00:00, model_arch='MiDashengLMModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'mispeech/midashenglm-7b'} 2025-12-04T09:50:45.2774848Z 2025-12-04T09:50:45.2776640Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:50:45.2776963Z config.json: 1.65kB [00:00, 9.37MB/s] 2025-12-04T09:50:45.3487171Z Unrecognized keys in `rope_scaling` for 'rope_type'='default': {'mrope_section'} 2025-12-04T09:50:45.4679561Z 2025-12-04T09:50:45.4680361Z preprocessor_config.json: 0% 0.00/349 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:50:55.2527875Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:55.2554794Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:55.2561917Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:55.2569521Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:55.2577266Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:55.2585992Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:50:55.2588369Z (EngineCore_DP0 pid=10011) INFO 12-04 09:50:55 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:50:55.5473416Z (EngineCore_DP0 pid=10011) INFO 12-04 09:50:55 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:50:55.6939810Z (EngineCore_DP0 pid=10011) 2025-12-04T09:50:55.6940230Z processor_config.json: 0% 0.00/268 [00:00, model_arch='ApertusForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'swiss-ai/Apertus-8B-Instruct-2509'} 2025-12-04T09:51:01.8314015Z 2025-12-04T09:51:01.8315100Z config.json: 0% 0.00/901 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:51:10.9472637Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:10.9499700Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:10.9507351Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:10.9515246Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:10.9523437Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:10.9531449Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:10.9534260Z (EngineCore_DP0 pid=10175) INFO 12-04 09:51:10 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:51:11.2331629Z (EngineCore_DP0 pid=10175) INFO 12-04 09:51:11 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:51:11.3394115Z (EngineCore_DP0 pid=10175) INFO 12-04 09:51:11 [gpu_model_runner.py:2840] Starting to load model swiss-ai/Apertus-8B-Instruct-2509... 2025-12-04T09:51:11.6140957Z (EngineCore_DP0 pid=10175) INFO 12-04 09:51:11 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:51:11.7103706Z (EngineCore_DP0 pid=10175) INFO 12-04 09:51:11 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:51:11.7361234Z (EngineCore_DP0 pid=10175) WARNING 12-04 09:51:11 [activation.py:420] CUDA-fused xIELU not available (No module named 'xielu') – falling back to a Python version. 2025-12-04T09:51:11.7361853Z 2025-12-04T09:51:11.7362490Z (EngineCore_DP0 pid=10175) WARNING 12-04 09:51:11 [activation.py:420] For CUDA xIELU (experimental), `pip install git+https://github.com/nickjbrowning/XIELU` 2025-12-04T09:51:12.3436011Z (EngineCore_DP0 pid=10175) INFO 12-04 09:51:12 [gpu_model_runner.py:2902] Model loading took 2.4219 GiB and 0.180947 seconds 2025-12-04T09:51:12.3440446Z (EngineCore_DP0 pid=10175) INFO 12-04 09:51:12 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T09:51:12.3441172Z (EngineCore_DP0 pid=10175) INFO 12-04 09:51:12 [kv_cache_utils.py:1204] Maximum concurrency for 65,536 tokens per request: 40.00x 2025-12-04T09:51:13.3452211Z (EngineCore_DP0 pid=10175) INFO 12-04 09:51:13 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:51:13.3466893Z INFO 12-04 09:51:13 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:51:13.8305619Z PASSED 2025-12-04T09:51:13.8461753Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2ForProcessRewardModel] Fork a new process to run a test 10245 2025-12-04T09:51:13.8472192Z Fork a new process to run a test 0 2025-12-04T09:51:13.8476121Z `transformers==4.56.2` installed, but `transformers<=4.53` is required to run this model. Reason: HF model uses remote code that is not compatible with latest Transformers 2025-12-04T09:51:14.1339182Z PASSED 2025-12-04T09:51:14.1488876Z models/test_initialization.py::test_can_initialize_large_subset[MiniMaxM1ForCausalLM] Fork a new process to run a test 10246 2025-12-04T09:51:14.1498049Z Fork a new process to run a test 0 2025-12-04T09:51:14.1760478Z INFO 12-04 09:51:14 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MiniMaxM1ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'MiniMaxAI/MiniMax-M1-40k'} 2025-12-04T09:51:14.2689807Z 2025-12-04T09:51:14.2691614Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:51:14.2691906Z config.json: 1.87kB [00:00, 10.6MB/s] 2025-12-04T09:51:14.3449740Z 2025-12-04T09:51:14.3451785Z configuration_minimax_m1.py: 0.00B [00:00, ?B/s] 2025-12-04T09:51:14.3452189Z configuration_minimax_m1.py: 7.30kB [00:00, 24.6MB/s] 2025-12-04T09:51:14.3544108Z A new version of the following files was downloaded from https://huggingface.co/MiniMaxAI/MiniMax-M1-40k: 2025-12-04T09:51:14.3544641Z - configuration_minimax_m1.py 2025-12-04T09:51:14.3545292Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:51:14.3837586Z You are using a model of type minimax_m1 to instantiate a model of type MiniMaxM1. This is not supported for all configurations of models and can yield errors. 2025-12-04T09:51:20.8521498Z INFO 12-04 09:51:20 [model.py:653] Resolved architecture: MiniMaxM1ForCausalLM 2025-12-04T09:51:20.8521981Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:51:20.9733246Z 2025-12-04T09:51:20.9756546Z model.safetensors.index.json: 0.00B [00:00, ?B/s] 2025-12-04T09:51:20.9757269Z model.safetensors.index.json: 823kB [00:00, 339MB/s] 2025-12-04T09:51:21.0030271Z 2025-12-04T09:51:21.1412344Z Parse safetensors files: 0% 0/413 [00:00= mamba page size. 2025-12-04T09:51:22.5350207Z 2025-12-04T09:51:22.5351351Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:51:22.5351757Z tokenizer_config.json: 1.97kB [00:00, 23.2MB/s] 2025-12-04T09:51:22.6453398Z 2025-12-04T09:51:22.6668553Z vocab.json: 0.00B [00:00, ?B/s] 2025-12-04T09:51:22.6668857Z vocab.json: 4.71MB [00:00, 220MB/s] 2025-12-04T09:51:22.7403369Z 2025-12-04T09:51:22.7541037Z merges.txt: 0.00B [00:00, ?B/s] 2025-12-04T09:51:22.7541388Z merges.txt: 2.41MB [00:00, 176MB/s] 2025-12-04T09:51:22.8574246Z 2025-12-04T09:51:22.8964964Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T09:51:22.8965649Z tokenizer.json: 9.73MB [00:00, 251MB/s] 2025-12-04T09:51:23.7347019Z You are using a model of type minimax_m1 to instantiate a model of type MiniMaxM1. This is not supported for all configurations of models and can yield errors. 2025-12-04T09:51:23.7576646Z (EngineCore_DP0 pid=10324) INFO 12-04 09:51:23 [core.py:727] Waiting for init message from front-end. 2025-12-04T09:51:23.7615664Z (EngineCore_DP0 pid=10324) INFO 12-04 09:51:23 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='MiniMaxAI/MiniMax-M1-40k', speculative_config=None, tokenizer='MiniMaxAI/MiniMax-M1-40k', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=10240000, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=MiniMaxAI/MiniMax-M1-40k, enable_prefix_caching=False, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:51:25.1011300Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:25.1040271Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:25.1048811Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:25.1056969Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:25.1066059Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:25.1074532Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:25.1077206Z (EngineCore_DP0 pid=10324) INFO 12-04 09:51:25 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:51:25.3866371Z (EngineCore_DP0 pid=10324) INFO 12-04 09:51:25 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:51:26.8740831Z (EngineCore_DP0 pid=10324) INFO 12-04 09:51:26 [gpu_model_runner.py:2840] Starting to load model MiniMaxAI/MiniMax-M1-40k... 2025-12-04T09:51:27.1490716Z (EngineCore_DP0 pid=10324) INFO 12-04 09:51:27 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:51:27.7999457Z (EngineCore_DP0 pid=10324) INFO 12-04 09:51:27 [gpu_model_runner.py:2902] Model loading took 7.9713 GiB and 0.109958 seconds 2025-12-04T09:51:27.8004198Z (EngineCore_DP0 pid=10324) INFO 12-04 09:51:27 [kv_cache_utils.py:1199] GPU KV cache size: 52,428,800,000 tokens 2025-12-04T09:51:27.8005040Z (EngineCore_DP0 pid=10324) INFO 12-04 09:51:27 [kv_cache_utils.py:1204] Maximum concurrency for 10,240,000 tokens per request: 5120.00x 2025-12-04T09:51:28.7242718Z (EngineCore_DP0 pid=10324) INFO 12-04 09:51:28 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:51:28.7258569Z INFO 12-04 09:51:28 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:51:30.1551836Z PASSED 2025-12-04T09:51:30.1701743Z models/test_initialization.py::test_can_initialize_large_subset[Glm4MoeForCausalLM] Fork a new process to run a test 10394 2025-12-04T09:51:30.1710949Z Fork a new process to run a test 0 2025-12-04T09:51:30.1988854Z INFO 12-04 09:51:30 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Glm4MoeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'zai-org/GLM-4.5'} 2025-12-04T09:51:30.3375221Z 2025-12-04T09:51:30.3377302Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:51:30.3377611Z config.json: 1.01kB [00:00, 5.72MB/s] 2025-12-04T09:51:36.8232875Z INFO 12-04 09:51:36 [model.py:653] Resolved architecture: Glm4MoeForCausalLM 2025-12-04T09:51:36.8233377Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:51:36.8485701Z INFO 12-04 09:51:36 [model.py:1714] Using max model len 131072 2025-12-04T09:51:37.0227714Z INFO 12-04 09:51:37 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:51:37.0624053Z 2025-12-04T09:51:37.0625859Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:51:37.0626274Z tokenizer_config.json: 7.31kB [00:00, 51.9MB/s] 2025-12-04T09:51:37.2169245Z 2025-12-04T09:51:37.5103391Z tokenizer.json: 0% 0.00/20.0M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:51:39.7323472Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:39.7350627Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:39.7357503Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:39.7364677Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:39.7372741Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:39.7380273Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:39.7382912Z (EngineCore_DP0 pid=10475) INFO 12-04 09:51:39 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:51:40.0293279Z (EngineCore_DP0 pid=10475) INFO 12-04 09:51:40 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:51:40.1567261Z (EngineCore_DP0 pid=10475) INFO 12-04 09:51:40 [gpu_model_runner.py:2840] Starting to load model zai-org/GLM-4.5... 2025-12-04T09:51:40.4421557Z (EngineCore_DP0 pid=10475) INFO 12-04 09:51:40 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:51:40.5226833Z (EngineCore_DP0 pid=10475) INFO 12-04 09:51:40 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:51:41.1288398Z (EngineCore_DP0 pid=10475) INFO 12-04 09:51:41 [gpu_model_runner.py:2902] Model loading took 3.2931 GiB and 0.140607 seconds 2025-12-04T09:51:41.1293386Z (EngineCore_DP0 pid=10475) INFO 12-04 09:51:41 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T09:51:41.1294360Z (EngineCore_DP0 pid=10475) INFO 12-04 09:51:41 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 20.00x 2025-12-04T09:51:42.2965087Z (EngineCore_DP0 pid=10475) INFO 12-04 09:51:42 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:51:42.2979507Z INFO 12-04 09:51:42 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:51:42.8197078Z PASSED 2025-12-04T09:51:42.8346694Z models/test_initialization.py::test_can_initialize_large_subset[JambaForCausalLM] Fork a new process to run a test 10545 2025-12-04T09:51:42.8355302Z Fork a new process to run a test 0 2025-12-04T09:51:42.8620050Z INFO 12-04 09:51:42 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='JambaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ai21labs/AI21-Jamba-1.5-Mini'} 2025-12-04T09:51:43.0898393Z 2025-12-04T09:51:43.0899283Z config.json: 0% 0.00/932 [00:00= mamba page size. 2025-12-04T09:51:49.8825007Z INFO 12-04 09:51:49 [config.py:453] Padding mamba page size by 5.26% to ensure that mamba page size and attention page size are exactly equal. 2025-12-04T09:51:49.9915040Z 2025-12-04T09:51:49.9916769Z tokenizer_config.json: 0% 0.00/14.2k [00:00. This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message. 2025-12-04T09:51:51.1036263Z 2025-12-04T09:51:51.1036641Z generation_config.json: 0% 0.00/126 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:51:52.4600177Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:52.4627083Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:52.4634063Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:52.4642523Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:52.4650514Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:52.4658573Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:51:52.4661390Z (EngineCore_DP0 pid=10624) INFO 12-04 09:51:52 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:51:52.7496367Z (EngineCore_DP0 pid=10624) INFO 12-04 09:51:52 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:51:52.8393097Z (EngineCore_DP0 pid=10624) INFO 12-04 09:51:52 [gpu_model_runner.py:2840] Starting to load model ai21labs/AI21-Jamba-1.5-Mini... 2025-12-04T09:51:53.1200128Z (EngineCore_DP0 pid=10624) INFO 12-04 09:51:53 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:51:53.7269449Z (EngineCore_DP0 pid=10624) INFO 12-04 09:51:53 [gpu_model_runner.py:2902] Model loading took 1.5245 GiB and 0.062215 seconds 2025-12-04T09:51:53.7274855Z (EngineCore_DP0 pid=10624) INFO 12-04 09:51:53 [kv_cache_utils.py:1199] GPU KV cache size: 8,589,934,592 tokens 2025-12-04T09:51:53.7275787Z (EngineCore_DP0 pid=10624) INFO 12-04 09:51:53 [kv_cache_utils.py:1204] Maximum concurrency for 262,144 tokens per request: 32768.00x 2025-12-04T09:51:54.3586190Z (EngineCore_DP0 pid=10624) INFO 12-04 09:51:54 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:51:54.3608882Z INFO 12-04 09:51:54 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:51:54.8385875Z PASSED 2025-12-04T09:51:54.8535010Z models/test_initialization.py::test_can_initialize_large_subset[ExaoneForCausalLM] Fork a new process to run a test 10694 2025-12-04T09:51:54.8544581Z Fork a new process to run a test 0 2025-12-04T09:51:54.8811076Z INFO 12-04 09:51:54 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='ExaoneForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct'} 2025-12-04T09:51:55.0648337Z 2025-12-04T09:51:55.0649219Z config.json: 0% 0.00/878 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:52:04.2077772Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:04.2104939Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:04.2112028Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:04.2118983Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:04.2127221Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:04.2134556Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:04.2137637Z (EngineCore_DP0 pid=10752) INFO 12-04 09:52:04 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:52:04.4965372Z (EngineCore_DP0 pid=10752) INFO 12-04 09:52:04 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:52:04.5630322Z (EngineCore_DP0 pid=10752) INFO 12-04 09:52:04 [gpu_model_runner.py:2840] Starting to load model LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct... 2025-12-04T09:52:04.8423012Z (EngineCore_DP0 pid=10752) INFO 12-04 09:52:04 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:52:04.9093679Z (EngineCore_DP0 pid=10752) INFO 12-04 09:52:04 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:52:05.5141410Z (EngineCore_DP0 pid=10752) INFO 12-04 09:52:05 [gpu_model_runner.py:2902] Model loading took 1.9698 GiB and 0.125616 seconds 2025-12-04T09:52:05.5145540Z (EngineCore_DP0 pid=10752) INFO 12-04 09:52:05 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T09:52:05.5146479Z (EngineCore_DP0 pid=10752) INFO 12-04 09:52:05 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 640.00x 2025-12-04T09:52:06.1549572Z (EngineCore_DP0 pid=10752) INFO 12-04 09:52:06 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:52:06.1572131Z INFO 12-04 09:52:06 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:52:06.6138591Z PASSED 2025-12-04T09:52:06.6288612Z models/test_initialization.py::test_can_initialize_large_subset[MistralModel] Fork a new process to run a test 10822 2025-12-04T09:52:06.6299039Z Fork a new process to run a test 0 2025-12-04T09:52:06.6567794Z INFO 12-04 09:52:06 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MistralModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'intfloat/e5-mistral-7b-instruct'} 2025-12-04T09:52:06.7354234Z 2025-12-04T09:52:06.7355776Z config.json: 0% 0.00/629 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:52:09.2299403Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:09.2326614Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:09.2333575Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:09.2341380Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:09.2349164Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:09.2356845Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:09.2359444Z (EngineCore_DP0 pid=10850) INFO 12-04 09:52:09 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:52:09.5207063Z (EngineCore_DP0 pid=10850) INFO 12-04 09:52:09 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:52:09.5678463Z (EngineCore_DP0 pid=10850) INFO 12-04 09:52:09 [gpu_model_runner.py:2840] Starting to load model intfloat/e5-mistral-7b-instruct... 2025-12-04T09:52:09.8466689Z (EngineCore_DP0 pid=10850) INFO 12-04 09:52:09 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:52:09.9139186Z (EngineCore_DP0 pid=10850) INFO 12-04 09:52:09 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:52:10.5250953Z (EngineCore_DP0 pid=10850) INFO 12-04 09:52:10 [gpu_model_runner.py:2902] Model loading took 0.6582 GiB and 0.127684 seconds 2025-12-04T09:52:10.5255262Z (EngineCore_DP0 pid=10850) INFO 12-04 09:52:10 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T09:52:10.5256163Z (EngineCore_DP0 pid=10850) INFO 12-04 09:52:10 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 213.06x 2025-12-04T09:52:11.1024723Z (EngineCore_DP0 pid=10850) INFO 12-04 09:52:11 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:52:11.1051935Z INFO 12-04 09:52:11 [llm.py:337] Supported tasks: ['embed'] 2025-12-04T09:52:11.5572862Z PASSED 2025-12-04T09:52:11.5723134Z models/test_initialization.py::test_can_initialize_large_subset[EagleDeepSeekMTPModel] Fork a new process to run a test 10920 2025-12-04T09:52:11.5731971Z Fork a new process to run a test 0 2025-12-04T09:52:11.6002665Z INFO 12-04 09:52:11 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='EagleDeepSeekMTPModel', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'eagle618/eagle-deepseek-v3-random', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'eagle618/deepseek-v3-random'} 2025-12-04T09:52:11.7683126Z 2025-12-04T09:52:11.7685327Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:52:11.7685709Z config.json: 1.95kB [00:00, 10.1MB/s] 2025-12-04T09:52:11.8210075Z INFO 12-04 09:52:11 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T09:52:11.8681825Z INFO 12-04 09:52:11 [model.py:653] Resolved architecture: DeepseekV3ForCausalLM 2025-12-04T09:52:11.8682293Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:52:11.8923146Z INFO 12-04 09:52:11 [model.py:1714] Using max model len 163840 2025-12-04T09:52:11.9856715Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T09:52:12.2185800Z 2025-12-04T09:52:12.2186602Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:52:12.2186936Z config.json: 1.79kB [00:00, 13.0MB/s] 2025-12-04T09:52:12.2506241Z INFO 12-04 09:52:12 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T09:52:12.3263624Z INFO 12-04 09:52:12 [model.py:653] Resolved architecture: DeepSeekMTPModel 2025-12-04T09:52:12.3265953Z INFO 12-04 09:52:12 [model.py:1714] Using max model len 163840 2025-12-04T09:52:12.3275492Z INFO 12-04 09:52:12 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:52:12.3623335Z 2025-12-04T09:52:12.3624988Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:52:12.3625342Z tokenizer_config.json: 3.13kB [00:00, 21.2MB/s] 2025-12-04T09:52:12.4732937Z 2025-12-04T09:52:12.5032175Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T09:52:12.5032511Z tokenizer.json: 7.85MB [00:00, 262MB/s] 2025-12-04T09:52:13.0448548Z INFO 12-04 09:52:13 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T09:52:13.0682251Z (EngineCore_DP0 pid=10928) INFO 12-04 09:52:13 [core.py:727] Waiting for init message from front-end. 2025-12-04T09:52:13.0744535Z (EngineCore_DP0 pid=10928) INFO 12-04 09:52:13 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='eagle618/deepseek-v3-random', speculative_config=SpeculativeConfig(method='eagle', model='eagle618/eagle-deepseek-v3-random', num_spec_tokens=1), tokenizer='eagle618/deepseek-v3-random', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=163840, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=fp8, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=eagle618/deepseek-v3-random, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+quant_fp8'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:52:14.4560010Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:14.4589393Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:14.4597491Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:14.4604487Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:14.4611965Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:14.4619828Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:14.4622805Z (EngineCore_DP0 pid=10928) INFO 12-04 09:52:14 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:52:14.7510337Z (EngineCore_DP0 pid=10928) INFO 12-04 09:52:14 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:52:14.7661309Z (EngineCore_DP0 pid=10928) WARNING 12-04 09:52:14 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-12-04T09:52:14.8251626Z (EngineCore_DP0 pid=10928) INFO 12-04 09:52:14 [gpu_model_runner.py:2840] Starting to load model eagle618/deepseek-v3-random... 2025-12-04T09:52:15.1096757Z (EngineCore_DP0 pid=10928) INFO 12-04 09:52:15 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:52:15.1903842Z (EngineCore_DP0 pid=10928) INFO 12-04 09:52:15 [cuda.py:328] Using Triton MLA backend on V1 engine. 2025-12-04T09:52:15.2061740Z (EngineCore_DP0 pid=10928) WARNING 12-04 09:52:15 [fp8.py:150] DeepGEMM backend requested but not available. 2025-12-04T09:52:15.2062566Z (EngineCore_DP0 pid=10928) INFO 12-04 09:52:15 [fp8.py:165] Using Triton backend for FP8 MoE 2025-12-04T09:52:15.2742624Z (EngineCore_DP0 pid=10928) WARNING 12-04 09:52:15 [fp8_utils.py:785] Using default W8A8 Block FP8 kernel config. Performance might be sub-optimal! Config file not found at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/model_executor/layers/quantization/utils/configs/N=8192,K=512,device_name=NVIDIA_L4,dtype=fp8_w8a8,block_shape=[128,128].json 2025-12-04T09:52:15.7723071Z (EngineCore_DP0 pid=10928) INFO 12-04 09:52:15 [gpu_model_runner.py:2879] Loading drafter model... 2025-12-04T09:52:15.7817864Z (EngineCore_DP0 pid=10928) INFO 12-04 09:52:15 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-12-04T09:52:15.7819413Z (EngineCore_DP0 pid=10928) INFO 12-04 09:52:15 [eagle.py:1027] Loading EAGLE LM head weights from the target model. 2025-12-04T09:52:16.3619908Z (EngineCore_DP0 pid=10928) INFO 12-04 09:52:16 [gpu_model_runner.py:2902] Model loading took 2.2133 GiB and 0.672375 seconds 2025-12-04T09:52:16.3625559Z (EngineCore_DP0 pid=10928) INFO 12-04 09:52:16 [kv_cache_utils.py:1199] GPU KV cache size: 4,660,336 tokens 2025-12-04T09:52:16.3626556Z (EngineCore_DP0 pid=10928) INFO 12-04 09:52:16 [kv_cache_utils.py:1204] Maximum concurrency for 163,840 tokens per request: 28.44x 2025-12-04T09:52:17.5379030Z (EngineCore_DP0 pid=10928) INFO 12-04 09:52:17 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:52:17.5392768Z INFO 12-04 09:52:17 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:52:18.0328786Z PASSED 2025-12-04T09:52:18.0479085Z models/test_initialization.py::test_can_initialize_large_subset[QWenLMHeadModel] Fork a new process to run a test 11002 2025-12-04T09:52:18.0487894Z Fork a new process to run a test 0 2025-12-04T09:52:18.0492118Z `transformers==4.56.2` installed, but `transformers<=4.53` is required to run this model. Reason: HF model uses remote code that is not compatible with latest Transformers 2025-12-04T09:52:18.3345148Z PASSED 2025-12-04T09:52:18.3496231Z models/test_initialization.py::test_can_initialize_large_subset[Dots1ForCausalLM] Fork a new process to run a test 11003 2025-12-04T09:52:18.3505514Z Fork a new process to run a test 0 2025-12-04T09:52:18.3783260Z INFO 12-04 09:52:18 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Dots1ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'rednote-hilab/dots.llm1.inst'} 2025-12-04T09:52:18.4554707Z 2025-12-04T09:52:18.4556550Z config.json: 0% 0.00/988 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:52:27.1117111Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:27.1143837Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:27.1150814Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:27.1157697Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:27.1165080Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:27.1172792Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:27.1175711Z (EngineCore_DP0 pid=11061) INFO 12-04 09:52:27 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:52:27.3943957Z (EngineCore_DP0 pid=11061) INFO 12-04 09:52:27 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:52:27.4628237Z (EngineCore_DP0 pid=11061) INFO 12-04 09:52:27 [gpu_model_runner.py:2840] Starting to load model rednote-hilab/dots.llm1.inst... 2025-12-04T09:52:27.7357037Z (EngineCore_DP0 pid=11061) INFO 12-04 09:52:27 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:52:27.8023335Z (EngineCore_DP0 pid=11061) INFO 12-04 09:52:27 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:52:28.3984537Z (EngineCore_DP0 pid=11061) INFO 12-04 09:52:28 [gpu_model_runner.py:2902] Model loading took 2.5821 GiB and 0.127136 seconds 2025-12-04T09:52:28.3989775Z (EngineCore_DP0 pid=11061) INFO 12-04 09:52:28 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-12-04T09:52:28.3990969Z (EngineCore_DP0 pid=11061) INFO 12-04 09:52:28 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 20.00x 2025-12-04T09:52:29.0514567Z (EngineCore_DP0 pid=11061) INFO 12-04 09:52:29 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:52:29.0537671Z INFO 12-04 09:52:29 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:52:29.5193274Z PASSED 2025-12-04T09:52:29.5343308Z models/test_initialization.py::test_can_initialize_large_subset[ModernBertModel] Fork a new process to run a test 11131 2025-12-04T09:52:29.5352476Z Fork a new process to run a test 0 2025-12-04T09:52:29.5618306Z INFO 12-04 09:52:29 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='ModernBertModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Alibaba-NLP/gte-modernbert-base'} 2025-12-04T09:52:29.6415110Z 2025-12-04T09:52:29.6417348Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:52:29.6417669Z config.json: 1.18kB [00:00, 6.77MB/s] 2025-12-04T09:52:36.1909227Z 2025-12-04T09:52:36.1911791Z modules.json: 0% 0.00/229 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:52:38.4706758Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:38.4737126Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:38.4744561Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:38.4752092Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:38.4759336Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:38.4766976Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:38.4769623Z (EngineCore_DP0 pid=11190) INFO 12-04 09:52:38 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:52:38.7567045Z (EngineCore_DP0 pid=11190) INFO 12-04 09:52:38 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:52:38.7770726Z (EngineCore_DP0 pid=11190) INFO 12-04 09:52:38 [gpu_model_runner.py:2840] Starting to load model Alibaba-NLP/gte-modernbert-base... 2025-12-04T09:52:39.0433001Z (EngineCore_DP0 pid=11190) INFO 12-04 09:52:39 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:52:39.1104239Z (EngineCore_DP0 pid=11190) INFO 12-04 09:52:39 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:52:39.7097779Z (EngineCore_DP0 pid=11190) INFO 12-04 09:52:39 [gpu_model_runner.py:2902] Model loading took 0.0826 GiB and 0.128138 seconds 2025-12-04T09:52:39.8879367Z (EngineCore_DP0 pid=11190) INFO 12-04 09:52:39 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T09:52:39.8907857Z (EngineCore_DP0 pid=11190) INFO 12-04 09:52:39 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T09:52:40.1690402Z (EngineCore_DP0 pid=11190) INFO 12-04 09:52:40 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:52:40.1710509Z INFO 12-04 09:52:40 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-12-04T09:52:40.6106117Z PASSED 2025-12-04T09:52:40.6257514Z models/test_initialization.py::test_can_initialize_large_subset[Ernie4_5_MoeForCausalLM] Fork a new process to run a test 11260 2025-12-04T09:52:40.6266732Z Fork a new process to run a test 0 2025-12-04T09:52:40.6544606Z INFO 12-04 09:52:40 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Ernie4_5_MoeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'baidu/ERNIE-4.5-21B-A3B-PT'} 2025-12-04T09:52:40.7690475Z 2025-12-04T09:52:40.7693153Z config.json: 0% 0.00/936 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:52:50.7476145Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:50.7502934Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:50.7509437Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:50.7518010Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:50.7525637Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:50.7533324Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:50.7536106Z (EngineCore_DP0 pid=11339) INFO 12-04 09:52:50 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:52:51.0409399Z (EngineCore_DP0 pid=11339) INFO 12-04 09:52:51 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:52:51.1243426Z (EngineCore_DP0 pid=11339) INFO 12-04 09:52:51 [gpu_model_runner.py:2840] Starting to load model baidu/ERNIE-4.5-21B-A3B-PT... 2025-12-04T09:52:51.4103464Z (EngineCore_DP0 pid=11339) INFO 12-04 09:52:51 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:52:51.4773801Z (EngineCore_DP0 pid=11339) INFO 12-04 09:52:51 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:52:52.0877222Z (EngineCore_DP0 pid=11339) INFO 12-04 09:52:52 [gpu_model_runner.py:2902] Model loading took 0.7310 GiB and 0.125072 seconds 2025-12-04T09:52:52.0881587Z (EngineCore_DP0 pid=11339) INFO 12-04 09:52:52 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T09:52:52.0882531Z (EngineCore_DP0 pid=11339) INFO 12-04 09:52:52 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 40.00x 2025-12-04T09:52:53.9905037Z (EngineCore_DP0 pid=11339) INFO 12-04 09:52:53 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:52:53.9919642Z INFO 12-04 09:52:53 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:52:54.4715062Z PASSED 2025-12-04T09:52:54.4865484Z models/test_initialization.py::test_can_initialize_large_subset[AquilaModel] Fork a new process to run a test 11409 2025-12-04T09:52:54.4874768Z Fork a new process to run a test 0 2025-12-04T09:52:54.5149252Z INFO 12-04 09:52:54 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='AquilaModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'BAAI/AquilaChat-7B'} 2025-12-04T09:52:54.6919877Z 2025-12-04T09:52:54.6920806Z config.json: 0% 0.00/707 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:52:57.2919444Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:57.2947046Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:57.2954572Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:57.2962471Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:57.2970195Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:57.2978449Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:52:57.2981870Z (EngineCore_DP0 pid=11417) INFO 12-04 09:52:57 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:52:57.5850217Z (EngineCore_DP0 pid=11417) INFO 12-04 09:52:57 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:52:57.6512834Z (EngineCore_DP0 pid=11417) INFO 12-04 09:52:57 [gpu_model_runner.py:2840] Starting to load model BAAI/AquilaChat-7B... 2025-12-04T09:52:57.9352176Z (EngineCore_DP0 pid=11417) INFO 12-04 09:52:57 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:52:58.0028472Z (EngineCore_DP0 pid=11417) INFO 12-04 09:52:58 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:52:58.6138470Z (EngineCore_DP0 pid=11417) INFO 12-04 09:52:58 [gpu_model_runner.py:2902] Model loading took 1.9048 GiB and 0.127605 seconds 2025-12-04T09:52:58.6143272Z (EngineCore_DP0 pid=11417) INFO 12-04 09:52:58 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-12-04T09:52:58.6144534Z (EngineCore_DP0 pid=11417) INFO 12-04 09:52:58 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 320.00x 2025-12-04T09:52:59.2447179Z (EngineCore_DP0 pid=11417) INFO 12-04 09:52:59 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:52:59.2468831Z INFO 12-04 09:52:59 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:52:59.7031628Z PASSED 2025-12-04T09:52:59.7181703Z models/test_initialization.py::test_can_initialize_large_subset[XLMRobertaForSequenceClassification] Fork a new process to run a test 11487 2025-12-04T09:52:59.7190673Z Fork a new process to run a test 0 2025-12-04T09:52:59.7465890Z INFO 12-04 09:52:59 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='XLMRobertaForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'BAAI/bge-reranker-v2-m3'} 2025-12-04T09:52:59.8170510Z 2025-12-04T09:52:59.8171755Z config.json: 0% 0.00/795 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:53:02.9984497Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:03.0015108Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:03.0022689Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:03.0030056Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:03.0037821Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:03.0045709Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:03.0048427Z (EngineCore_DP0 pid=11516) INFO 12-04 09:53:03 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:53:03.2844429Z (EngineCore_DP0 pid=11516) INFO 12-04 09:53:03 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:53:03.3109698Z (EngineCore_DP0 pid=11516) INFO 12-04 09:53:03 [gpu_model_runner.py:2840] Starting to load model BAAI/bge-reranker-v2-m3... 2025-12-04T09:53:03.5810564Z (EngineCore_DP0 pid=11516) INFO 12-04 09:53:03 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:53:03.5867379Z (EngineCore_DP0 pid=11516) INFO 12-04 09:53:03 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:53:04.1909361Z (EngineCore_DP0 pid=11516) INFO 12-04 09:53:04 [gpu_model_runner.py:2902] Model loading took 0.5200 GiB and 0.067528 seconds 2025-12-04T09:53:04.9712780Z (EngineCore_DP0 pid=11516) INFO 12-04 09:53:04 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T09:53:04.9769040Z (EngineCore_DP0 pid=11516) INFO 12-04 09:53:04 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T09:53:05.2602477Z (EngineCore_DP0 pid=11516) INFO 12-04 09:53:05 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:53:05.2619760Z INFO 12-04 09:53:05 [llm.py:337] Supported tasks: ['encode', 'classify', 'score'] 2025-12-04T09:53:05.7517822Z PASSED 2025-12-04T09:53:05.7668554Z models/test_initialization.py::test_can_initialize_large_subset[MiniCPMV] Fork a new process to run a test 11586 2025-12-04T09:53:05.7678242Z Fork a new process to run a test 0 2025-12-04T09:53:05.7952349Z INFO 12-04 09:53:05 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MiniCPMV', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'openbmb/MiniCPM-Llama3-V-2_5'} 2025-12-04T09:53:05.8842975Z 2025-12-04T09:53:05.8844917Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:53:05.8845231Z config.json: 1.39kB [00:00, 7.37MB/s] 2025-12-04T09:53:05.9509501Z 2025-12-04T09:53:05.9511282Z configuration_minicpm.py: 0.00B [00:00, ?B/s] 2025-12-04T09:53:05.9511685Z configuration_minicpm.py: 4.06kB [00:00, 18.6MB/s] 2025-12-04T09:53:05.9612323Z A new version of the following files was downloaded from https://huggingface.co/openbmb/MiniCPM-Llama3-V-2_5: 2025-12-04T09:53:05.9612861Z - configuration_minicpm.py 2025-12-04T09:53:05.9613510Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:53:06.2114841Z 2025-12-04T09:53:06.2115336Z preprocessor_config.json: 0% 0.00/599 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:53:15.2201073Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:15.2230187Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:15.2237701Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:15.2245880Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:15.2254677Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:15.2262957Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:15.2265690Z (EngineCore_DP0 pid=11644) INFO 12-04 09:53:15 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:53:15.5265797Z (EngineCore_DP0 pid=11644) INFO 12-04 09:53:15 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:53:15.7463320Z (EngineCore_DP0 pid=11644) 2025-12-04T09:53:15.7464182Z processing_minicpmv.py: 0.00B [00:00, ?B/s] 2025-12-04T09:53:15.7464571Z processing_minicpmv.py: 11.9kB [00:00, 55.7MB/s] 2025-12-04T09:53:15.7886402Z (EngineCore_DP0 pid=11644) 2025-12-04T09:53:15.7888574Z image_processing_minicpmv.py: 0.00B [00:00, ?B/s] 2025-12-04T09:53:15.7889001Z image_processing_minicpmv.py: 15.5kB [00:00, 74.7MB/s] 2025-12-04T09:53:15.8826135Z (EngineCore_DP0 pid=11644) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-12-04T09:53:18.2899884Z (EngineCore_DP0 pid=11644) INFO 12-04 09:53:18 [gpu_model_runner.py:2840] Starting to load model openbmb/MiniCPM-Llama3-V-2_5... 2025-12-04T09:53:18.6552369Z (EngineCore_DP0 pid=11644) INFO 12-04 09:53:18 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:53:18.6793971Z (EngineCore_DP0 pid=11644) INFO 12-04 09:53:18 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:53:18.7548505Z (EngineCore_DP0 pid=11644) INFO 12-04 09:53:18 [layer.py:497] MultiHeadAttention attn_backend: _Backend.XFORMERS, use_upstream_fa: False 2025-12-04T09:53:19.4239226Z (EngineCore_DP0 pid=11644) INFO 12-04 09:53:19 [gpu_model_runner.py:2902] Model loading took 2.6087 GiB and 0.191056 seconds 2025-12-04T09:53:19.4243801Z (EngineCore_DP0 pid=11644) INFO 12-04 09:53:19 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T09:53:19.4245537Z (EngineCore_DP0 pid=11644) INFO 12-04 09:53:19 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 320.00x 2025-12-04T09:53:20.2899959Z (EngineCore_DP0 pid=11644) INFO 12-04 09:53:20 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:53:20.2915483Z INFO 12-04 09:53:20 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:53:20.7808810Z PASSED 2025-12-04T09:53:20.7958679Z models/test_initialization.py::test_can_initialize_large_subset[Qwen3NextMTP] Fork a new process to run a test 11730 2025-12-04T09:53:20.7968660Z Fork a new process to run a test 0 2025-12-04T09:53:20.7972153Z `transformers==4.56.2` installed, but `transformers>=4.56.3` is required to run this model. 2025-12-04T09:53:21.0823652Z PASSED 2025-12-04T09:53:21.0972964Z models/test_initialization.py::test_can_initialize_large_subset[JAISLMHeadModel] Fork a new process to run a test 11731 2025-12-04T09:53:21.0981922Z Fork a new process to run a test 0 2025-12-04T09:53:21.1249057Z INFO 12-04 09:53:21 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='JAISLMHeadModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'inceptionai/jais-13b-chat'} 2025-12-04T09:53:21.3094464Z 2025-12-04T09:53:21.3095515Z config.json: 0% 0.00/1.26k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:53:30.2695851Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:30.2723037Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:30.2730217Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:30.2738054Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:30.2746187Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:30.2753647Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:30.2756821Z (EngineCore_DP0 pid=11789) INFO 12-04 09:53:30 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:53:30.5559264Z (EngineCore_DP0 pid=11789) INFO 12-04 09:53:30 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:53:30.6508751Z (EngineCore_DP0 pid=11789) INFO 12-04 09:53:30 [gpu_model_runner.py:2840] Starting to load model inceptionai/jais-13b-chat... 2025-12-04T09:53:30.9228748Z (EngineCore_DP0 pid=11789) INFO 12-04 09:53:30 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:53:30.9257465Z (EngineCore_DP0 pid=11789) INFO 12-04 09:53:30 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:53:31.5239793Z (EngineCore_DP0 pid=11789) INFO 12-04 09:53:31 [gpu_model_runner.py:2902] Model loading took 1.3986 GiB and 0.062409 seconds 2025-12-04T09:53:31.5243737Z (EngineCore_DP0 pid=11789) INFO 12-04 09:53:31 [kv_cache_utils.py:1199] GPU KV cache size: 524,288 tokens 2025-12-04T09:53:31.5244669Z (EngineCore_DP0 pid=11789) INFO 12-04 09:53:31 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 256.00x 2025-12-04T09:53:32.1089314Z (EngineCore_DP0 pid=11789) INFO 12-04 09:53:32 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:53:32.1110558Z INFO 12-04 09:53:32 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:53:32.5687056Z PASSED 2025-12-04T09:53:32.5834699Z models/test_initialization.py::test_can_initialize_large_subset[Gemma3nForConditionalGeneration] Fork a new process to run a test 11859 2025-12-04T09:53:32.5843156Z Fork a new process to run a test 0 2025-12-04T09:53:32.6108691Z INFO 12-04 09:53:32 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Gemma3nForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'google/gemma-3n-E2B-it'} 2025-12-04T09:53:32.9671429Z INFO 12-04 09:53:32 [model.py:653] Resolved architecture: Gemma3nForConditionalGeneration 2025-12-04T09:53:32.9672257Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:53:32.9913559Z INFO 12-04 09:53:32 [model.py:1714] Using max model len 32768 2025-12-04T09:53:33.1572239Z INFO 12-04 09:53:33 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:53:36.3609486Z (EngineCore_DP0 pid=11866) INFO 12-04 09:53:36 [core.py:727] Waiting for init message from front-end. 2025-12-04T09:53:36.3645644Z (EngineCore_DP0 pid=11866) INFO 12-04 09:53:36 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='google/gemma-3n-E2B-it', speculative_config=None, tokenizer='google/gemma-3n-E2B-it', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.bfloat16, max_seq_len=32768, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=google/gemma-3n-E2B-it, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:53:37.7286221Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:37.7315517Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:37.7322725Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:37.7330897Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:37.7338972Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:37.7347204Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:37.7349993Z (EngineCore_DP0 pid=11866) INFO 12-04 09:53:37 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:53:38.0361518Z (EngineCore_DP0 pid=11866) INFO 12-04 09:53:38 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:53:44.3674512Z (EngineCore_DP0 pid=11866) INFO 12-04 09:53:44 [gpu_model_runner.py:2840] Starting to load model google/gemma-3n-E2B-it... 2025-12-04T09:53:44.8062464Z (EngineCore_DP0 pid=11866) INFO 12-04 09:53:44 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:53:45.0246470Z (EngineCore_DP0 pid=11866) INFO 12-04 09:53:45 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:53:45.7027925Z (EngineCore_DP0 pid=11866) INFO 12-04 09:53:45 [gpu_model_runner.py:2902] Model loading took 3.6358 GiB and 0.319489 seconds 2025-12-04T09:53:45.7032867Z (EngineCore_DP0 pid=11866) INFO 12-04 09:53:45 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T09:53:45.7033761Z (EngineCore_DP0 pid=11866) INFO 12-04 09:53:45 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 300.62x 2025-12-04T09:53:47.4975151Z (EngineCore_DP0 pid=11866) INFO 12-04 09:53:47 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:53:47.4990647Z INFO 12-04 09:53:47 [llm.py:337] Supported tasks: ['generate', 'transcription'] 2025-12-04T09:53:48.0446027Z PASSED 2025-12-04T09:53:48.0594513Z models/test_initialization.py::test_can_initialize_large_subset[Glm4MoeMTPModel] Fork a new process to run a test 11967 2025-12-04T09:53:48.0604271Z Fork a new process to run a test 0 2025-12-04T09:53:48.0606237Z Model is not available online 2025-12-04T09:53:48.3431921Z PASSED 2025-12-04T09:53:48.3580660Z models/test_initialization.py::test_can_initialize_large_subset[GteModel] Fork a new process to run a test 11968 2025-12-04T09:53:48.3591055Z Fork a new process to run a test 0 2025-12-04T09:53:48.3862452Z INFO 12-04 09:53:48 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GteModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Snowflake/snowflake-arctic-embed-m-v2.0'} 2025-12-04T09:53:48.4797570Z 2025-12-04T09:53:48.4799525Z config.json: 0% 0.00/971 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:53:58.1907553Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:58.1935974Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:58.1944232Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:58.1951913Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:58.1960354Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:58.1968311Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:53:58.1971641Z (EngineCore_DP0 pid=12049) INFO 12-04 09:53:58 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:53:58.4818409Z (EngineCore_DP0 pid=12049) INFO 12-04 09:53:58 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:53:58.5083438Z (EngineCore_DP0 pid=12049) INFO 12-04 09:53:58 [gpu_model_runner.py:2840] Starting to load model Snowflake/snowflake-arctic-embed-m-v2.0... 2025-12-04T09:53:58.7787750Z (EngineCore_DP0 pid=12049) INFO 12-04 09:53:58 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:53:58.8445976Z (EngineCore_DP0 pid=12049) INFO 12-04 09:53:58 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:53:59.4596200Z (EngineCore_DP0 pid=12049) INFO 12-04 09:53:59 [gpu_model_runner.py:2902] Model loading took 0.3769 GiB and 0.129779 seconds 2025-12-04T09:54:00.3754498Z (EngineCore_DP0 pid=12049) INFO 12-04 09:54:00 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T09:54:00.4588160Z (EngineCore_DP0 pid=12049) INFO 12-04 09:54:00 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T09:54:00.7428354Z (EngineCore_DP0 pid=12049) INFO 12-04 09:54:00 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:54:00.7443847Z INFO 12-04 09:54:00 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-12-04T09:54:01.2437993Z PASSED 2025-12-04T09:54:01.2585364Z models/test_initialization.py::test_can_initialize_large_subset[Mamba2ForCausalLM] Fork a new process to run a test 12119 2025-12-04T09:54:01.2594505Z Fork a new process to run a test 0 2025-12-04T09:54:01.2870070Z INFO 12-04 09:54:01 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Mamba2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'mistralai/Mamba-Codestral-7B-v0.1'} 2025-12-04T09:54:01.3616732Z 2025-12-04T09:54:01.3618557Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:54:01.3618874Z config.json: 1.01kB [00:00, 6.66MB/s] 2025-12-04T09:54:07.8238800Z INFO 12-04 09:54:07 [model.py:653] Resolved architecture: Mamba2ForCausalLM 2025-12-04T09:54:07.8239284Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:54:07.8494770Z WARNING 12-04 09:54:07 [model.py:2042] The model's config.json does not contain any of the following keys to determine the original maximum length of the model: ['max_position_embeddings', 'n_positions', 'max_seq_len', 'seq_length', 'model_max_length', 'max_target_positions', 'max_sequence_length', 'max_seq_length', 'seq_len']. Assuming the model's maximum length is 2048. 2025-12-04T09:54:07.8496173Z INFO 12-04 09:54:07 [model.py:1714] Using max model len 2048 2025-12-04T09:54:08.0213863Z INFO 12-04 09:54:08 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:54:08.0214779Z INFO 12-04 09:54:08 [config.py:300] Warning: Prefix caching is currently enabled. Its support for Mamba2 layers is experimental. Please report any issues you may observe. 2025-12-04T09:54:08.0216012Z INFO 12-04 09:54:08 [config.py:313] Disabling cascade attention since it is not supported for hybrid models. 2025-12-04T09:54:08.0947148Z 2025-12-04T09:54:08.0952757Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:54:08.0953131Z tokenizer_config.json: 137kB [00:00, 241MB/s] 2025-12-04T09:54:08.2181019Z 2025-12-04T09:54:08.4967689Z tokenizer.model: 0% 0.00/588k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:54:10.4064672Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:10.4091891Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:10.4098837Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:10.4106655Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:10.4113703Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:10.4121284Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:10.4123865Z (EngineCore_DP0 pid=12197) INFO 12-04 09:54:10 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:54:10.6973285Z (EngineCore_DP0 pid=12197) INFO 12-04 09:54:10 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:54:10.9030259Z (EngineCore_DP0 pid=12197) INFO 12-04 09:54:10 [gpu_model_runner.py:2840] Starting to load model mistralai/Mamba-Codestral-7B-v0.1... 2025-12-04T09:54:11.1831179Z (EngineCore_DP0 pid=12197) INFO 12-04 09:54:11 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:54:11.7976347Z (EngineCore_DP0 pid=12197) INFO 12-04 09:54:11 [gpu_model_runner.py:2902] Model loading took 0.7052 GiB and 0.061400 seconds 2025-12-04T09:54:11.7981444Z (EngineCore_DP0 pid=12197) INFO 12-04 09:54:11 [kv_cache_utils.py:1199] GPU KV cache size: 10,186,752 tokens 2025-12-04T09:54:11.7982205Z (EngineCore_DP0 pid=12197) INFO 12-04 09:54:11 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 4974.00x 2025-12-04T09:54:12.4718738Z (EngineCore_DP0 pid=12197) INFO 12-04 09:54:12 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:54:12.4742144Z INFO 12-04 09:54:12 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:54:12.9091330Z PASSED 2025-12-04T09:54:12.9238646Z models/test_initialization.py::test_can_initialize_large_subset[GraniteMoeHybridForCausalLM] Fork a new process to run a test 12267 2025-12-04T09:54:12.9247919Z Fork a new process to run a test 0 2025-12-04T09:54:12.9519642Z INFO 12-04 09:54:12 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GraniteMoeHybridForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ibm-granite/granite-4.0-tiny-preview'} 2025-12-04T09:54:13.0371803Z 2025-12-04T09:54:13.0373934Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:54:13.0374282Z config.json: 1.80kB [00:00, 9.54MB/s] 2025-12-04T09:54:19.5475656Z INFO 12-04 09:54:19 [model.py:653] Resolved architecture: GraniteMoeHybridForCausalLM 2025-12-04T09:54:19.5476498Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:54:19.6445444Z 2025-12-04T09:54:19.6448342Z model.safetensors.index.json: 0.00B [00:00, ?B/s] 2025-12-04T09:54:19.6448741Z model.safetensors.index.json: 48.9kB [00:00, 105MB/s] 2025-12-04T09:54:19.6486868Z 2025-12-04T09:54:19.6951408Z Parse safetensors files: 0% 0/3 [00:00= mamba page size. 2025-12-04T09:54:19.9516498Z INFO 12-04 09:54:19 [config.py:453] Padding mamba page size by 1.59% to ensure that mamba page size and attention page size are exactly equal. 2025-12-04T09:54:19.9824531Z 2025-12-04T09:54:19.9826861Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:54:19.9827199Z tokenizer_config.json: 10.1kB [00:00, 44.8MB/s] 2025-12-04T09:54:20.1526735Z 2025-12-04T09:54:20.1680275Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T09:54:20.1680604Z tokenizer.json: 3.48MB [00:00, 225MB/s] 2025-12-04T09:54:20.2549161Z 2025-12-04T09:54:20.2550265Z special_tokens_map.json: 0% 0.00/801 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:54:21.7351135Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:21.7378401Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:21.7386011Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:21.7393096Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:21.7400391Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:21.7407921Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:21.7410625Z (EngineCore_DP0 pid=12328) INFO 12-04 09:54:21 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:54:22.0256896Z (EngineCore_DP0 pid=12328) INFO 12-04 09:54:22 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:54:22.0836340Z (EngineCore_DP0 pid=12328) INFO 12-04 09:54:22 [gpu_model_runner.py:2840] Starting to load model ibm-granite/granite-4.0-tiny-preview... 2025-12-04T09:54:22.3595977Z (EngineCore_DP0 pid=12328) INFO 12-04 09:54:22 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:54:22.9623985Z (EngineCore_DP0 pid=12328) INFO 12-04 09:54:22 [gpu_model_runner.py:2902] Model loading took 0.1857 GiB and 0.061840 seconds 2025-12-04T09:54:22.9630162Z (EngineCore_DP0 pid=12328) INFO 12-04 09:54:22 [kv_cache_utils.py:1199] GPU KV cache size: 1,717,960,704 tokens 2025-12-04T09:54:22.9631143Z (EngineCore_DP0 pid=12328) INFO 12-04 09:54:22 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 13107.00x 2025-12-04T09:54:23.4967633Z (EngineCore_DP0 pid=12328) INFO 12-04 09:54:23 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:54:23.4991621Z INFO 12-04 09:54:23 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:54:23.9395019Z PASSED 2025-12-04T09:54:23.9543115Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2VLForConditionalGeneration] Fork a new process to run a test 12398 2025-12-04T09:54:23.9553319Z Fork a new process to run a test 0 2025-12-04T09:54:23.9830250Z INFO 12-04 09:54:23 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2VLForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen2-VL-2B-Instruct'} 2025-12-04T09:54:24.2200706Z 2025-12-04T09:54:24.2202415Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:54:24.2202713Z config.json: 1.20kB [00:00, 7.83MB/s] 2025-12-04T09:54:24.3191528Z 2025-12-04T09:54:24.3192455Z preprocessor_config.json: 0% 0.00/347 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:54:26.9307980Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:26.9348441Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:26.9355802Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:26.9363338Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:26.9371364Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:26.9379109Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:26.9382291Z (EngineCore_DP0 pid=12406) INFO 12-04 09:54:26 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:54:27.2205733Z (EngineCore_DP0 pid=12406) INFO 12-04 09:54:27 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:54:28.6739553Z (EngineCore_DP0 pid=12406) 2025-12-04T09:54:28.6740919Z chat_template.json: 0.00B [00:00, ?B/s] 2025-12-04T09:54:28.6741358Z chat_template.json: 1.05kB [00:00, 3.28MB/s] 2025-12-04T09:54:29.2086028Z (EngineCore_DP0 pid=12406) INFO 12-04 09:54:29 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen2-VL-2B-Instruct... 2025-12-04T09:54:29.5606450Z (EngineCore_DP0 pid=12406) INFO 12-04 09:54:29 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:54:29.6294765Z (EngineCore_DP0 pid=12406) INFO 12-04 09:54:29 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:54:30.2397815Z (EngineCore_DP0 pid=12406) INFO 12-04 09:54:30 [gpu_model_runner.py:2902] Model loading took 1.7941 GiB and 0.134234 seconds 2025-12-04T09:54:30.2401909Z (EngineCore_DP0 pid=12406) INFO 12-04 09:54:30 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T09:54:30.2402846Z (EngineCore_DP0 pid=12406) INFO 12-04 09:54:30 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 320.00x 2025-12-04T09:54:31.8075394Z (EngineCore_DP0 pid=12406) INFO 12-04 09:54:31 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:54:31.8090501Z INFO 12-04 09:54:31 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:54:32.2807064Z PASSED 2025-12-04T09:54:32.2952988Z models/test_initialization.py::test_can_initialize_large_subset[GraniteForCausalLM] Fork a new process to run a test 12476 2025-12-04T09:54:32.2962783Z Fork a new process to run a test 0 2025-12-04T09:54:32.3228183Z INFO 12-04 09:54:32 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GraniteForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ibm/PowerLM-3b'} 2025-12-04T09:54:32.5029158Z 2025-12-04T09:54:32.5030509Z config.json: 0% 0.00/755 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:54:41.6187143Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:41.6214891Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:41.6222250Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:41.6229526Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:41.6237711Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:41.6245777Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:41.6248688Z (EngineCore_DP0 pid=12537) INFO 12-04 09:54:41 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:54:41.9098795Z (EngineCore_DP0 pid=12537) INFO 12-04 09:54:41 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:54:41.9730928Z (EngineCore_DP0 pid=12537) INFO 12-04 09:54:41 [gpu_model_runner.py:2840] Starting to load model ibm/PowerLM-3b... 2025-12-04T09:54:42.2510355Z (EngineCore_DP0 pid=12537) INFO 12-04 09:54:42 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:54:42.3177771Z (EngineCore_DP0 pid=12537) INFO 12-04 09:54:42 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:54:42.9245282Z (EngineCore_DP0 pid=12537) INFO 12-04 09:54:42 [gpu_model_runner.py:2902] Model loading took 0.3707 GiB and 0.124931 seconds 2025-12-04T09:54:42.9249641Z (EngineCore_DP0 pid=12537) INFO 12-04 09:54:42 [kv_cache_utils.py:1199] GPU KV cache size: 1,165,072 tokens 2025-12-04T09:54:42.9250422Z (EngineCore_DP0 pid=12537) INFO 12-04 09:54:42 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 284.44x 2025-12-04T09:54:43.7475987Z (EngineCore_DP0 pid=12537) INFO 12-04 09:54:43 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:54:43.7499480Z INFO 12-04 09:54:43 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:54:44.1944179Z PASSED 2025-12-04T09:54:44.2090734Z models/test_initialization.py::test_can_initialize_large_subset[BailingMoeForCausalLM] Fork a new process to run a test 12607 2025-12-04T09:54:44.2101033Z Fork a new process to run a test 0 2025-12-04T09:54:44.2381125Z INFO 12-04 09:54:44 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='BailingMoeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'inclusionAI/Ling-lite-1.5'} 2025-12-04T09:54:44.4771411Z 2025-12-04T09:54:44.4773621Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:54:44.4774083Z config.json: 1.28kB [00:00, 7.12MB/s] 2025-12-04T09:54:44.5429921Z 2025-12-04T09:54:44.5431494Z configuration_bailing_moe.py: 0.00B [00:00, ?B/s] 2025-12-04T09:54:44.5431907Z configuration_bailing_moe.py: 2.86kB [00:00, 16.7MB/s] 2025-12-04T09:54:44.5531388Z A new version of the following files was downloaded from https://huggingface.co/inclusionAI/Ling-lite-1.5: 2025-12-04T09:54:44.5531943Z - configuration_bailing_moe.py 2025-12-04T09:54:44.5532577Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:54:51.1452099Z INFO 12-04 09:54:51 [model.py:653] Resolved architecture: BailingMoeForCausalLM 2025-12-04T09:54:51.1452632Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:54:51.1707716Z INFO 12-04 09:54:51 [model.py:1714] Using max model len 32768 2025-12-04T09:54:51.1709753Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T09:54:51.2286492Z INFO 12-04 09:54:51 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:54:51.2761261Z 2025-12-04T09:54:51.2765697Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:54:51.2766160Z tokenizer_config.json: 51.1kB [00:00, 102MB/s] 2025-12-04T09:54:51.3567869Z 2025-12-04T09:54:51.3930242Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T09:54:51.3930648Z tokenizer.json: 9.75MB [00:00, 270MB/s] 2025-12-04T09:54:51.4884399Z 2025-12-04T09:54:51.4886072Z special_tokens_map.json: 0% 0.00/747 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:54:53.3956720Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:53.3984749Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:53.3992233Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:53.3999605Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:53.4007266Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:53.4015189Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:53.4018156Z (EngineCore_DP0 pid=12665) INFO 12-04 09:54:53 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:54:53.6825165Z (EngineCore_DP0 pid=12665) INFO 12-04 09:54:53 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:54:53.7441604Z (EngineCore_DP0 pid=12665) INFO 12-04 09:54:53 [gpu_model_runner.py:2840] Starting to load model inclusionAI/Ling-lite-1.5... 2025-12-04T09:54:54.0170522Z (EngineCore_DP0 pid=12665) INFO 12-04 09:54:54 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:54:54.0846541Z (EngineCore_DP0 pid=12665) INFO 12-04 09:54:54 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:54:54.6810081Z (EngineCore_DP0 pid=12665) INFO 12-04 09:54:54 [gpu_model_runner.py:2902] Model loading took 1.0567 GiB and 0.126909 seconds 2025-12-04T09:54:54.6814521Z (EngineCore_DP0 pid=12665) INFO 12-04 09:54:54 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T09:54:54.6815570Z (EngineCore_DP0 pid=12665) INFO 12-04 09:54:54 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 426.11x 2025-12-04T09:54:55.8217838Z (EngineCore_DP0 pid=12665) INFO 12-04 09:54:55 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:54:55.8234572Z INFO 12-04 09:54:55 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:54:56.2794147Z PASSED 2025-12-04T09:54:56.2941951Z models/test_initialization.py::test_can_initialize_large_subset[BertModel] Fork a new process to run a test 12735 2025-12-04T09:54:56.2952640Z Fork a new process to run a test 0 2025-12-04T09:54:56.3226731Z INFO 12-04 09:54:56 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='BertModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'BAAI/bge-base-en-v1.5'} 2025-12-04T09:54:56.4320541Z 2025-12-04T09:54:56.4321126Z config.json: 0% 0.00/777 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:54:58.4674863Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:58.4705733Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:58.4713506Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:58.4721346Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:58.4729803Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:58.4737454Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:54:58.4740009Z (EngineCore_DP0 pid=12743) INFO 12-04 09:54:58 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:54:58.7538603Z (EngineCore_DP0 pid=12743) INFO 12-04 09:54:58 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:54:58.7640713Z (EngineCore_DP0 pid=12743) INFO 12-04 09:54:58 [gpu_model_runner.py:2840] Starting to load model BAAI/bge-base-en-v1.5... 2025-12-04T09:54:59.0286250Z (EngineCore_DP0 pid=12743) INFO 12-04 09:54:59 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:54:59.0332090Z (EngineCore_DP0 pid=12743) INFO 12-04 09:54:59 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:54:59.6300572Z (EngineCore_DP0 pid=12743) INFO 12-04 09:54:59 [gpu_model_runner.py:2902] Model loading took 0.0579 GiB and 0.065544 seconds 2025-12-04T09:54:59.7266040Z (EngineCore_DP0 pid=12743) INFO 12-04 09:54:59 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T09:54:59.7294648Z (EngineCore_DP0 pid=12743) INFO 12-04 09:54:59 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T09:55:00.0065870Z (EngineCore_DP0 pid=12743) INFO 12-04 09:55:00 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:55:00.0079572Z INFO 12-04 09:55:00 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-12-04T09:55:00.4303492Z PASSED 2025-12-04T09:55:00.4450288Z models/test_initialization.py::test_can_initialize_large_subset[MolmoForCausalLM] Fork a new process to run a test 12813 2025-12-04T09:55:00.4459990Z Fork a new process to run a test 0 2025-12-04T09:55:00.4463969Z `transformers==4.56.2` installed, but `transformers<=4.48` is required to run this model. Reason: Incorrectly-detected `tensorflow` import. 2025-12-04T09:55:00.7255679Z PASSED 2025-12-04T09:55:00.7401177Z models/test_initialization.py::test_can_initialize_large_subset[DeepseekVLV2ForCausalLM] Fork a new process to run a test 12814 2025-12-04T09:55:00.7410418Z Fork a new process to run a test 0 2025-12-04T09:55:00.7414022Z `transformers==4.56.2` installed, but `transformers<=4.48` is required to run this model. Reason: HF model is not compatible. 2025-12-04T09:55:01.0209990Z PASSED 2025-12-04T09:55:01.0357294Z models/test_initialization.py::test_can_initialize_large_subset[Plamo2ForCausalLM] Fork a new process to run a test 12815 2025-12-04T09:55:01.0367279Z Fork a new process to run a test 0 2025-12-04T09:55:01.0371923Z `transformers==4.56.2` installed, but `transformers<=4.55.4` is required to run this model. Reason: HF model uses remote code that is not compatible with latest Transformers 2025-12-04T09:55:01.3145227Z PASSED 2025-12-04T09:55:01.3292118Z models/test_initialization.py::test_can_initialize_large_subset[DotsOCRForCausalLM] Fork a new process to run a test 12816 2025-12-04T09:55:01.3302006Z Fork a new process to run a test 0 2025-12-04T09:55:01.3572795Z INFO 12-04 09:55:01 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='DotsOCRForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'rednote-hilab/dots.ocr'} 2025-12-04T09:55:01.4330535Z 2025-12-04T09:55:01.4332714Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:55:01.4333042Z config.json: 1.47kB [00:00, 8.59MB/s] 2025-12-04T09:55:01.5050688Z 2025-12-04T09:55:01.5053032Z configuration_dots.py: 0.00B [00:00, ?B/s] 2025-12-04T09:55:01.5053512Z configuration_dots.py: 3.04kB [00:00, 17.5MB/s] 2025-12-04T09:55:01.5152331Z A new version of the following files was downloaded from https://huggingface.co/rednote-hilab/dots.ocr: 2025-12-04T09:55:01.5152893Z - configuration_dots.py 2025-12-04T09:55:01.5153497Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:55:01.7207690Z 2025-12-04T09:55:01.7208177Z preprocessor_config.json: 0% 0.00/432 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:55:10.5284880Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:10.5313815Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:10.5321353Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:10.5329597Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:10.5337741Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:10.5346673Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:10.5349602Z (EngineCore_DP0 pid=12874) INFO 12-04 09:55:10 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:55:10.8277933Z (EngineCore_DP0 pid=12874) INFO 12-04 09:55:10 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:55:10.9624464Z (EngineCore_DP0 pid=12874) The image processor of type `Qwen2VLImageProcessor` is now loaded as a fast processor by default, even if the model checkpoint was saved with a slow processor. This is a breaking change and may produce slightly different outputs. To continue using the slow processor, instantiate this class with `use_fast=False`. Note that this behavior will be extended to all models in a future release. 2025-12-04T09:55:12.3120784Z (EngineCore_DP0 pid=12874) 2025-12-04T09:55:12.3122884Z chat_template.json: 0.00B [00:00, ?B/s] 2025-12-04T09:55:12.3123587Z chat_template.json: 1.11kB [00:00, 3.26MB/s] 2025-12-04T09:55:12.9105411Z (EngineCore_DP0 pid=12874) INFO 12-04 09:55:12 [gpu_model_runner.py:2840] Starting to load model rednote-hilab/dots.ocr... 2025-12-04T09:55:13.1980411Z (EngineCore_DP0 pid=12874) INFO 12-04 09:55:13 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:55:13.2282415Z (EngineCore_DP0 pid=12874) INFO 12-04 09:55:13 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:55:13.8585816Z (EngineCore_DP0 pid=12874) INFO 12-04 09:55:13 [gpu_model_runner.py:2902] Model loading took 1.1334 GiB and 0.088607 seconds 2025-12-04T09:55:13.8590610Z (EngineCore_DP0 pid=12874) INFO 12-04 09:55:13 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T09:55:13.8591741Z (EngineCore_DP0 pid=12874) INFO 12-04 09:55:13 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 80.00x 2025-12-04T09:55:15.4797785Z (EngineCore_DP0 pid=12874) INFO 12-04 09:55:15 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:55:15.4817179Z INFO 12-04 09:55:15 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:55:15.9605866Z PASSED 2025-12-04T09:55:15.9751451Z models/test_initialization.py::test_can_initialize_large_subset[Gemma2ForCausalLM] Fork a new process to run a test 12944 2025-12-04T09:55:15.9761279Z Fork a new process to run a test 0 2025-12-04T09:55:16.0032001Z INFO 12-04 09:55:16 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Gemma2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'google/gemma-2-9b'} 2025-12-04T09:55:16.1215598Z 2025-12-04T09:55:16.1216673Z config.json: 0% 0.00/856 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:55:19.9426107Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:19.9454329Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:19.9461953Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:19.9469519Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:19.9477447Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:19.9484649Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:19.9487642Z (EngineCore_DP0 pid=12975) INFO 12-04 09:55:19 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:55:20.2368797Z (EngineCore_DP0 pid=12975) INFO 12-04 09:55:20 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:55:20.3140107Z (EngineCore_DP0 pid=12975) INFO 12-04 09:55:20 [gpu_model_runner.py:2840] Starting to load model google/gemma-2-9b... 2025-12-04T09:55:20.5911972Z (EngineCore_DP0 pid=12975) INFO 12-04 09:55:20 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:55:20.6611374Z (EngineCore_DP0 pid=12975) INFO 12-04 09:55:20 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:55:21.2680759Z (EngineCore_DP0 pid=12975) INFO 12-04 09:55:21 [gpu_model_runner.py:2902] Model loading took 2.0821 GiB and 0.129393 seconds 2025-12-04T09:55:21.2685650Z (EngineCore_DP0 pid=12975) INFO 12-04 09:55:21 [kv_cache_utils.py:1199] GPU KV cache size: 1,310,720 tokens 2025-12-04T09:55:21.2686788Z (EngineCore_DP0 pid=12975) INFO 12-04 09:55:21 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 159.69x 2025-12-04T09:55:22.5968137Z (EngineCore_DP0 pid=12975) INFO 12-04 09:55:22 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:55:22.5981402Z INFO 12-04 09:55:22 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:55:23.0813592Z PASSED 2025-12-04T09:55:23.0960109Z models/test_initialization.py::test_can_initialize_large_subset[LlamaForCausalLM] Fork a new process to run a test 13045 2025-12-04T09:55:23.0969207Z Fork a new process to run a test 0 2025-12-04T09:55:23.1238385Z INFO 12-04 09:55:23 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='LlamaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'meta-llama/Llama-3.2-1B-Instruct'} 2025-12-04T09:55:23.2571257Z INFO 12-04 09:55:23 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-12-04T09:55:23.2571751Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:55:23.2813179Z INFO 12-04 09:55:23 [model.py:1714] Using max model len 131072 2025-12-04T09:55:23.4440012Z INFO 12-04 09:55:23 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:55:23.9459026Z (EngineCore_DP0 pid=13052) INFO 12-04 09:55:23 [core.py:727] Waiting for init message from front-end. 2025-12-04T09:55:23.9497095Z (EngineCore_DP0 pid=13052) INFO 12-04 09:55:23 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.bfloat16, max_seq_len=131072, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:55:25.2227019Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:25.2254049Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:25.2261441Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:25.2268131Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:25.2275639Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:25.2283266Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:25.2285825Z (EngineCore_DP0 pid=13052) INFO 12-04 09:55:25 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:55:25.5182287Z (EngineCore_DP0 pid=13052) INFO 12-04 09:55:25 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:55:25.5804295Z (EngineCore_DP0 pid=13052) INFO 12-04 09:55:25 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-12-04T09:55:25.8629814Z (EngineCore_DP0 pid=13052) INFO 12-04 09:55:25 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:55:25.9560856Z (EngineCore_DP0 pid=13052) INFO 12-04 09:55:25 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:55:26.5647963Z (EngineCore_DP0 pid=13052) INFO 12-04 09:55:26 [gpu_model_runner.py:2902] Model loading took 0.6192 GiB and 0.150526 seconds 2025-12-04T09:55:26.5652571Z (EngineCore_DP0 pid=13052) INFO 12-04 09:55:26 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T09:55:26.5653822Z (EngineCore_DP0 pid=13052) INFO 12-04 09:55:26 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 40.00x 2025-12-04T09:55:27.7406443Z (EngineCore_DP0 pid=13052) INFO 12-04 09:55:27 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:55:27.7421783Z INFO 12-04 09:55:27 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:55:28.2112798Z PASSED 2025-12-04T09:55:28.2259502Z models/test_initialization.py::test_can_initialize_large_subset[H2OVLChatModel] Fork a new process to run a test 13122 2025-12-04T09:55:28.2270395Z Fork a new process to run a test 0 2025-12-04T09:55:28.2274505Z `transformers==4.56.2` installed, but `transformers<=4.48` is required to run this model. Reason: HF model is not compatible. 2025-12-04T09:55:28.5049168Z PASSED 2025-12-04T09:55:28.5196623Z models/test_initialization.py::test_can_initialize_large_subset[EagleMiniCPMForCausalLM] Fork a new process to run a test 13123 2025-12-04T09:55:28.5207110Z Fork a new process to run a test 0 2025-12-04T09:55:28.5208384Z Model is not available online 2025-12-04T09:55:28.8007935Z PASSED 2025-12-04T09:55:28.8155201Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2_5OmniForConditionalGeneration] Fork a new process to run a test 13124 2025-12-04T09:55:28.8164932Z Fork a new process to run a test 0 2025-12-04T09:55:28.8440336Z INFO 12-04 09:55:28 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2_5OmniForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen2.5-Omni-7B-AWQ'} 2025-12-04T09:55:28.9195492Z 2025-12-04T09:55:28.9198915Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:55:28.9199235Z config.json: 15.4kB [00:00, 48.6MB/s] 2025-12-04T09:55:28.9722251Z Unrecognized keys in `rope_scaling` for 'rope_type'='default': {'mrope_section'} 2025-12-04T09:55:29.0391806Z 2025-12-04T09:55:29.0393053Z preprocessor_config.json: 0% 0.00/667 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:55:38.6385274Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:38.6415343Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:38.6432389Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:38.6432956Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:38.6438422Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:38.6446229Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:38.6449273Z (EngineCore_DP0 pid=13203) INFO 12-04 09:55:38 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:55:38.9362557Z (EngineCore_DP0 pid=13203) INFO 12-04 09:55:38 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:55:41.0936184Z (EngineCore_DP0 pid=13203) INFO 12-04 09:55:41 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen2.5-Omni-7B-AWQ... 2025-12-04T09:55:41.4666146Z (EngineCore_DP0 pid=13203) INFO 12-04 09:55:41 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:55:41.4667813Z (EngineCore_DP0 pid=13203) WARNING 12-04 09:55:41 [qwen2_5_omni_thinker.py:865] flash_attn is not available, the model may not yield the exactly same result as the transformers implementation in the audio tower part. 2025-12-04T09:55:41.5931521Z (EngineCore_DP0 pid=13203) INFO 12-04 09:55:41 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:55:42.2718387Z (EngineCore_DP0 pid=13203) INFO 12-04 09:55:42 [gpu_model_runner.py:2902] Model loading took 4.6412 GiB and 0.236989 seconds 2025-12-04T09:55:42.2722556Z (EngineCore_DP0 pid=13203) INFO 12-04 09:55:42 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T09:55:42.2723445Z (EngineCore_DP0 pid=13203) INFO 12-04 09:55:42 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 160.00x 2025-12-04T09:55:43.5571150Z (EngineCore_DP0 pid=13203) INFO 12-04 09:55:43 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:55:43.5585999Z INFO 12-04 09:55:43 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:55:44.0570536Z PASSED 2025-12-04T09:55:44.0717509Z models/test_initialization.py::test_can_initialize_large_subset[ArcticForCausalLM] Fork a new process to run a test 13273 2025-12-04T09:55:44.0726599Z Fork a new process to run a test 0 2025-12-04T09:55:44.0997631Z INFO 12-04 09:55:44 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='ArcticForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Snowflake/snowflake-arctic-instruct'} 2025-12-04T09:55:44.2770730Z 2025-12-04T09:55:44.2773414Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:55:44.2773921Z config.json: 1.28kB [00:00, 6.01MB/s] 2025-12-04T09:55:44.3882886Z 2025-12-04T09:55:44.3885238Z configuration_arctic.py: 0.00B [00:00, ?B/s] 2025-12-04T09:55:44.3885819Z configuration_arctic.py: 9.46kB [00:00, 28.9MB/s] 2025-12-04T09:55:44.3982081Z A new version of the following files was downloaded from https://huggingface.co/Snowflake/snowflake-arctic-instruct: 2025-12-04T09:55:44.3983008Z - configuration_arctic.py 2025-12-04T09:55:44.3983932Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:55:50.9821258Z INFO 12-04 09:55:50 [model.py:653] Resolved architecture: ArcticForCausalLM 2025-12-04T09:55:50.9821736Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:55:51.0075590Z INFO 12-04 09:55:51 [model.py:1714] Using max model len 4096 2025-12-04T09:55:51.0077580Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T09:55:51.0628002Z INFO 12-04 09:55:51 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:55:51.1045239Z 2025-12-04T09:55:51.1047671Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:55:51.1048203Z tokenizer_config.json: 1.58kB [00:00, 7.92MB/s] 2025-12-04T09:55:51.1399272Z 2025-12-04T09:55:51.1400504Z tokenization_arctic.py: 0.00B [00:00, ?B/s] 2025-12-04T09:55:51.1400858Z tokenization_arctic.py: 1.80kB [00:00, 14.1MB/s] 2025-12-04T09:55:51.1421796Z A new version of the following files was downloaded from https://huggingface.co/Snowflake/snowflake-arctic-instruct: 2025-12-04T09:55:51.1422378Z - tokenization_arctic.py 2025-12-04T09:55:51.1422998Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:55:51.2335964Z 2025-12-04T09:55:51.3346240Z tokenizer.model: 0% 0.00/500k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:55:53.0434636Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:53.0461942Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:53.0469461Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:53.0477375Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:53.0484628Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:53.0492097Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:55:53.0494619Z (EngineCore_DP0 pid=13351) INFO 12-04 09:55:53 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:55:53.3298600Z (EngineCore_DP0 pid=13351) INFO 12-04 09:55:53 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:55:53.4413399Z (EngineCore_DP0 pid=13351) INFO 12-04 09:55:53 [gpu_model_runner.py:2840] Starting to load model Snowflake/snowflake-arctic-instruct... 2025-12-04T09:55:53.7163123Z (EngineCore_DP0 pid=13351) INFO 12-04 09:55:53 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:55:53.7813621Z (EngineCore_DP0 pid=13351) INFO 12-04 09:55:53 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:55:54.3816932Z (EngineCore_DP0 pid=13351) INFO 12-04 09:55:54 [gpu_model_runner.py:2902] Model loading took 1.7530 GiB and 0.124453 seconds 2025-12-04T09:55:54.3820781Z (EngineCore_DP0 pid=13351) INFO 12-04 09:55:54 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T09:55:54.3821542Z (EngineCore_DP0 pid=13351) INFO 12-04 09:55:54 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 640.00x 2025-12-04T09:55:54.5479722Z (EngineCore_DP0 pid=13351) WARNING 12-04 09:55:54 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-12-04T09:55:54.9939483Z (EngineCore_DP0 pid=13351) INFO 12-04 09:55:54 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:55:54.9962693Z INFO 12-04 09:55:54 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:55:55.4549723Z PASSED 2025-12-04T09:55:55.4696182Z models/test_initialization.py::test_can_initialize_large_subset[KeyeForConditionalGeneration] Fork a new process to run a test 13421 2025-12-04T09:55:55.4706512Z Fork a new process to run a test 0 2025-12-04T09:55:55.4978071Z INFO 12-04 09:55:55 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='KeyeForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Kwai-Keye/Keye-VL-8B-Preview'} 2025-12-04T09:55:55.5821538Z 2025-12-04T09:55:55.5823551Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:55:55.5823902Z config.json: 1.87kB [00:00, 11.7MB/s] 2025-12-04T09:55:55.6430394Z 2025-12-04T09:55:55.6432906Z configuration_keye.py: 0.00B [00:00, ?B/s] 2025-12-04T09:55:55.6433207Z configuration_keye.py: 12.3kB [00:00, 40.7MB/s] 2025-12-04T09:55:55.6531369Z A new version of the following files was downloaded from https://huggingface.co/Kwai-Keye/Keye-VL-8B-Preview: 2025-12-04T09:55:55.6531946Z - configuration_keye.py 2025-12-04T09:55:55.6532584Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:55:55.8733999Z 2025-12-04T09:55:55.8734495Z preprocessor_config.json: 0% 0.00/380 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:56:05.4257946Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:05.4286873Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:05.4293820Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:05.4302438Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:05.4310414Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:05.4318012Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:05.4321005Z (EngineCore_DP0 pid=13499) INFO 12-04 09:56:05 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:56:05.7122277Z (EngineCore_DP0 pid=13499) INFO 12-04 09:56:05 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:56:05.8388627Z (EngineCore_DP0 pid=13499) 2025-12-04T09:56:05.8390064Z processor_config.json: 0% 0.00/115 [00:00, model_arch='Glm4ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'zai-org/GLM-4-9B-0414'} 2025-12-04T09:56:10.9464488Z 2025-12-04T09:56:10.9466028Z config.json: 0% 0.00/689 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:56:20.1716107Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:20.1745231Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:20.1752032Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:20.1759426Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:20.1767046Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:20.1775884Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:20.1778802Z (EngineCore_DP0 pid=13647) INFO 12-04 09:56:20 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:56:20.4629031Z (EngineCore_DP0 pid=13647) INFO 12-04 09:56:20 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:56:20.5510035Z (EngineCore_DP0 pid=13647) INFO 12-04 09:56:20 [gpu_model_runner.py:2840] Starting to load model zai-org/GLM-4-9B-0414... 2025-12-04T09:56:20.8277454Z (EngineCore_DP0 pid=13647) INFO 12-04 09:56:20 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:56:20.9130221Z (EngineCore_DP0 pid=13647) INFO 12-04 09:56:20 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:56:21.5180331Z (EngineCore_DP0 pid=13647) INFO 12-04 09:56:21 [gpu_model_runner.py:2902] Model loading took 2.6973 GiB and 0.146149 seconds 2025-12-04T09:56:21.5184883Z (EngineCore_DP0 pid=13647) INFO 12-04 09:56:21 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T09:56:21.5185839Z (EngineCore_DP0 pid=13647) INFO 12-04 09:56:21 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 320.00x 2025-12-04T09:56:23.4618416Z (EngineCore_DP0 pid=13647) INFO 12-04 09:56:23 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:56:23.4633154Z INFO 12-04 09:56:23 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:56:23.9610551Z PASSED 2025-12-04T09:56:23.9760669Z models/test_initialization.py::test_can_initialize_large_subset[OlmoeForCausalLM] Fork a new process to run a test 13717 2025-12-04T09:56:23.9771243Z Fork a new process to run a test 0 2025-12-04T09:56:24.0048927Z INFO 12-04 09:56:24 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='OlmoeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'allenai/OLMoE-1B-7B-0924-Instruct'} 2025-12-04T09:56:24.1246035Z 2025-12-04T09:56:24.1247351Z config.json: 0% 0.00/759 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:56:32.6463786Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:32.6489440Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:32.6496615Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:32.6504429Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:32.6512614Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:32.6520392Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:32.6523590Z (EngineCore_DP0 pid=13775) INFO 12-04 09:56:32 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:56:32.9349851Z (EngineCore_DP0 pid=13775) INFO 12-04 09:56:32 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:56:32.9829818Z (EngineCore_DP0 pid=13775) INFO 12-04 09:56:32 [gpu_model_runner.py:2840] Starting to load model allenai/OLMoE-1B-7B-0924-Instruct... 2025-12-04T09:56:33.2558920Z (EngineCore_DP0 pid=13775) INFO 12-04 09:56:33 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:56:33.3216280Z (EngineCore_DP0 pid=13775) INFO 12-04 09:56:33 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:56:33.9285922Z (EngineCore_DP0 pid=13775) INFO 12-04 09:56:33 [gpu_model_runner.py:2902] Model loading took 0.4395 GiB and 0.124689 seconds 2025-12-04T09:56:33.9291040Z (EngineCore_DP0 pid=13775) INFO 12-04 09:56:33 [kv_cache_utils.py:1199] GPU KV cache size: 1,310,720 tokens 2025-12-04T09:56:33.9291981Z (EngineCore_DP0 pid=13775) INFO 12-04 09:56:33 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 320.00x 2025-12-04T09:56:34.4079919Z (EngineCore_DP0 pid=13775) INFO 12-04 09:56:34 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:56:34.4100063Z INFO 12-04 09:56:34 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:56:34.8504589Z PASSED 2025-12-04T09:56:34.8654459Z models/test_initialization.py::test_can_initialize_large_subset[Glm4vMoeForConditionalGeneration] Fork a new process to run a test 13845 2025-12-04T09:56:34.8663999Z Fork a new process to run a test 0 2025-12-04T09:56:34.8938567Z INFO 12-04 09:56:34 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Glm4vMoeForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'zai-org/GLM-4.5V'} 2025-12-04T09:56:34.9679391Z 2025-12-04T09:56:34.9681484Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:56:34.9681793Z config.json: 1.85kB [00:00, 10.7MB/s] 2025-12-04T09:56:35.0839946Z 2025-12-04T09:56:35.0841079Z preprocessor_config.json: 0% 0.00/364 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:56:44.7869882Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:44.7907547Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:44.7914979Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:44.7923138Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:44.7930794Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:44.7938617Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:56:44.7941461Z (EngineCore_DP0 pid=13928) INFO 12-04 09:56:44 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:56:45.0747064Z (EngineCore_DP0 pid=13928) INFO 12-04 09:56:45 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:56:45.3403890Z (EngineCore_DP0 pid=13928) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-12-04T09:56:46.5884797Z (EngineCore_DP0 pid=13928) 2025-12-04T09:56:46.5885363Z video_preprocessor_config.json: 0% 0.00/365 [00:00, model_arch='PixtralForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'mistralai/Pixtral-12B-2409'} 2025-12-04T09:57:10.8769647Z 2025-12-04T09:57:10.8770889Z params.json: 0% 0.00/552 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:57:30.8315343Z (EngineCore_DP0 pid=14095) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:357: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-12-04T09:57:30.8316669Z (EngineCore_DP0 pid=14095) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-12-04T09:57:30.8335269Z (EngineCore_DP0 pid=14095) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-12-04T09:57:30.8336887Z (EngineCore_DP0 pid=14095) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 2025-12-04T09:57:30.8337642Z (EngineCore_DP0 pid=14095) dispatch key: ADInplaceOrView 2025-12-04T09:57:30.8338168Z (EngineCore_DP0 pid=14095) previous kernel: no debug info 2025-12-04T09:57:30.8339874Z (EngineCore_DP0 pid=14095) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-12-04T09:57:30.8341009Z (EngineCore_DP0 pid=14095) self.m.impl( 2025-12-04T09:57:31.9207404Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:31.9234374Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:31.9240977Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:31.9248254Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:31.9255084Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:31.9262068Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:31.9265012Z (EngineCore_DP0 pid=14095) INFO 12-04 09:57:31 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:57:32.1873671Z (EngineCore_DP0 pid=14095) INFO 12-04 09:57:32 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:57:32.3942460Z (EngineCore_DP0 pid=14095) INFO 12-04 09:57:32 [gpu_model_runner.py:2840] Starting to load model mistralai/Pixtral-12B-2409... 2025-12-04T09:57:32.6467887Z (EngineCore_DP0 pid=14095) INFO 12-04 09:57:32 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:57:32.7047584Z (EngineCore_DP0 pid=14095) INFO 12-04 09:57:32 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:57:33.2632172Z (EngineCore_DP0 pid=14095) INFO 12-04 09:57:33 [gpu_model_runner.py:2902] Model loading took 3.1304 GiB and 0.112781 seconds 2025-12-04T09:57:33.5087783Z (EngineCore_DP0 pid=14095) INFO 12-04 09:57:33 [gpu_model_runner.py:3647] Encoder cache will be initialized with a budget of 8192 tokens, and profiled with 1 image items of the maximum feature size. 2025-12-04T09:57:33.5588345Z (EngineCore_DP0 pid=14095) WARNING 12-04 09:57:33 [processing.py:1091] PixtralProcessorAdapter did not return `BatchFeature`. Make sure to match the behaviour of `ProcessorMixin` when implementing custom processors. 2025-12-04T09:57:35.1037246Z (EngineCore_DP0 pid=14095) INFO 12-04 09:57:35 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/1ec3568f6a/rank_0_0/backbone for vLLM's torch.compile 2025-12-04T09:57:35.1042366Z (EngineCore_DP0 pid=14095) INFO 12-04 09:57:35 [backends.py:608] Dynamo bytecode transform time: 1.24 s 2025-12-04T09:57:35.7372473Z (EngineCore_DP0 pid=14095) [rank0]:W1204 09:57:35.736000 14095 site-packages/torch/_inductor/utils.py:1703] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-12-04T09:57:37.0610913Z (EngineCore_DP0 pid=14095) INFO 12-04 09:57:37 [backends.py:214] Cache the graph for dynamic shape for later use 2025-12-04T09:57:37.9985753Z (EngineCore_DP0 pid=14095) INFO 12-04 09:57:37 [backends.py:241] Compiling a graph for dynamic shape takes 2.69 s 2025-12-04T09:57:38.5433299Z (EngineCore_DP0 pid=14095) INFO 12-04 09:57:38 [monitor.py:33] torch.compile takes 3.93 s in total 2025-12-04T09:57:39.6663940Z (EngineCore_DP0 pid=14095) INFO 12-04 09:57:39 [gpu_worker.py:314] Available KV cache memory: 13.48 GiB 2025-12-04T09:57:39.9645993Z (EngineCore_DP0 pid=14095) INFO 12-04 09:57:39 [kv_cache_utils.py:1199] GPU KV cache size: 3,534,544 tokens 2025-12-04T09:57:39.9646923Z (EngineCore_DP0 pid=14095) INFO 12-04 09:57:39 [kv_cache_utils.py:1204] Maximum concurrency for 128,000 tokens per request: 27.61x 2025-12-04T09:57:40.2669413Z (EngineCore_DP0 pid=14095) 2025-12-04T09:57:40.3694085Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/67 [00:00, model_arch='StableLmForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'stabilityai/stablelm-3b-4e1t'} 2025-12-04T09:57:43.5235718Z 2025-12-04T09:57:43.5237306Z config.json: 0% 0.00/599 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:57:45.7483613Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:45.7510876Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:45.7518495Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:45.7525892Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:45.7533769Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:45.7541678Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:45.7544552Z (EngineCore_DP0 pid=14377) INFO 12-04 09:57:45 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:57:46.0366833Z (EngineCore_DP0 pid=14377) INFO 12-04 09:57:46 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:57:46.1005954Z (EngineCore_DP0 pid=14377) INFO 12-04 09:57:46 [gpu_model_runner.py:2840] Starting to load model stabilityai/stablelm-3b-4e1t... 2025-12-04T09:57:46.3792006Z (EngineCore_DP0 pid=14377) INFO 12-04 09:57:46 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:57:46.4454676Z (EngineCore_DP0 pid=14377) INFO 12-04 09:57:46 [cuda.py:420] Using FlexAttention backend for head_size=80 on V1 engine. 2025-12-04T09:57:46.5481100Z (EngineCore_DP0 pid=14377) WARNING 12-04 09:57:46 [vllm.py:821] `torch.compile` is turned on, but the model stabilityai/stablelm-3b-4e1t does not support it. Please open an issue on GitHub if you want it to be supported. 2025-12-04T09:57:47.0904951Z (EngineCore_DP0 pid=14377) INFO 12-04 09:57:47 [gpu_model_runner.py:2902] Model loading took 0.6296 GiB and 0.171528 seconds 2025-12-04T09:57:47.0909257Z (EngineCore_DP0 pid=14377) INFO 12-04 09:57:47 [kv_cache_utils.py:1199] GPU KV cache size: 1,048,576 tokens 2025-12-04T09:57:47.0910150Z (EngineCore_DP0 pid=14377) INFO 12-04 09:57:47 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 256.00x 2025-12-04T09:57:47.6023257Z (EngineCore_DP0 pid=14377) INFO 12-04 09:57:47 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:57:47.6046670Z INFO 12-04 09:57:47 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:57:48.0578459Z PASSED 2025-12-04T09:57:48.0727964Z models/test_initialization.py::test_can_initialize_large_subset[DeepseekV32ForCausalLM] Fork a new process to run a test 14447 2025-12-04T09:57:48.0737702Z Fork a new process to run a test 0 2025-12-04T09:57:48.1007808Z INFO 12-04 09:57:48 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='DeepseekV32ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'deepseek-ai/DeepSeek-V3.2-Exp'} 2025-12-04T09:57:48.3396809Z 2025-12-04T09:57:48.3398419Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:57:48.3398833Z config.json: 1.55kB [00:00, 9.94MB/s] 2025-12-04T09:57:48.3826917Z You are using a model of type deepseek_v32 to instantiate a model of type deepseek_v3. This is not supported for all configurations of models and can yield errors. 2025-12-04T09:57:48.3835745Z INFO 12-04 09:57:48 [config.py:637] Detected quantization_config.scale_fmt=ue8m0; enabling UE8M0 for DeepGEMM. 2025-12-04T09:57:48.3837140Z INFO 12-04 09:57:48 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T09:57:48.4508751Z INFO 12-04 09:57:48 [model.py:653] Resolved architecture: DeepseekV32ForCausalLM 2025-12-04T09:57:48.4509333Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:57:48.4745522Z INFO 12-04 09:57:48 [model.py:1714] Using max model len 163840 2025-12-04T09:57:48.7488755Z INFO 12-04 09:57:48 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:57:48.7489611Z INFO 12-04 09:57:48 [config.py:480] Using custom fp8 kv-cache format for DeepSeekV3.2 2025-12-04T09:57:48.7519718Z INFO 12-04 09:57:48 [cuda.py:186] Forcing kv cache block size to 64 for FlashMLASparse backend. 2025-12-04T09:57:48.9103624Z 2025-12-04T09:57:48.9104889Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:57:48.9105272Z tokenizer_config.json: 4.06kB [00:00, 18.3MB/s] 2025-12-04T09:57:49.0237607Z 2025-12-04T09:57:49.0527957Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T09:57:49.0528304Z tokenizer.json: 7.85MB [00:00, 271MB/s] 2025-12-04T09:57:49.5662729Z 2025-12-04T09:57:49.5664080Z generation_config.json: 0% 0.00/171 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:57:50.9537656Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:50.9564762Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:50.9572112Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:50.9579744Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:50.9587038Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:50.9594501Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:50.9597291Z (EngineCore_DP0 pid=14455) INFO 12-04 09:57:50 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:57:51.2502307Z (EngineCore_DP0 pid=14455) INFO 12-04 09:57:51 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:57:51.3647413Z (EngineCore_DP0 pid=14455) INFO 12-04 09:57:51 [gpu_model_runner.py:2840] Starting to load model deepseek-ai/DeepSeek-V3.2-Exp... 2025-12-04T09:57:51.6538782Z (EngineCore_DP0 pid=14455) INFO 12-04 09:57:51 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:57:51.7459479Z (EngineCore_DP0 pid=14455) INFO 12-04 09:57:51 [cuda.py:275] Using Sparse MLA backend on V1 engine. 2025-12-04T09:57:51.7645929Z (EngineCore_DP0 pid=14455) WARNING 12-04 09:57:51 [fp8.py:150] DeepGEMM backend requested but not available. 2025-12-04T09:57:51.7646759Z (EngineCore_DP0 pid=14455) INFO 12-04 09:57:51 [fp8.py:165] Using Triton backend for FP8 MoE 2025-12-04T09:57:51.8862124Z (EngineCore_DP0 pid=14455) WARNING 12-04 09:57:51 [fp8_utils.py:785] Using default W8A8 Block FP8 kernel config. Performance might be sub-optimal! Config file not found at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/model_executor/layers/quantization/utils/configs/N=32768,K=512,device_name=NVIDIA_L4,dtype=fp8_w8a8,block_shape=[128,128].json 2025-12-04T09:57:52.9571178Z (EngineCore_DP0 pid=14455) INFO 12-04 09:57:52 [gpu_model_runner.py:2902] Model loading took 4.4518 GiB and 0.730603 seconds 2025-12-04T09:57:52.9576171Z (EngineCore_DP0 pid=14455) INFO 12-04 09:57:52 [kv_cache_utils.py:1199] GPU KV cache size: 13,626,112 tokens 2025-12-04T09:57:52.9577845Z (EngineCore_DP0 pid=14455) INFO 12-04 09:57:52 [kv_cache_utils.py:1204] Maximum concurrency for 163,840 tokens per request: 83.17x 2025-12-04T09:57:53.8028030Z (EngineCore_DP0 pid=14455) INFO 12-04 09:57:53 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:57:53.8042690Z INFO 12-04 09:57:53 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:57:54.3109058Z PASSED 2025-12-04T09:57:54.3259716Z models/test_initialization.py::test_can_initialize_large_subset[InternLM3ForCausalLM] Fork a new process to run a test 14529 2025-12-04T09:57:54.3269175Z Fork a new process to run a test 0 2025-12-04T09:57:54.3544138Z INFO 12-04 09:57:54 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='InternLM3ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'internlm/internlm3-8b-instruct'} 2025-12-04T09:57:54.4458527Z 2025-12-04T09:57:54.4460687Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:57:54.4461001Z config.json: 1.01kB [00:00, 5.16MB/s] 2025-12-04T09:57:54.5091039Z 2025-12-04T09:57:54.5093342Z configuration_internlm3.py: 0.00B [00:00, ?B/s] 2025-12-04T09:57:54.5093818Z configuration_internlm3.py: 10.5kB [00:00, 51.7MB/s] 2025-12-04T09:57:54.5186915Z A new version of the following files was downloaded from https://huggingface.co/internlm/internlm3-8b-instruct: 2025-12-04T09:57:54.5187492Z - configuration_internlm3.py 2025-12-04T09:57:54.5188133Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:57:54.7402385Z INFO 12-04 09:57:54 [model.py:653] Resolved architecture: InternLM3ForCausalLM 2025-12-04T09:57:54.7402791Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:57:54.7647554Z INFO 12-04 09:57:54 [model.py:1714] Using max model len 196608 2025-12-04T09:57:54.7649807Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T09:57:54.8220107Z INFO 12-04 09:57:54 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:57:54.8570630Z 2025-12-04T09:57:54.8572064Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:57:54.8572657Z tokenizer_config.json: 5.69kB [00:00, 41.9MB/s] 2025-12-04T09:57:54.8980690Z 2025-12-04T09:57:54.8982827Z tokenization_internlm3.py: 0.00B [00:00, ?B/s] 2025-12-04T09:57:54.8983217Z tokenization_internlm3.py: 13.2kB [00:00, 69.3MB/s] 2025-12-04T09:57:54.9037877Z A new version of the following files was downloaded from https://huggingface.co/internlm/internlm3-8b-instruct: 2025-12-04T09:57:54.9038474Z - tokenization_internlm3.py 2025-12-04T09:57:54.9039115Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:57:55.0010948Z 2025-12-04T09:57:55.1676603Z tokenizer.model: 0% 0.00/2.48M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:57:57.3206323Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:57.3234091Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:57.3241611Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:57.3249370Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:57.3256743Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:57.3264263Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:57:57.3267300Z (EngineCore_DP0 pid=14557) INFO 12-04 09:57:57 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:57:57.6069503Z (EngineCore_DP0 pid=14557) INFO 12-04 09:57:57 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:57:57.7021603Z (EngineCore_DP0 pid=14557) INFO 12-04 09:57:57 [gpu_model_runner.py:2840] Starting to load model internlm/internlm3-8b-instruct... 2025-12-04T09:57:57.9774903Z (EngineCore_DP0 pid=14557) INFO 12-04 09:57:57 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:57:58.0445451Z (EngineCore_DP0 pid=14557) INFO 12-04 09:57:58 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:57:58.6388822Z (EngineCore_DP0 pid=14557) INFO 12-04 09:57:58 [gpu_model_runner.py:2902] Model loading took 2.3086 GiB and 0.126745 seconds 2025-12-04T09:57:58.6392781Z (EngineCore_DP0 pid=14557) INFO 12-04 09:57:58 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T09:57:58.6393579Z (EngineCore_DP0 pid=14557) INFO 12-04 09:57:58 [kv_cache_utils.py:1204] Maximum concurrency for 196,608 tokens per request: 53.33x 2025-12-04T09:57:59.0418680Z (EngineCore_DP0 pid=14557) WARNING 12-04 09:57:59 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-12-04T09:58:00.4803913Z (EngineCore_DP0 pid=14557) INFO 12-04 09:58:00 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:58:00.4818124Z INFO 12-04 09:58:00 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:58:01.0246111Z PASSED 2025-12-04T09:58:01.0396619Z models/test_initialization.py::test_can_initialize_large_subset[ChatGLMForConditionalGeneration] Fork a new process to run a test 14627 2025-12-04T09:58:01.0407132Z Fork a new process to run a test 0 2025-12-04T09:58:01.0677441Z INFO 12-04 09:58:01 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='ChatGLMForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'thu-coai/ShieldLM-6B-chatglm3'} 2025-12-04T09:58:01.1826221Z 2025-12-04T09:58:01.1827973Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:58:01.1828289Z config.json: 1.51kB [00:00, 9.48MB/s] 2025-12-04T09:58:07.6288776Z INFO 12-04 09:58:07 [model.py:653] Resolved architecture: ChatGLMForConditionalGeneration 2025-12-04T09:58:07.6289334Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:58:07.6539778Z INFO 12-04 09:58:07 [model.py:1714] Using max model len 8192 2025-12-04T09:58:07.6541906Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T09:58:07.8247707Z INFO 12-04 09:58:07 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:58:07.8683280Z 2025-12-04T09:58:07.8684299Z tokenizer_config.json: 0% 0.00/244 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:58:09.9346916Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:09.9374100Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:09.9381612Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:09.9389599Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:09.9397621Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:09.9405869Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:09.9408773Z (EngineCore_DP0 pid=14705) INFO 12-04 09:58:09 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:58:10.2224510Z (EngineCore_DP0 pid=14705) INFO 12-04 09:58:10 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:58:10.2921042Z (EngineCore_DP0 pid=14705) INFO 12-04 09:58:10 [gpu_model_runner.py:2840] Starting to load model thu-coai/ShieldLM-6B-chatglm3... 2025-12-04T09:58:10.5722370Z (EngineCore_DP0 pid=14705) INFO 12-04 09:58:10 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:58:10.6397442Z (EngineCore_DP0 pid=14705) INFO 12-04 09:58:10 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:58:11.2518198Z (EngineCore_DP0 pid=14705) INFO 12-04 09:58:11 [gpu_model_runner.py:2902] Model loading took 1.3741 GiB and 0.127526 seconds 2025-12-04T09:58:11.2523314Z (EngineCore_DP0 pid=14705) INFO 12-04 09:58:11 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T09:58:11.2524079Z (EngineCore_DP0 pid=14705) INFO 12-04 09:58:11 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 1280.00x 2025-12-04T09:58:11.5283591Z (EngineCore_DP0 pid=14705) WARNING 12-04 09:58:11 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-12-04T09:58:12.9475121Z (EngineCore_DP0 pid=14705) INFO 12-04 09:58:12 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:58:12.9490112Z INFO 12-04 09:58:12 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:58:13.4155673Z PASSED 2025-12-04T09:58:13.4306822Z models/test_initialization.py::test_can_initialize_large_subset[MiniMaxForCausalLM] Fork a new process to run a test 14775 2025-12-04T09:58:13.4316279Z Fork a new process to run a test 0 2025-12-04T09:58:13.4590403Z INFO 12-04 09:58:13 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MiniMaxForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'MiniMaxAI/MiniMax-Text-01-hf'} 2025-12-04T09:58:13.6078055Z 2025-12-04T09:58:13.6080514Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:58:13.6080983Z config.json: 3.00kB [00:00, 15.8MB/s] 2025-12-04T09:58:13.7092412Z INFO 12-04 09:58:13 [model.py:653] Resolved architecture: MiniMaxForCausalLM 2025-12-04T09:58:13.7092914Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:58:13.8203292Z 2025-12-04T09:58:13.8227672Z model.safetensors.index.json: 0.00B [00:00, ?B/s] 2025-12-04T09:58:13.8228147Z model.safetensors.index.json: 823kB [00:00, 321MB/s] 2025-12-04T09:58:13.8495584Z 2025-12-04T09:58:13.9753209Z Parse safetensors files: 0% 0/413 [00:00= mamba page size. 2025-12-04T09:58:15.5319352Z 2025-12-04T09:58:15.5320521Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:58:15.5320947Z tokenizer_config.json: 1.39kB [00:00, 14.4MB/s] 2025-12-04T09:58:15.6779903Z 2025-12-04T09:58:15.6982175Z vocab.json: 0.00B [00:00, ?B/s] 2025-12-04T09:58:15.6982627Z vocab.json: 4.71MB [00:00, 233MB/s] 2025-12-04T09:58:15.7573049Z 2025-12-04T09:58:15.7694102Z merges.txt: 0.00B [00:00, ?B/s] 2025-12-04T09:58:15.7694533Z merges.txt: 2.41MB [00:00, 198MB/s] 2025-12-04T09:58:15.8456305Z 2025-12-04T09:58:15.8823891Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T09:58:15.8824393Z tokenizer.json: 9.72MB [00:00, 265MB/s] 2025-12-04T09:58:16.6697344Z (EngineCore_DP0 pid=14803) INFO 12-04 09:58:16 [core.py:727] Waiting for init message from front-end. 2025-12-04T09:58:16.6733493Z (EngineCore_DP0 pid=14803) INFO 12-04 09:58:16 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='MiniMaxAI/MiniMax-Text-01-hf', speculative_config=None, tokenizer='MiniMaxAI/MiniMax-Text-01-hf', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.bfloat16, max_seq_len=10240000, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=MiniMaxAI/MiniMax-Text-01-hf, enable_prefix_caching=False, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:58:18.0011799Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:18.0042804Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:18.0051587Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:18.0060734Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:18.0069936Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:18.0079093Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:18.0082306Z (EngineCore_DP0 pid=14803) INFO 12-04 09:58:18 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:58:18.3021935Z (EngineCore_DP0 pid=14803) INFO 12-04 09:58:18 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:58:19.2662651Z (EngineCore_DP0 pid=14803) INFO 12-04 09:58:19 [gpu_model_runner.py:2840] Starting to load model MiniMaxAI/MiniMax-Text-01-hf... 2025-12-04T09:58:19.5488339Z (EngineCore_DP0 pid=14803) INFO 12-04 09:58:19 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:58:20.2142782Z (EngineCore_DP0 pid=14803) INFO 12-04 09:58:20 [gpu_model_runner.py:2902] Model loading took 7.9713 GiB and 0.110637 seconds 2025-12-04T09:58:20.2148338Z (EngineCore_DP0 pid=14803) INFO 12-04 09:58:20 [kv_cache_utils.py:1199] GPU KV cache size: 52,428,800,000 tokens 2025-12-04T09:58:20.2149330Z (EngineCore_DP0 pid=14803) INFO 12-04 09:58:20 [kv_cache_utils.py:1204] Maximum concurrency for 10,240,000 tokens per request: 5120.00x 2025-12-04T09:58:21.1350971Z (EngineCore_DP0 pid=14803) INFO 12-04 09:58:21 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:58:21.1366624Z INFO 12-04 09:58:21 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:58:22.5474711Z PASSED 2025-12-04T09:58:22.5625128Z models/test_initialization.py::test_can_initialize_large_subset[Zamba2ForCausalLM] Fork a new process to run a test 14873 2025-12-04T09:58:22.5633434Z Fork a new process to run a test 0 2025-12-04T09:58:22.5906327Z INFO 12-04 09:58:22 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Zamba2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Zyphra/Zamba2-7B-instruct'} 2025-12-04T09:58:22.7842891Z 2025-12-04T09:58:22.7844951Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:58:22.7845718Z config.json: 2.41kB [00:00, 14.8MB/s] 2025-12-04T09:58:29.3418830Z INFO 12-04 09:58:29 [model.py:653] Resolved architecture: Zamba2ForCausalLM 2025-12-04T09:58:29.3419637Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:58:29.5463213Z 2025-12-04T09:58:29.5466207Z model.safetensors.index.json: 0.00B [00:00, ?B/s] 2025-12-04T09:58:29.5466779Z model.safetensors.index.json: 63.3kB [00:00, 241MB/s] 2025-12-04T09:58:29.5502993Z 2025-12-04T09:58:29.6577689Z Parse safetensors files: 0% 0/3 [00:00= mamba page size. 2025-12-04T09:58:29.9469432Z INFO 12-04 09:58:29 [config.py:453] Padding mamba page size by 43.05% to ensure that mamba page size and attention page size are exactly equal. 2025-12-04T09:58:30.0037330Z 2025-12-04T09:58:30.0038345Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:58:30.0039060Z tokenizer_config.json: 1.37kB [00:00, 14.3MB/s] 2025-12-04T09:58:30.1860727Z 2025-12-04T09:58:30.1956649Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T09:58:30.1957012Z tokenizer.json: 1.80MB [00:00, 185MB/s] 2025-12-04T09:58:30.3003717Z 2025-12-04T09:58:30.3004459Z special_tokens_map.json: 0% 0.00/504 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:58:31.7950008Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:31.7977172Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:31.7984799Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:31.7992417Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:31.8000342Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:31.8008067Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:31.8010544Z (EngineCore_DP0 pid=14934) INFO 12-04 09:58:31 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:58:32.0809899Z (EngineCore_DP0 pid=14934) INFO 12-04 09:58:32 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:58:32.1452490Z (EngineCore_DP0 pid=14934) INFO 12-04 09:58:32 [gpu_model_runner.py:2840] Starting to load model Zyphra/Zamba2-7B-instruct... 2025-12-04T09:58:32.4177938Z (EngineCore_DP0 pid=14934) INFO 12-04 09:58:32 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:58:32.4204919Z (EngineCore_DP0 pid=14934) INFO 12-04 09:58:32 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:58:33.2034204Z (EngineCore_DP0 pid=14934) INFO 12-04 09:58:33 [gpu_model_runner.py:2902] Model loading took 13.7859 GiB and 0.237595 seconds 2025-12-04T09:58:33.2051875Z (EngineCore_DP0 pid=14934) WARNING 12-04 09:58:33 [kv_cache_utils.py:949] Add 10 padding layers, may waste at most 12.35% KV cache memory 2025-12-04T09:58:33.2060368Z (EngineCore_DP0 pid=14934) INFO 12-04 09:58:33 [kv_cache_utils.py:1199] GPU KV cache size: 3,600 tokens 2025-12-04T09:58:33.2062348Z (EngineCore_DP0 pid=14934) INFO 12-04 09:58:33 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 6.45x 2025-12-04T09:58:33.7526604Z (EngineCore_DP0 pid=14934) INFO 12-04 09:58:33 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:58:33.7550000Z INFO 12-04 09:58:33 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:58:34.2188383Z PASSED 2025-12-04T09:58:34.2337884Z models/test_initialization.py::test_can_initialize_large_subset[JambaForSequenceClassification] Fork a new process to run a test 15004 2025-12-04T09:58:34.2348599Z Fork a new process to run a test 0 2025-12-04T09:58:34.2627961Z INFO 12-04 09:58:34 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='JambaForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ai21labs/Jamba-tiny-reward-dev'} 2025-12-04T09:58:34.3852998Z 2025-12-04T09:58:34.3855330Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:58:34.3855667Z config.json: 1.14kB [00:00, 5.71MB/s] 2025-12-04T09:58:40.8764881Z INFO 12-04 09:58:40 [model.py:915] Resolved `--runner auto` to `--runner pooling`. Pass the value explicitly to silence this message. 2025-12-04T09:58:40.8765895Z INFO 12-04 09:58:40 [model.py:653] Resolved architecture: JambaForSequenceClassification 2025-12-04T09:58:40.8766845Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:58:40.9017772Z INFO 12-04 09:58:40 [model.py:1714] Using max model len 262144 2025-12-04T09:58:40.9306621Z INFO 12-04 09:58:40 [arg_utils.py:1725] (Enabling) chunked prefill by default 2025-12-04T09:58:40.9307294Z INFO 12-04 09:58:40 [arg_utils.py:1728] (Enabling) prefix caching by default 2025-12-04T09:58:41.0717588Z INFO 12-04 09:58:41 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:58:41.0718668Z INFO 12-04 09:58:41 [config.py:306] Hybrid or mamba-based model detected without support for prefix caching: disabling. 2025-12-04T09:58:41.0719612Z INFO 12-04 09:58:41 [config.py:313] Disabling cascade attention since it is not supported for hybrid models. 2025-12-04T09:58:41.1265211Z INFO 12-04 09:58:41 [config.py:429] Setting attention block size to 80 tokens to ensure that attention page size is >= mamba page size. 2025-12-04T09:58:41.1268095Z INFO 12-04 09:58:41 [config.py:453] Padding mamba page size by 5.26% to ensure that mamba page size and attention page size are exactly equal. 2025-12-04T09:58:41.1681501Z 2025-12-04T09:58:41.1684786Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:58:41.1685275Z tokenizer_config.json: 14.3kB [00:00, 42.5MB/s] 2025-12-04T09:58:41.2684544Z 2025-12-04T09:58:41.5099467Z tokenizer.model: 0% 0.00/1.12M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:58:43.3461610Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:43.3489032Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:43.3495635Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:43.3503579Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:43.3511371Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:43.3518954Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:43.3521825Z (EngineCore_DP0 pid=15083) INFO 12-04 09:58:43 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:58:43.6410397Z (EngineCore_DP0 pid=15083) INFO 12-04 09:58:43 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:58:43.6789954Z (EngineCore_DP0 pid=15083) INFO 12-04 09:58:43 [gpu_model_runner.py:2840] Starting to load model ai21labs/Jamba-tiny-reward-dev... 2025-12-04T09:58:43.9651432Z (EngineCore_DP0 pid=15083) INFO 12-04 09:58:43 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:58:44.5907929Z (EngineCore_DP0 pid=15083) INFO 12-04 09:58:44 [gpu_model_runner.py:2902] Model loading took 0.1348 GiB and 0.064181 seconds 2025-12-04T09:58:44.5913035Z (EngineCore_DP0 pid=15083) INFO 12-04 09:58:44 [kv_cache_utils.py:1199] GPU KV cache size: 68,719,476,736 tokens 2025-12-04T09:58:44.5913998Z (EngineCore_DP0 pid=15083) INFO 12-04 09:58:44 [kv_cache_utils.py:1204] Maximum concurrency for 262,144 tokens per request: 262144.00x 2025-12-04T09:58:45.6263707Z (EngineCore_DP0 pid=15083) INFO 12-04 09:58:45 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:58:45.6280177Z INFO 12-04 09:58:45 [llm.py:337] Supported tasks: ['classify'] 2025-12-04T09:58:46.0921416Z PASSED 2025-12-04T09:58:46.1071296Z models/test_initialization.py::test_can_initialize_large_subset[Cohere2VisionForConditionalGeneration] Fork a new process to run a test 15153 2025-12-04T09:58:46.1081263Z Fork a new process to run a test 0 2025-12-04T09:58:46.1358368Z INFO 12-04 09:58:46 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Cohere2VisionForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'CohereLabs/command-a-vision-07-2025'} 2025-12-04T09:58:46.3042806Z 2025-12-04T09:58:46.3043857Z config.json: 0% 0.00/3.59k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:58:56.7808578Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:56.7838642Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:56.7845741Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:56.7853189Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:56.7861109Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:56.7869414Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:58:56.7872766Z (EngineCore_DP0 pid=15231) INFO 12-04 09:58:56 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:58:57.0703297Z (EngineCore_DP0 pid=15231) INFO 12-04 09:58:57 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:59:01.3669833Z (EngineCore_DP0 pid=15231) INFO 12-04 09:59:01 [gpu_model_runner.py:2840] Starting to load model CohereLabs/command-a-vision-07-2025... 2025-12-04T09:59:01.7913078Z (EngineCore_DP0 pid=15231) INFO 12-04 09:59:01 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:59:01.8036266Z (EngineCore_DP0 pid=15231) INFO 12-04 09:59:01 [layer.py:497] MultiHeadAttention attn_backend: _Backend.XFORMERS, use_upstream_fa: False 2025-12-04T09:59:01.8462218Z (EngineCore_DP0 pid=15231) INFO 12-04 09:59:01 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:59:02.4402813Z (EngineCore_DP0 pid=15231) INFO 12-04 09:59:02 [gpu_model_runner.py:2902] Model loading took 9.8899 GiB and 0.103932 seconds 2025-12-04T09:59:02.4406809Z (EngineCore_DP0 pid=15231) INFO 12-04 09:59:02 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T09:59:02.4407758Z (EngineCore_DP0 pid=15231) INFO 12-04 09:59:02 [kv_cache_utils.py:1204] Maximum concurrency for 500,000 tokens per request: 213.06x 2025-12-04T09:59:03.8394324Z (EngineCore_DP0 pid=15231) INFO 12-04 09:59:03 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:59:03.8409088Z INFO 12-04 09:59:03 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:59:04.4646196Z PASSED 2025-12-04T09:59:04.4795691Z models/test_initialization.py::test_can_initialize_large_subset[LongCatFlashMTPModel] Fork a new process to run a test 15317 2025-12-04T09:59:04.4806908Z Fork a new process to run a test 0 2025-12-04T09:59:04.5085541Z INFO 12-04 09:59:04 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='LongCatFlashMTPModel', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'meituan-longcat/LongCat-Flash-Chat', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'meituan-longcat/LongCat-Flash-Chat'} 2025-12-04T09:59:04.7152240Z 2025-12-04T09:59:04.7154435Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:59:04.7154769Z config.json: 1.00kB [00:00, 5.01MB/s] 2025-12-04T09:59:04.7884170Z 2025-12-04T09:59:04.7888122Z configuration_longcat_flash.py: 0.00B [00:00, ?B/s] 2025-12-04T09:59:04.7888552Z configuration_longcat_flash.py: 9.69kB [00:00, 28.2MB/s] 2025-12-04T09:59:04.7995884Z A new version of the following files was downloaded from https://huggingface.co/meituan-longcat/LongCat-Flash-Chat: 2025-12-04T09:59:04.7996448Z - configuration_longcat_flash.py 2025-12-04T09:59:04.7997136Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:59:11.3731752Z INFO 12-04 09:59:11 [model.py:653] Resolved architecture: LongcatFlashForCausalLM 2025-12-04T09:59:11.3732265Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:59:11.4707812Z 2025-12-04T09:59:11.4805208Z model.safetensors.index.json: 0.00B [00:00, ?B/s] 2025-12-04T09:59:11.4805658Z model.safetensors.index.json: 4.14MB [00:00, 424MB/s] 2025-12-04T09:59:11.5248133Z 2025-12-04T09:59:11.6351041Z Parse safetensors files: 0% 0/75 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:59:21.4227212Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:21.4257668Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:21.4267088Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:21.4276778Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:21.4286602Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:21.4296112Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:21.4299512Z (EngineCore_DP0 pid=15465) INFO 12-04 09:59:21 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:59:21.7153353Z (EngineCore_DP0 pid=15465) INFO 12-04 09:59:21 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:59:21.7301619Z (EngineCore_DP0 pid=15465) WARNING 12-04 09:59:21 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-12-04T09:59:21.8186567Z (EngineCore_DP0 pid=15465) INFO 12-04 09:59:21 [gpu_model_runner.py:2840] Starting to load model meituan-longcat/LongCat-Flash-Chat... 2025-12-04T09:59:22.0987412Z (EngineCore_DP0 pid=15465) INFO 12-04 09:59:22 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:59:22.1551942Z (EngineCore_DP0 pid=15465) INFO 12-04 09:59:22 [cuda.py:328] Using Triton MLA backend on V1 engine. 2025-12-04T09:59:22.2354556Z (EngineCore_DP0 pid=15465) INFO 12-04 09:59:22 [gpu_model_runner.py:2879] Loading drafter model... 2025-12-04T09:59:22.2566023Z (EngineCore_DP0 pid=15465) WARNING 12-04 09:59:22 [vllm.py:821] `torch.compile` is turned on, but the model meituan-longcat/LongCat-Flash-Chat does not support it. Please open an issue on GitHub if you want it to be supported. 2025-12-04T09:59:22.2575190Z (EngineCore_DP0 pid=15465) INFO 12-04 09:59:22 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-12-04T09:59:22.2576173Z (EngineCore_DP0 pid=15465) INFO 12-04 09:59:22 [eagle.py:1027] Loading EAGLE LM head weights from the target model. 2025-12-04T09:59:22.8406907Z (EngineCore_DP0 pid=15465) INFO 12-04 09:59:22 [gpu_model_runner.py:2902] Model loading took 5.0746 GiB and 0.159111 seconds 2025-12-04T09:59:22.8407826Z (EngineCore_DP0 pid=15465) INFO 12-04 09:59:22 [kv_cache_utils.py:1199] GPU KV cache size: 3,106,880 tokens 2025-12-04T09:59:22.8408546Z (EngineCore_DP0 pid=15465) INFO 12-04 09:59:22 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 23.70x 2025-12-04T09:59:23.8269033Z (EngineCore_DP0 pid=15465) INFO 12-04 09:59:23 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:59:23.8282641Z INFO 12-04 09:59:23 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:59:24.3403301Z PASSED 2025-12-04T09:59:24.3552714Z models/test_initialization.py::test_can_initialize_large_subset[HunYuanMoEV1ForCausalLM] Fork a new process to run a test 15535 2025-12-04T09:59:24.3562606Z Fork a new process to run a test 0 2025-12-04T09:59:24.3835613Z INFO 12-04 09:59:24 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='HunYuanMoEV1ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'tencent/Hunyuan-A13B-Instruct'} 2025-12-04T09:59:24.4777867Z 2025-12-04T09:59:24.4779969Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:59:24.4780391Z config.json: 3.38kB [00:00, 19.6MB/s] 2025-12-04T09:59:24.5560213Z 2025-12-04T09:59:24.5562197Z configuration_hunyuan.py: 0.00B [00:00, ?B/s] 2025-12-04T09:59:24.5562722Z configuration_hunyuan.py: 15.0kB [00:00, 59.0MB/s] 2025-12-04T09:59:24.5677536Z A new version of the following files was downloaded from https://huggingface.co/tencent/Hunyuan-A13B-Instruct: 2025-12-04T09:59:24.5678139Z - configuration_hunyuan.py 2025-12-04T09:59:24.5678786Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:59:24.6146957Z You are using a model of type hunyuan_v1_moe to instantiate a model of type hunyuan. This is not supported for all configurations of models and can yield errors. 2025-12-04T09:59:24.6159859Z INFO 12-04 09:59:24 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T09:59:31.1634984Z INFO 12-04 09:59:31 [model.py:653] Resolved architecture: HunYuanMoEV1ForCausalLM 2025-12-04T09:59:31.1635492Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:59:31.1891546Z INFO 12-04 09:59:31 [model.py:1714] Using max model len 32768 2025-12-04T09:59:31.1893124Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T09:59:31.2483387Z INFO 12-04 09:59:31 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:59:31.2813575Z 2025-12-04T09:59:31.2814873Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:59:31.2815227Z tokenizer_config.json: 1.02kB [00:00, 8.90MB/s] 2025-12-04T09:59:31.3718299Z 2025-12-04T09:59:31.5931637Z tokenizer.json: 0% 0.00/16.4M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:59:33.7993221Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:33.8020898Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:33.8028220Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:33.8039737Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:33.8049779Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:33.8057384Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:33.8060474Z (EngineCore_DP0 pid=15613) INFO 12-04 09:59:33 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:59:34.0956788Z (EngineCore_DP0 pid=15613) INFO 12-04 09:59:34 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:59:34.1865661Z (EngineCore_DP0 pid=15613) INFO 12-04 09:59:34 [gpu_model_runner.py:2840] Starting to load model tencent/Hunyuan-A13B-Instruct... 2025-12-04T09:59:34.4699970Z (EngineCore_DP0 pid=15613) INFO 12-04 09:59:34 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:59:34.5484920Z (EngineCore_DP0 pid=15613) INFO 12-04 09:59:34 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:59:35.1548925Z (EngineCore_DP0 pid=15613) INFO 12-04 09:59:35 [gpu_model_runner.py:2902] Model loading took 1.2754 GiB and 0.138283 seconds 2025-12-04T09:59:35.1554074Z (EngineCore_DP0 pid=15613) INFO 12-04 09:59:35 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T09:59:35.1555043Z (EngineCore_DP0 pid=15613) INFO 12-04 09:59:35 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 80.00x 2025-12-04T09:59:36.3673176Z (EngineCore_DP0 pid=15613) INFO 12-04 09:59:36 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:59:36.3687858Z INFO 12-04 09:59:36 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T09:59:36.8576168Z PASSED 2025-12-04T09:59:36.8725371Z models/test_initialization.py::test_can_initialize_large_subset[NomicBertModel] Fork a new process to run a test 15683 2025-12-04T09:59:36.8735449Z Fork a new process to run a test 0 2025-12-04T09:59:36.9007546Z INFO 12-04 09:59:36 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='NomicBertModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'nomic-ai/nomic-embed-text-v2-moe'} 2025-12-04T09:59:36.9709703Z 2025-12-04T09:59:36.9711632Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:59:36.9712041Z config.json: 2.48kB [00:00, 16.0MB/s] 2025-12-04T09:59:37.0391858Z 2025-12-04T09:59:37.0392745Z configuration_hf_nomic_bert.py: 0.00B [00:00, ?B/s] 2025-12-04T09:59:37.0393221Z configuration_hf_nomic_bert.py: 1.96kB [00:00, 12.3MB/s] 2025-12-04T09:59:37.0487940Z A new version of the following files was downloaded from https://huggingface.co/nomic-ai/nomic-bert-2048: 2025-12-04T09:59:37.0488477Z - configuration_hf_nomic_bert.py 2025-12-04T09:59:37.0489115Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:59:37.2551733Z 2025-12-04T09:59:37.2552271Z sentence_bert_config.json: 0% 0.00/53.0 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T09:59:46.6212249Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:46.6239198Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:46.6247770Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:46.6254940Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:46.6262608Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:46.6270005Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T09:59:46.6273498Z (EngineCore_DP0 pid=15764) INFO 12-04 09:59:46 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T09:59:46.9070505Z (EngineCore_DP0 pid=15764) INFO 12-04 09:59:46 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T09:59:46.9217484Z (EngineCore_DP0 pid=15764) INFO 12-04 09:59:46 [gpu_model_runner.py:2840] Starting to load model nomic-ai/nomic-embed-text-v2-moe... 2025-12-04T09:59:47.1960354Z (EngineCore_DP0 pid=15764) INFO 12-04 09:59:47 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T09:59:47.2615669Z (EngineCore_DP0 pid=15764) INFO 12-04 09:59:47 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T09:59:47.8788210Z (EngineCore_DP0 pid=15764) INFO 12-04 09:59:47 [gpu_model_runner.py:2902] Model loading took 0.3718 GiB and 0.127305 seconds 2025-12-04T09:59:48.6604003Z (EngineCore_DP0 pid=15764) INFO 12-04 09:59:48 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T09:59:48.7289123Z (EngineCore_DP0 pid=15764) INFO 12-04 09:59:48 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T09:59:49.0114677Z (EngineCore_DP0 pid=15764) INFO 12-04 09:59:49 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T09:59:49.0139362Z INFO 12-04 09:59:49 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-12-04T09:59:49.5057712Z PASSED 2025-12-04T09:59:49.5207352Z models/test_initialization.py::test_can_initialize_large_subset[MptForCausalLM] Fork a new process to run a test 15834 2025-12-04T09:59:49.5217753Z Fork a new process to run a test 0 2025-12-04T09:59:49.5219985Z Model is not available online 2025-12-04T09:59:49.8050998Z PASSED 2025-12-04T09:59:49.8200444Z models/test_initialization.py::test_can_initialize_large_subset[TeleChat2ForCausalLM] Fork a new process to run a test 15835 2025-12-04T09:59:49.8210293Z Fork a new process to run a test 0 2025-12-04T09:59:49.8477173Z INFO 12-04 09:59:49 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='TeleChat2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Tele-AI/TeleChat2-3B'} 2025-12-04T09:59:49.9313341Z 2025-12-04T09:59:49.9315687Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:59:49.9316184Z config.json: 1.04kB [00:00, 5.85MB/s] 2025-12-04T09:59:49.9922642Z 2025-12-04T09:59:49.9924597Z configuration_telechat2.py: 0.00B [00:00, ?B/s] 2025-12-04T09:59:49.9925168Z configuration_telechat2.py: 4.48kB [00:00, 17.7MB/s] 2025-12-04T09:59:50.0020110Z A new version of the following files was downloaded from https://huggingface.co/Tele-AI/TeleChat2-3B: 2025-12-04T09:59:50.0020772Z - configuration_telechat2.py 2025-12-04T09:59:50.0021432Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:59:56.4974984Z INFO 12-04 09:59:56 [model.py:653] Resolved architecture: TeleChat2ForCausalLM 2025-12-04T09:59:56.4975794Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T09:59:56.5440567Z ERROR 12-04 09:59:56 [config.py:298] Error retrieving safetensors: 'Tele-AI/TeleChat2-3B' is not a safetensors repo. Couldn't find 'model.safetensors.index.json' or 'model.safetensors' files., retrying 1 of 2 2025-12-04T09:59:58.6010721Z ERROR 12-04 09:59:58 [config.py:296] Error retrieving safetensors: 'Tele-AI/TeleChat2-3B' is not a safetensors repo. Couldn't find 'model.safetensors.index.json' or 'model.safetensors' files. 2025-12-04T09:59:58.6261918Z INFO 12-04 09:59:58 [model.py:1939] Downcasting torch.float32 to torch.bfloat16. 2025-12-04T09:59:58.6267023Z INFO 12-04 09:59:58 [model.py:1714] Using max model len 32768 2025-12-04T09:59:58.6269183Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T09:59:58.6878881Z INFO 12-04 09:59:58 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T09:59:58.7253438Z 2025-12-04T09:59:58.7254989Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T09:59:58.7255342Z tokenizer_config.json: 5.07kB [00:00, 39.7MB/s] 2025-12-04T09:59:58.7620807Z 2025-12-04T09:59:58.7622566Z tokenization_telechat2.py: 0.00B [00:00, ?B/s] 2025-12-04T09:59:58.7622917Z tokenization_telechat2.py: 8.82kB [00:00, 40.3MB/s] 2025-12-04T09:59:58.7673666Z A new version of the following files was downloaded from https://huggingface.co/Tele-AI/TeleChat2-3B: 2025-12-04T09:59:58.7674473Z - tokenization_telechat2.py 2025-12-04T09:59:58.7675155Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T09:59:58.8578995Z 2025-12-04T09:59:59.0541839Z tokenizer.model: 0% 0.00/2.20M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:00:00.9761871Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:00.9789634Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:00.9796697Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:00.9806957Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:00.9816846Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:00.9824945Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:00.9827261Z (EngineCore_DP0 pid=15913) INFO 12-04 10:00:00 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:00:01.2624560Z (EngineCore_DP0 pid=15913) INFO 12-04 10:00:01 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:00:01.3320037Z (EngineCore_DP0 pid=15913) INFO 12-04 10:00:01 [gpu_model_runner.py:2840] Starting to load model Tele-AI/TeleChat2-3B... 2025-12-04T10:00:01.6051450Z (EngineCore_DP0 pid=15913) INFO 12-04 10:00:01 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:00:01.6726907Z (EngineCore_DP0 pid=15913) INFO 12-04 10:00:01 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:00:02.2689892Z (EngineCore_DP0 pid=15913) INFO 12-04 10:00:02 [gpu_model_runner.py:2902] Model loading took 1.6778 GiB and 0.127432 seconds 2025-12-04T10:00:02.2694581Z (EngineCore_DP0 pid=15913) INFO 12-04 10:00:02 [kv_cache_utils.py:1199] GPU KV cache size: 873,808 tokens 2025-12-04T10:00:02.2695356Z (EngineCore_DP0 pid=15913) INFO 12-04 10:00:02 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 26.67x 2025-12-04T10:00:02.6463256Z (EngineCore_DP0 pid=15913) WARNING 12-04 10:00:02 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-12-04T10:00:03.1062259Z (EngineCore_DP0 pid=15913) INFO 12-04 10:00:03 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:00:03.1077593Z INFO 12-04 10:00:03 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:00:03.5873873Z PASSED 2025-12-04T10:00:03.6022336Z models/test_initialization.py::test_can_initialize_large_subset[GteNewModel] Fork a new process to run a test 15983 2025-12-04T10:00:03.6031911Z Fork a new process to run a test 0 2025-12-04T10:00:03.6299039Z INFO 12-04 10:00:03 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GteNewModel', exist_overrides={'architectures': ['GteNewModel']}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Alibaba-NLP/gte-base-en-v1.5'} 2025-12-04T10:00:03.7083289Z 2025-12-04T10:00:03.7085413Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:00:03.7085707Z config.json: 1.35kB [00:00, 7.38MB/s] 2025-12-04T10:00:03.8101907Z INFO 12-04 10:00:03 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T10:00:03.9879997Z 2025-12-04T10:00:03.9880647Z sentence_bert_config.json: 0% 0.00/54.0 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:00:12.2559695Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:12.2587827Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:12.2594940Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:12.2602384Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:12.2609708Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:12.2617291Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:12.2620126Z (EngineCore_DP0 pid=16042) INFO 12-04 10:00:12 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:00:12.5432528Z (EngineCore_DP0 pid=16042) INFO 12-04 10:00:12 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:00:12.5657213Z (EngineCore_DP0 pid=16042) INFO 12-04 10:00:12 [gpu_model_runner.py:2840] Starting to load model Alibaba-NLP/gte-base-en-v1.5... 2025-12-04T10:00:12.8349059Z (EngineCore_DP0 pid=16042) INFO 12-04 10:00:12 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:00:12.9022572Z (EngineCore_DP0 pid=16042) INFO 12-04 10:00:12 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:00:13.5055773Z (EngineCore_DP0 pid=16042) INFO 12-04 10:00:13 [gpu_model_runner.py:2902] Model loading took 0.0624 GiB and 0.128457 seconds 2025-12-04T10:00:13.6152762Z (EngineCore_DP0 pid=16042) INFO 12-04 10:00:13 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T10:00:13.6179857Z (EngineCore_DP0 pid=16042) INFO 12-04 10:00:13 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T10:00:13.8995335Z (EngineCore_DP0 pid=16042) INFO 12-04 10:00:13 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:00:13.9009172Z INFO 12-04 10:00:13 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-12-04T10:00:14.3341423Z PASSED 2025-12-04T10:00:14.3490768Z models/test_initialization.py::test_can_initialize_large_subset[SmolVLMForConditionalGeneration] Fork a new process to run a test 16112 2025-12-04T10:00:14.3500664Z Fork a new process to run a test 0 2025-12-04T10:00:14.3779043Z INFO 12-04 10:00:14 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='SmolVLMForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'HuggingFaceTB/SmolVLM2-2.2B-Instruct'} 2025-12-04T10:00:14.4566299Z 2025-12-04T10:00:14.4568191Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:00:14.4568506Z config.json: 3.64kB [00:00, 20.8MB/s] 2025-12-04T10:00:14.5807313Z 2025-12-04T10:00:14.5807618Z preprocessor_config.json: 0% 0.00/599 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:00:23.3946119Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:23.3974690Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:23.3981815Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:23.3990085Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:23.3996947Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:23.4005184Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:23.4008024Z (EngineCore_DP0 pid=16170) INFO 12-04 10:00:23 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:00:23.6811823Z (EngineCore_DP0 pid=16170) INFO 12-04 10:00:23 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:00:23.8089641Z (EngineCore_DP0 pid=16170) 2025-12-04T10:00:23.8090427Z processor_config.json: 0% 0.00/67.0 [00:00, model_arch='GPTBigCodeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'bigcode/starcoder'} 2025-12-04T10:00:27.1532674Z 2025-12-04T10:00:27.1533672Z config.json: 0% 0.00/1.05k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:00:35.9867956Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:35.9895975Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:35.9903590Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:35.9910886Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:35.9918643Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:35.9926699Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:35.9930065Z (EngineCore_DP0 pid=16314) INFO 12-04 10:00:35 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:00:36.2864320Z (EngineCore_DP0 pid=16314) INFO 12-04 10:00:36 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:00:36.3829600Z (EngineCore_DP0 pid=16314) INFO 12-04 10:00:36 [gpu_model_runner.py:2840] Starting to load model bigcode/starcoder... 2025-12-04T10:00:36.6701277Z (EngineCore_DP0 pid=16314) INFO 12-04 10:00:36 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:00:36.6744389Z (EngineCore_DP0 pid=16314) INFO 12-04 10:00:36 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:00:37.2864133Z (EngineCore_DP0 pid=16314) INFO 12-04 10:00:37 [gpu_model_runner.py:2902] Model loading took 1.3634 GiB and 0.063262 seconds 2025-12-04T10:00:37.2868172Z (EngineCore_DP0 pid=16314) INFO 12-04 10:00:37 [kv_cache_utils.py:1199] GPU KV cache size: 20,971,520 tokens 2025-12-04T10:00:37.2869077Z (EngineCore_DP0 pid=16314) INFO 12-04 10:00:37 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 2560.00x 2025-12-04T10:00:39.8014471Z (EngineCore_DP0 pid=16314) INFO 12-04 10:00:39 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:00:39.8028933Z INFO 12-04 10:00:39 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:00:40.2769006Z PASSED 2025-12-04T10:00:40.2919463Z models/test_initialization.py::test_can_initialize_large_subset[HunYuanDenseV1ForCausalLM] Fork a new process to run a test 16384 2025-12-04T10:00:40.2930082Z Fork a new process to run a test 0 2025-12-04T10:00:40.2931808Z Model is not available online 2025-12-04T10:00:40.5776291Z PASSED 2025-12-04T10:00:40.5926089Z models/test_initialization.py::test_can_initialize_large_subset[MedusaModel] Fork a new process to run a test 16385 2025-12-04T10:00:40.5935625Z Fork a new process to run a test 0 2025-12-04T10:00:40.6207434Z INFO 12-04 10:00:40 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MedusaModel', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'abhigoyal/vllm-medusa-llama-68m-random', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'JackFram/llama-68m'} 2025-12-04T10:00:40.7218860Z 2025-12-04T10:00:40.7220741Z config.json: 0% 0.00/545 [00:00. This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message 2025-12-04T10:00:48.3839117Z You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message. 2025-12-04T10:00:48.6017632Z 2025-12-04T10:00:48.6018022Z generation_config.json: 0% 0.00/137 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:00:49.9610603Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:49.9637467Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:49.9644334Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:49.9651825Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:49.9659290Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:49.9667191Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:49.9669972Z (EngineCore_DP0 pid=16463) INFO 12-04 10:00:49 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:00:50.2513633Z (EngineCore_DP0 pid=16463) INFO 12-04 10:00:50 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:00:50.2532915Z (EngineCore_DP0 pid=16463) WARNING 12-04 10:00:50 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-12-04T10:00:50.2668752Z (EngineCore_DP0 pid=16463) INFO 12-04 10:00:50 [gpu_model_runner.py:2840] Starting to load model JackFram/llama-68m... 2025-12-04T10:00:50.5396154Z (EngineCore_DP0 pid=16463) INFO 12-04 10:00:50 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:00:50.6066296Z (EngineCore_DP0 pid=16463) INFO 12-04 10:00:50 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:00:50.6659757Z (EngineCore_DP0 pid=16463) INFO 12-04 10:00:50 [gpu_model_runner.py:2879] Loading drafter model... 2025-12-04T10:00:50.6671860Z (EngineCore_DP0 pid=16463) WARNING 12-04 10:00:50 [vllm.py:821] `torch.compile` is turned on, but the model JackFram/llama-68m does not support it. Please open an issue on GitHub if you want it to be supported. 2025-12-04T10:00:51.2151555Z (EngineCore_DP0 pid=16463) INFO 12-04 10:00:51 [gpu_model_runner.py:2902] Model loading took 0.1563 GiB and 0.128094 seconds 2025-12-04T10:00:51.2156011Z (EngineCore_DP0 pid=16463) INFO 12-04 10:00:51 [kv_cache_utils.py:1199] GPU KV cache size: 3,495,248 tokens 2025-12-04T10:00:51.2156844Z (EngineCore_DP0 pid=16463) INFO 12-04 10:00:51 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1706.66x 2025-12-04T10:00:52.3241335Z (EngineCore_DP0 pid=16463) INFO 12-04 10:00:52 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:00:52.3255841Z INFO 12-04 10:00:52 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:00:52.7691017Z PASSED 2025-12-04T10:00:52.7840208Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2ForRewardModel] Fork a new process to run a test 16533 2025-12-04T10:00:52.7850924Z Fork a new process to run a test 0 2025-12-04T10:00:52.7854897Z `transformers==4.56.2` installed, but `transformers<=4.53` is required to run this model. Reason: HF model uses remote code that is not compatible with latest Transformers 2025-12-04T10:00:53.0690698Z PASSED 2025-12-04T10:00:53.0840728Z models/test_initialization.py::test_can_initialize_large_subset[Qwen3VLMoeForConditionalGeneration] Fork a new process to run a test 16534 2025-12-04T10:00:53.0850555Z Fork a new process to run a test 0 2025-12-04T10:00:53.0851972Z Model is not available online 2025-12-04T10:00:53.3674349Z PASSED 2025-12-04T10:00:53.3823798Z models/test_initialization.py::test_can_initialize_large_subset[RWForCausalLM] Fork a new process to run a test 16535 2025-12-04T10:00:53.3834188Z Fork a new process to run a test 0 2025-12-04T10:00:53.4108472Z INFO 12-04 10:00:53 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='RWForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'tiiuae/falcon-40b'} 2025-12-04T10:00:53.5681685Z 2025-12-04T10:00:53.5683940Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:00:53.5684242Z config.json: 1.05kB [00:00, 5.11MB/s] 2025-12-04T10:00:53.6696889Z INFO 12-04 10:00:53 [model.py:653] Resolved architecture: FalconForCausalLM 2025-12-04T10:00:53.6697396Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:00:53.6960422Z INFO 12-04 10:00:53 [model.py:1714] Using max model len 2048 2025-12-04T10:00:53.8778113Z INFO 12-04 10:00:53 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:00:53.9218923Z 2025-12-04T10:00:53.9221239Z tokenizer_config.json: 0% 0.00/242 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:00:55.6978442Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:55.7005200Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:55.7012653Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:55.7020446Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:55.7028080Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:55.7035665Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:00:55.7038430Z (EngineCore_DP0 pid=16543) INFO 12-04 10:00:55 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:00:55.9866608Z (EngineCore_DP0 pid=16543) INFO 12-04 10:00:55 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:00:56.0986701Z (EngineCore_DP0 pid=16543) INFO 12-04 10:00:56 [gpu_model_runner.py:2840] Starting to load model tiiuae/falcon-40b... 2025-12-04T10:00:56.3757095Z (EngineCore_DP0 pid=16543) INFO 12-04 10:00:56 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:00:56.4422590Z (EngineCore_DP0 pid=16543) INFO 12-04 10:00:56 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:00:57.0491885Z (EngineCore_DP0 pid=16543) INFO 12-04 10:00:57 [gpu_model_runner.py:2902] Model loading took 2.2582 GiB and 0.127158 seconds 2025-12-04T10:00:57.0496546Z (EngineCore_DP0 pid=16543) INFO 12-04 10:00:57 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:00:57.0497428Z (EngineCore_DP0 pid=16543) INFO 12-04 10:00:57 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 2560.00x 2025-12-04T10:00:58.0459945Z (EngineCore_DP0 pid=16543) INFO 12-04 10:00:58 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:00:58.0474383Z INFO 12-04 10:00:58 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:00:58.5205970Z PASSED 2025-12-04T10:00:58.5354950Z models/test_initialization.py::test_can_initialize_large_subset[GraniteMoeForCausalLM] Fork a new process to run a test 16613 2025-12-04T10:00:58.5364016Z Fork a new process to run a test 0 2025-12-04T10:00:58.5631814Z INFO 12-04 10:00:58 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GraniteMoeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ibm/PowerMoE-3b'} 2025-12-04T10:00:58.8106309Z 2025-12-04T10:00:58.8107135Z config.json: 0% 0.00/928 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:01:07.6280327Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:07.6307233Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:07.6314545Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:07.6321977Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:07.6329689Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:07.6337307Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:07.6340452Z (EngineCore_DP0 pid=16671) INFO 12-04 10:01:07 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:01:07.9167738Z (EngineCore_DP0 pid=16671) INFO 12-04 10:01:07 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:01:07.9645985Z (EngineCore_DP0 pid=16671) INFO 12-04 10:01:07 [gpu_model_runner.py:2840] Starting to load model ibm/PowerMoE-3b... 2025-12-04T10:01:08.2420556Z (EngineCore_DP0 pid=16671) INFO 12-04 10:01:08 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:01:08.3089350Z (EngineCore_DP0 pid=16671) INFO 12-04 10:01:08 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:01:08.9089349Z (EngineCore_DP0 pid=16671) INFO 12-04 10:01:08 [gpu_model_runner.py:2902] Model loading took 0.1616 GiB and 0.125450 seconds 2025-12-04T10:01:08.9093484Z (EngineCore_DP0 pid=16671) INFO 12-04 10:01:08 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:01:08.9094256Z (EngineCore_DP0 pid=16671) INFO 12-04 10:01:08 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 1280.00x 2025-12-04T10:01:09.8818503Z (EngineCore_DP0 pid=16671) INFO 12-04 10:01:09 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:01:09.8834409Z INFO 12-04 10:01:09 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:01:10.3333999Z PASSED 2025-12-04T10:01:10.3483477Z models/test_initialization.py::test_can_initialize_large_subset[ModernBertForSequenceClassification] Fork a new process to run a test 16741 2025-12-04T10:01:10.3493063Z Fork a new process to run a test 0 2025-12-04T10:01:10.3761964Z INFO 12-04 10:01:10 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='ModernBertForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Alibaba-NLP/gte-reranker-modernbert-base'} 2025-12-04T10:01:10.4536179Z 2025-12-04T10:01:10.4538516Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:01:10.4538810Z config.json: 1.33kB [00:00, 6.71MB/s] 2025-12-04T10:01:16.8998835Z INFO 12-04 10:01:16 [model.py:915] Resolved `--runner auto` to `--runner pooling`. Pass the value explicitly to silence this message. 2025-12-04T10:01:16.8999637Z INFO 12-04 10:01:16 [model.py:653] Resolved architecture: ModernBertForSequenceClassification 2025-12-04T10:01:16.9000156Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:01:16.9250884Z INFO 12-04 10:01:16 [model.py:1939] Downcasting torch.float32 to torch.float16. 2025-12-04T10:01:16.9583615Z 2025-12-04T10:01:16.9585992Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:01:16.9586359Z tokenizer_config.json: 21.0kB [00:00, 109MB/s] 2025-12-04T10:01:16.9594867Z INFO 12-04 10:01:16 [model.py:1714] Using max model len 8192 2025-12-04T10:01:16.9878414Z INFO 12-04 10:01:16 [arg_utils.py:1725] (Disabling) chunked prefill by default 2025-12-04T10:01:16.9878949Z INFO 12-04 10:01:16 [arg_utils.py:1728] (Disabling) prefix caching by default 2025-12-04T10:01:17.1298557Z INFO 12-04 10:01:17 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T10:01:17.2705105Z 2025-12-04T10:01:17.2848928Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T10:01:17.2849280Z tokenizer.json: 3.58MB [00:00, 250MB/s] 2025-12-04T10:01:17.3776032Z 2025-12-04T10:01:17.3778249Z special_tokens_map.json: 0% 0.00/694 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:01:19.0111946Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:19.0143056Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:19.0150695Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:19.0157778Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:19.0165286Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:19.0173244Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:19.0176244Z (EngineCore_DP0 pid=16800) INFO 12-04 10:01:19 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:01:19.2968348Z (EngineCore_DP0 pid=16800) INFO 12-04 10:01:19 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:01:19.3168794Z (EngineCore_DP0 pid=16800) INFO 12-04 10:01:19 [gpu_model_runner.py:2840] Starting to load model Alibaba-NLP/gte-reranker-modernbert-base... 2025-12-04T10:01:19.5965311Z (EngineCore_DP0 pid=16800) INFO 12-04 10:01:19 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:01:19.6633711Z (EngineCore_DP0 pid=16800) INFO 12-04 10:01:19 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:01:20.2695498Z (EngineCore_DP0 pid=16800) INFO 12-04 10:01:20 [gpu_model_runner.py:2902] Model loading took 0.0837 GiB and 0.127460 seconds 2025-12-04T10:01:20.4428585Z (EngineCore_DP0 pid=16800) INFO 12-04 10:01:20 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T10:01:20.4455775Z (EngineCore_DP0 pid=16800) INFO 12-04 10:01:20 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T10:01:20.7268945Z (EngineCore_DP0 pid=16800) INFO 12-04 10:01:20 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:01:20.7288670Z INFO 12-04 10:01:20 [llm.py:337] Supported tasks: ['encode', 'classify', 'score'] 2025-12-04T10:01:21.1675464Z PASSED 2025-12-04T10:01:21.3874046Z models/test_initialization.py::test_can_initialize_large_subset[Lfm2ForCausalLM] SKIPPED 2025-12-04T10:01:21.4022120Z models/test_initialization.py::test_can_initialize_large_subset[Cohere2ForCausalLM] Fork a new process to run a test 16870 2025-12-04T10:01:21.4032392Z Fork a new process to run a test 0 2025-12-04T10:01:21.4307700Z INFO 12-04 10:01:21 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Cohere2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'CohereForAI/c4ai-command-r7b-12-2024'} 2025-12-04T10:01:21.6202703Z 2025-12-04T10:01:21.6203747Z config.json: 0% 0.00/1.02k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:01:25.4395849Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:25.4423993Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:25.4431442Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:25.4439658Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:25.4447578Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:25.4455014Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:25.4458229Z (EngineCore_DP0 pid=16898) INFO 12-04 10:01:25 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:01:25.7245647Z (EngineCore_DP0 pid=16898) INFO 12-04 10:01:25 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:01:25.8396079Z (EngineCore_DP0 pid=16898) INFO 12-04 10:01:25 [gpu_model_runner.py:2840] Starting to load model CohereForAI/c4ai-command-r7b-12-2024... 2025-12-04T10:01:26.1134583Z (EngineCore_DP0 pid=16898) INFO 12-04 10:01:26 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:01:26.1972387Z (EngineCore_DP0 pid=16898) INFO 12-04 10:01:26 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:01:26.7826741Z (EngineCore_DP0 pid=16898) INFO 12-04 10:01:26 [gpu_model_runner.py:2902] Model loading took 2.3909 GiB and 0.131228 seconds 2025-12-04T10:01:26.7831708Z (EngineCore_DP0 pid=16898) INFO 12-04 10:01:26 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:01:28.0774309Z (EngineCore_DP0 pid=16898) INFO 12-04 10:01:26 [kv_cache_utils.py:1204] Maximum concurrency for 132,096 tokens per request: 213.06x 2025-12-04T10:01:28.0775247Z (EngineCore_DP0 pid=16898) INFO 12-04 10:01:28 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:01:28.0788120Z INFO 12-04 10:01:28 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:01:28.5949419Z PASSED 2025-12-04T10:01:28.6098873Z models/test_initialization.py::test_can_initialize_large_subset[FuyuForCausalLM] Fork a new process to run a test 16968 2025-12-04T10:01:28.6108802Z Fork a new process to run a test 0 2025-12-04T10:01:28.6383161Z INFO 12-04 10:01:28 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='FuyuForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'adept/fuyu-8b'} 2025-12-04T10:01:28.7079119Z 2025-12-04T10:01:28.7080351Z config.json: 0% 0.00/768 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:01:39.2583713Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:39.2612164Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:39.2619728Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:39.2627503Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:39.2635342Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:39.2643684Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:39.2646727Z (EngineCore_DP0 pid=17047) INFO 12-04 10:01:39 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:01:39.5513829Z (EngineCore_DP0 pid=17047) INFO 12-04 10:01:39 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:01:43.7009786Z (EngineCore_DP0 pid=17047) INFO 12-04 10:01:43 [gpu_model_runner.py:2840] Starting to load model adept/fuyu-8b... 2025-12-04T10:01:43.9806845Z (EngineCore_DP0 pid=17047) INFO 12-04 10:01:43 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:01:44.0109581Z (EngineCore_DP0 pid=17047) INFO 12-04 10:01:44 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:01:44.6201926Z (EngineCore_DP0 pid=17047) INFO 12-04 10:01:44 [gpu_model_runner.py:2902] Model loading took 4.3976 GiB and 0.089338 seconds 2025-12-04T10:01:44.6205876Z (EngineCore_DP0 pid=17047) INFO 12-04 10:01:44 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-12-04T10:01:44.6206642Z (EngineCore_DP0 pid=17047) INFO 12-04 10:01:44 [kv_cache_utils.py:1204] Maximum concurrency for 16,384 tokens per request: 40.00x 2025-12-04T10:01:45.7386501Z (EngineCore_DP0 pid=17047) INFO 12-04 10:01:45 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:01:45.7401989Z INFO 12-04 10:01:45 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:01:46.3856313Z PASSED 2025-12-04T10:01:46.4004548Z models/test_initialization.py::test_can_initialize_large_subset[TransformersMoEForSequenceClassification] Fork a new process to run a test 17133 2025-12-04T10:01:46.4014814Z Fork a new process to run a test 0 2025-12-04T10:01:46.4018804Z `transformers==4.56.2` installed, but `transformers>=4.57.0.dev0` is required to run this model. 2025-12-04T10:01:46.6866439Z PASSED 2025-12-04T10:01:46.7015160Z models/test_initialization.py::test_can_initialize_large_subset[BloomForCausalLM] Fork a new process to run a test 17134 2025-12-04T10:01:46.7024899Z Fork a new process to run a test 0 2025-12-04T10:01:46.7301139Z INFO 12-04 10:01:46 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='BloomForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'bigscience/bloom-560m'} 2025-12-04T10:01:46.8858003Z 2025-12-04T10:01:46.8859048Z config.json: 0% 0.00/693 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:01:56.4204824Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:56.4232147Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:56.4239448Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:56.4247170Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:56.4255009Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:56.4262821Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:01:56.4266096Z (EngineCore_DP0 pid=17212) INFO 12-04 10:01:56 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:01:56.7094553Z (EngineCore_DP0 pid=17212) INFO 12-04 10:01:56 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:01:56.7551015Z (EngineCore_DP0 pid=17212) INFO 12-04 10:01:56 [gpu_model_runner.py:2840] Starting to load model bigscience/bloom-560m... 2025-12-04T10:01:57.0273633Z (EngineCore_DP0 pid=17212) INFO 12-04 10:01:57 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:01:57.0670760Z (EngineCore_DP0 pid=17212) INFO 12-04 10:01:57 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:01:57.6700791Z (EngineCore_DP0 pid=17212) INFO 12-04 10:01:57 [gpu_model_runner.py:2902] Model loading took 0.5020 GiB and 0.098258 seconds 2025-12-04T10:01:57.6705760Z (EngineCore_DP0 pid=17212) INFO 12-04 10:01:57 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:01:57.6706719Z (EngineCore_DP0 pid=17212) INFO 12-04 10:01:57 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1280.00x 2025-12-04T10:01:59.0516844Z (EngineCore_DP0 pid=17212) INFO 12-04 10:01:59 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:01:59.0531369Z INFO 12-04 10:01:59 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:01:59.5241239Z PASSED 2025-12-04T10:01:59.5390520Z models/test_initialization.py::test_can_initialize_large_subset[GlmForCausalLM] Fork a new process to run a test 17282 2025-12-04T10:01:59.5401094Z Fork a new process to run a test 0 2025-12-04T10:01:59.5677314Z INFO 12-04 10:01:59 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GlmForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'zai-org/glm-4-9b-chat-hf'} 2025-12-04T10:01:59.6679217Z 2025-12-04T10:01:59.6680504Z config.json: 0% 0.00/661 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:02:09.1221928Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:09.1249639Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:09.1257143Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:09.1264683Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:09.1272366Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:09.1279865Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:09.1283551Z (EngineCore_DP0 pid=17360) INFO 12-04 10:02:09 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:02:09.4082872Z (EngineCore_DP0 pid=17360) INFO 12-04 10:02:09 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:02:09.5069341Z (EngineCore_DP0 pid=17360) INFO 12-04 10:02:09 [gpu_model_runner.py:2840] Starting to load model zai-org/glm-4-9b-chat-hf... 2025-12-04T10:02:09.7816854Z (EngineCore_DP0 pid=17360) INFO 12-04 10:02:09 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:02:09.8684008Z (EngineCore_DP0 pid=17360) INFO 12-04 10:02:09 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:02:10.4866717Z (EngineCore_DP0 pid=17360) INFO 12-04 10:02:10 [gpu_model_runner.py:2902] Model loading took 2.7090 GiB and 0.169022 seconds 2025-12-04T10:02:10.4871094Z (EngineCore_DP0 pid=17360) INFO 12-04 10:02:10 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T10:02:10.4872650Z (EngineCore_DP0 pid=17360) INFO 12-04 10:02:10 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 80.00x 2025-12-04T10:02:12.3601802Z (EngineCore_DP0 pid=17360) INFO 12-04 10:02:12 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:02:12.3616110Z INFO 12-04 10:02:12 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:02:12.8702054Z PASSED 2025-12-04T10:02:12.8852269Z models/test_initialization.py::test_can_initialize_large_subset[SmolLM3ForCausalLM] Fork a new process to run a test 17430 2025-12-04T10:02:12.8863339Z Fork a new process to run a test 0 2025-12-04T10:02:12.9134400Z INFO 12-04 10:02:12 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='SmolLM3ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'HuggingFaceTB/SmolLM3-3B'} 2025-12-04T10:02:13.0228733Z 2025-12-04T10:02:13.0231132Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:02:13.0231469Z config.json: 1.92kB [00:00, 9.74MB/s] 2025-12-04T10:02:13.1333170Z INFO 12-04 10:02:13 [model.py:653] Resolved architecture: SmolLM3ForCausalLM 2025-12-04T10:02:13.1333627Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:02:13.1578845Z INFO 12-04 10:02:13 [model.py:1714] Using max model len 65536 2025-12-04T10:02:13.3217696Z INFO 12-04 10:02:13 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:02:13.3570854Z 2025-12-04T10:02:13.3574547Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:02:13.3574918Z tokenizer_config.json: 50.4kB [00:00, 118MB/s] 2025-12-04T10:02:13.4412717Z 2025-12-04T10:02:13.6857839Z tokenizer.json: 0% 0.00/17.2M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:02:15.8478385Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:15.8505644Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:15.8512722Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:15.8519734Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:15.8527448Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:15.8534612Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:15.8537310Z (EngineCore_DP0 pid=17459) INFO 12-04 10:02:15 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:02:16.1377455Z (EngineCore_DP0 pid=17459) INFO 12-04 10:02:16 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:02:16.2040882Z (EngineCore_DP0 pid=17459) INFO 12-04 10:02:16 [gpu_model_runner.py:2840] Starting to load model HuggingFaceTB/SmolLM3-3B... 2025-12-04T10:02:16.4838560Z (EngineCore_DP0 pid=17459) INFO 12-04 10:02:16 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:02:16.4840401Z (EngineCore_DP0 pid=17459) INFO 12-04 10:02:16 [transformers.py:493] Using Transformers backend. 2025-12-04T10:02:16.5206616Z (EngineCore_DP0 pid=17459) `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:02:16.6023731Z (EngineCore_DP0 pid=17459) INFO 12-04 10:02:16 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:02:17.2122041Z (EngineCore_DP0 pid=17459) INFO 12-04 10:02:17 [gpu_model_runner.py:2902] Model loading took 0.6358 GiB and 0.177932 seconds 2025-12-04T10:02:17.2126705Z (EngineCore_DP0 pid=17459) INFO 12-04 10:02:17 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:02:17.2127669Z (EngineCore_DP0 pid=17459) INFO 12-04 10:02:17 [kv_cache_utils.py:1204] Maximum concurrency for 65,536 tokens per request: 80.00x 2025-12-04T10:02:18.5371254Z (EngineCore_DP0 pid=17459) INFO 12-04 10:02:18 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:02:18.5385234Z INFO 12-04 10:02:18 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:02:19.0245133Z PASSED 2025-12-04T10:02:19.0394468Z models/test_initialization.py::test_can_initialize_large_subset[ChatGLMModel] Fork a new process to run a test 17529 2025-12-04T10:02:19.0404338Z Fork a new process to run a test 0 2025-12-04T10:02:19.0408316Z `transformers==4.56.2` installed, but `transformers<=4.48` is required to run this model. 2025-12-04T10:02:19.3234399Z PASSED 2025-12-04T10:02:19.3384618Z models/test_initialization.py::test_can_initialize_large_subset[TransformersMoEForMultimodalLM] Fork a new process to run a test 17530 2025-12-04T10:02:19.3394606Z Fork a new process to run a test 0 2025-12-04T10:02:19.3398440Z `transformers==4.56.2` installed, but `transformers>=4.57.0.dev0` is required to run this model. 2025-12-04T10:02:19.6219512Z PASSED 2025-12-04T10:02:19.6368849Z models/test_initialization.py::test_can_initialize_large_subset[Lfm2MoeForCausalLM] Fork a new process to run a test 17531 2025-12-04T10:02:19.6379972Z Fork a new process to run a test 0 2025-12-04T10:02:19.6383453Z `transformers==4.56.2` installed, but `transformers>=4.58` is required to run this model. 2025-12-04T10:02:19.9211992Z PASSED 2025-12-04T10:02:19.9361190Z models/test_initialization.py::test_can_initialize_large_subset[Eagle3LlamaForCausalLM] Fork a new process to run a test 17532 2025-12-04T10:02:19.9371902Z Fork a new process to run a test 0 2025-12-04T10:02:19.9645428Z INFO 12-04 10:02:19 [utils.py:239] non-default args: {'tokenizer': 'meta-llama/Llama-3.1-8B-Instruct', 'trust_remote_code': True, 'load_format': 'dummy', 'max_model_len': 10240, 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Eagle3LlamaForCausalLM', exist_overrides={}, use_original_num_layers=True), 'speculative_config': {'model': 'yuhuili/EAGLE3-LLaMA3.1-Instruct-8B', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'meta-llama/Llama-3.1-8B-Instruct'} 2025-12-04T10:02:20.0597840Z 2025-12-04T10:02:20.0599154Z config.json: 0% 0.00/855 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:02:23.2530542Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:23.2558559Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:23.2565999Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:23.2573521Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:23.2581413Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:23.2589083Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:23.2591906Z (EngineCore_DP0 pid=17540) INFO 12-04 10:02:23 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:02:23.5503333Z (EngineCore_DP0 pid=17540) INFO 12-04 10:02:23 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:02:23.5657049Z (EngineCore_DP0 pid=17540) WARNING 12-04 10:02:23 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-12-04T10:02:23.6036932Z (EngineCore_DP0 pid=17540) INFO 12-04 10:02:23 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.1-8B-Instruct... 2025-12-04T10:02:23.8832090Z (EngineCore_DP0 pid=17540) INFO 12-04 10:02:23 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:02:23.9669089Z (EngineCore_DP0 pid=17540) INFO 12-04 10:02:23 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:02:24.0803125Z (EngineCore_DP0 pid=17540) INFO 12-04 10:02:24 [gpu_model_runner.py:2879] Loading drafter model... 2025-12-04T10:02:24.1470618Z (EngineCore_DP0 pid=17540) INFO 12-04 10:02:24 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-12-04T10:02:24.1471834Z (EngineCore_DP0 pid=17540) INFO 12-04 10:02:24 [eagle.py:1043] The EAGLE head's lm_head will be loaded separately from the target model. 2025-12-04T10:02:24.6986646Z (EngineCore_DP0 pid=17540) INFO 12-04 10:02:24 [gpu_model_runner.py:2902] Model loading took 15.7807 GiB and 0.264226 seconds 2025-12-04T10:02:24.6994505Z (EngineCore_DP0 pid=17540) INFO 12-04 10:02:24 [kv_cache_utils.py:1199] GPU KV cache size: 79,424 tokens 2025-12-04T10:02:24.6995487Z (EngineCore_DP0 pid=17540) INFO 12-04 10:02:24 [kv_cache_utils.py:1204] Maximum concurrency for 10,240 tokens per request: 7.76x 2025-12-04T10:02:25.3950923Z (EngineCore_DP0 pid=17540) INFO 12-04 10:02:25 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:02:25.3966179Z INFO 12-04 10:02:25 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:02:25.8730207Z PASSED 2025-12-04T10:02:25.8879311Z models/test_initialization.py::test_can_initialize_large_subset[OrionForCausalLM] Fork a new process to run a test 17610 2025-12-04T10:02:25.8888345Z Fork a new process to run a test 0 2025-12-04T10:02:25.9161968Z INFO 12-04 10:02:25 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='OrionForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'OrionStarAI/Orion-14B-Chat'} 2025-12-04T10:02:26.0472565Z 2025-12-04T10:02:26.0476184Z config.json: 0% 0.00/771 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:02:35.1466920Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:35.1494079Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:35.1501352Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:35.1509126Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:35.1516585Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:35.1524050Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:35.1526624Z (EngineCore_DP0 pid=17688) INFO 12-04 10:02:35 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:02:35.4390876Z (EngineCore_DP0 pid=17688) INFO 12-04 10:02:35 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:02:35.5371120Z (EngineCore_DP0 pid=17688) INFO 12-04 10:02:35 [gpu_model_runner.py:2840] Starting to load model OrionStarAI/Orion-14B-Chat... 2025-12-04T10:02:35.8209618Z (EngineCore_DP0 pid=17688) INFO 12-04 10:02:35 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:02:35.8865186Z (EngineCore_DP0 pid=17688) INFO 12-04 10:02:35 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:02:36.5043201Z (EngineCore_DP0 pid=17688) INFO 12-04 10:02:36 [gpu_model_runner.py:2902] Model loading took 2.2496 GiB and 0.124662 seconds 2025-12-04T10:02:36.5047805Z (EngineCore_DP0 pid=17688) INFO 12-04 10:02:36 [kv_cache_utils.py:1199] GPU KV cache size: 524,288 tokens 2025-12-04T10:02:36.5048746Z (EngineCore_DP0 pid=17688) INFO 12-04 10:02:36 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 128.00x 2025-12-04T10:02:36.7607760Z (EngineCore_DP0 pid=17688) WARNING 12-04 10:02:36 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-12-04T10:02:37.2654432Z (EngineCore_DP0 pid=17688) INFO 12-04 10:02:37 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:02:37.2671549Z INFO 12-04 10:02:37 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:02:37.7528556Z PASSED 2025-12-04T10:02:37.7677650Z models/test_initialization.py::test_can_initialize_large_subset[PhiMoEForCausalLM] Fork a new process to run a test 17758 2025-12-04T10:02:37.7688005Z Fork a new process to run a test 0 2025-12-04T10:02:37.7963390Z INFO 12-04 10:02:37 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='PhiMoEForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'microsoft/Phi-3.5-MoE-instruct'} 2025-12-04T10:02:37.9032225Z 2025-12-04T10:02:37.9034387Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:02:37.9034771Z config.json: 4.53kB [00:00, 24.6MB/s] 2025-12-04T10:02:37.9781705Z 2025-12-04T10:02:37.9784552Z configuration_phimoe.py: 0.00B [00:00, ?B/s] 2025-12-04T10:02:37.9784966Z configuration_phimoe.py: 12.3kB [00:00, 40.5MB/s] 2025-12-04T10:02:37.9889059Z A new version of the following files was downloaded from https://huggingface.co/microsoft/Phi-3.5-MoE-instruct: 2025-12-04T10:02:37.9889648Z - configuration_phimoe.py 2025-12-04T10:02:37.9890280Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:02:38.0208396Z INFO 12-04 10:02:38 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T10:02:44.4958583Z INFO 12-04 10:02:44 [model.py:653] Resolved architecture: PhiMoEForCausalLM 2025-12-04T10:02:44.4959079Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:02:44.5216795Z INFO 12-04 10:02:44 [model.py:1714] Using max model len 131072 2025-12-04T10:02:44.5218474Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T10:02:44.5804301Z INFO 12-04 10:02:44 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:02:44.6147870Z 2025-12-04T10:02:44.6149421Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:02:44.6149977Z tokenizer_config.json: 3.98kB [00:00, 19.9MB/s] 2025-12-04T10:02:44.7340761Z 2025-12-04T10:02:44.8671659Z tokenizer.model: 0% 0.00/500k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:02:46.5834322Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:46.5861650Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:46.5868945Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:46.5876989Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:46.5884691Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:46.5892293Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:02:46.5895189Z (EngineCore_DP0 pid=17836) INFO 12-04 10:02:46 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:02:46.8698506Z (EngineCore_DP0 pid=17836) INFO 12-04 10:02:46 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:02:46.9523015Z (EngineCore_DP0 pid=17836) INFO 12-04 10:02:46 [gpu_model_runner.py:2840] Starting to load model microsoft/Phi-3.5-MoE-instruct... 2025-12-04T10:02:47.2289221Z (EngineCore_DP0 pid=17836) INFO 12-04 10:02:47 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:02:47.2977327Z (EngineCore_DP0 pid=17836) INFO 12-04 10:02:47 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:02:47.9005103Z (EngineCore_DP0 pid=17836) INFO 12-04 10:02:47 [gpu_model_runner.py:2902] Model loading took 0.8927 GiB and 0.128166 seconds 2025-12-04T10:02:47.9010302Z (EngineCore_DP0 pid=17836) INFO 12-04 10:02:47 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:02:47.9011118Z (EngineCore_DP0 pid=17836) INFO 12-04 10:02:47 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 20.00x 2025-12-04T10:02:48.5205454Z (EngineCore_DP0 pid=17836) INFO 12-04 10:02:48 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:02:48.5227615Z INFO 12-04 10:02:48 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:02:48.9890175Z PASSED 2025-12-04T10:02:49.0039498Z models/test_initialization.py::test_can_initialize_large_subset[Emu3ForConditionalGeneration] Fork a new process to run a test 17906 2025-12-04T10:02:49.0049005Z Fork a new process to run a test 0 2025-12-04T10:02:49.0315962Z INFO 12-04 10:02:49 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Emu3ForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'BAAI/Emu3-Chat-hf'} 2025-12-04T10:02:49.6285392Z INFO 12-04 10:02:49 [model.py:653] Resolved architecture: Emu3ForConditionalGeneration 2025-12-04T10:02:49.6285932Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:02:49.6520325Z INFO 12-04 10:02:49 [model.py:1939] Downcasting torch.float32 to torch.bfloat16. 2025-12-04T10:02:49.6521212Z INFO 12-04 10:02:49 [model.py:1714] Using max model len 131072 2025-12-04T10:02:49.8778069Z INFO 12-04 10:02:49 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:03:11.4554943Z (EngineCore_DP0 pid=17913) INFO 12-04 10:03:11 [core.py:727] Waiting for init message from front-end. 2025-12-04T10:03:11.4590854Z (EngineCore_DP0 pid=17913) INFO 12-04 10:03:11 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='BAAI/Emu3-Chat-hf', speculative_config=None, tokenizer='BAAI/Emu3-Chat-hf', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.bfloat16, max_seq_len=131072, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=BAAI/Emu3-Chat-hf, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:03:12.8401058Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:03:12.8429189Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:03:12.8436338Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:03:12.8444145Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:03:12.8451555Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:03:12.8458844Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:03:12.8461808Z (EngineCore_DP0 pid=17913) INFO 12-04 10:03:12 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:03:13.1321300Z (EngineCore_DP0 pid=17913) INFO 12-04 10:03:13 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:03:46.2830792Z (EngineCore_DP0 pid=17913) INFO 12-04 10:03:46 [gpu_model_runner.py:2840] Starting to load model BAAI/Emu3-Chat-hf... 2025-12-04T10:03:46.7045064Z (EngineCore_DP0 pid=17913) INFO 12-04 10:03:46 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:03:46.7046294Z (EngineCore_DP0 pid=17913) INFO 12-04 10:03:46 [transformers.py:493] Using Transformers backend. 2025-12-04T10:03:46.7411511Z (EngineCore_DP0 pid=17913) `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:03:46.9401302Z (EngineCore_DP0 pid=17913) INFO 12-04 10:03:46 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:03:47.5663875Z (EngineCore_DP0 pid=17913) INFO 12-04 10:03:47 [gpu_model_runner.py:2902] Model loading took 3.7283 GiB and 0.319966 seconds 2025-12-04T10:03:47.5668657Z (EngineCore_DP0 pid=17913) INFO 12-04 10:03:47 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:03:47.5669649Z (EngineCore_DP0 pid=17913) INFO 12-04 10:03:47 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 20.00x 2025-12-04T10:03:58.7353652Z (EngineCore_DP0 pid=17913) INFO 12-04 10:03:58 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:03:58.7368046Z INFO 12-04 10:03:58 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:03:59.2609201Z PASSED 2025-12-04T10:03:59.2757479Z models/test_initialization.py::test_can_initialize_large_subset[SeedOssForCausalLM] Fork a new process to run a test 17983 2025-12-04T10:03:59.2768723Z Fork a new process to run a test 0 2025-12-04T10:03:59.2769815Z Model is not available online 2025-12-04T10:03:59.5583636Z PASSED 2025-12-04T10:03:59.5733404Z models/test_initialization.py::test_can_initialize_large_subset[HCXVisionForCausalLM] Fork a new process to run a test 17984 2025-12-04T10:03:59.5743086Z Fork a new process to run a test 0 2025-12-04T10:03:59.6017261Z INFO 12-04 10:03:59 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='HCXVisionForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'naver-hyperclovax/HyperCLOVAX-SEED-Vision-Instruct-3B'} 2025-12-04T10:03:59.7486563Z 2025-12-04T10:03:59.7488802Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:03:59.7489130Z config.json: 5.51kB [00:00, 26.5MB/s] 2025-12-04T10:03:59.8088302Z 2025-12-04T10:03:59.8089064Z configuration_hyperclovax.py: 0.00B [00:00, ?B/s] 2025-12-04T10:03:59.8089483Z configuration_hyperclovax.py: 2.44kB [00:00, 26.7MB/s] 2025-12-04T10:03:59.8187494Z A new version of the following files was downloaded from https://huggingface.co/naver-hyperclovax/HyperCLOVAX-SEED-Vision-Instruct-3B: 2025-12-04T10:03:59.8188209Z - configuration_hyperclovax.py 2025-12-04T10:03:59.8188854Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:04:00.0513949Z 2025-12-04T10:04:00.0514437Z preprocessor_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:04:00.0514886Z preprocessor_config.json: 1.76kB [00:00, 19.4MB/s] 2025-12-04T10:04:06.6055965Z INFO 12-04 10:04:06 [model.py:653] Resolved architecture: HCXVisionForCausalLM 2025-12-04T10:04:06.6056482Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:04:06.6315295Z INFO 12-04 10:04:06 [model.py:1714] Using max model len 131072 2025-12-04T10:04:06.6317808Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T10:04:06.6912898Z INFO 12-04 10:04:06 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:04:06.7276640Z 2025-12-04T10:04:06.7278160Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:04:06.7278501Z tokenizer_config.json: 11.7kB [00:00, 85.7MB/s] 2025-12-04T10:04:06.8105279Z 2025-12-04T10:04:06.8223792Z vocab.json: 0.00B [00:00, ?B/s] 2025-12-04T10:04:06.8224078Z vocab.json: 1.86MB [00:00, 158MB/s] 2025-12-04T10:04:06.8674987Z 2025-12-04T10:04:06.8731379Z merges.txt: 0.00B [00:00, ?B/s] 2025-12-04T10:04:06.8731675Z merges.txt: 1.08MB [00:00, 193MB/s] 2025-12-04T10:04:06.9116359Z 2025-12-04T10:04:06.9366864Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T10:04:06.9367162Z tokenizer.json: 8.03MB [00:00, 322MB/s] 2025-12-04T10:04:06.9780999Z 2025-12-04T10:04:06.9782437Z added_tokens.json: 0% 0.00/925 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:04:09.3532294Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:09.3560640Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:09.3567416Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:09.3575426Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:09.3583289Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:09.3591288Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:09.3594336Z (EngineCore_DP0 pid=18043) INFO 12-04 10:04:09 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:04:09.6565745Z (EngineCore_DP0 pid=18043) INFO 12-04 10:04:09 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:04:10.0976666Z (EngineCore_DP0 pid=18043) 2025-12-04T10:04:10.0977059Z processor_config.json: 0% 0.00/120 [00:00, model_arch='Eagle3Qwen2_5vlForCausalLM', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'Rayzl/qwen2.5-vl-7b-eagle3-sgl', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'Qwen/Qwen2.5-VL-7B-Instruct'} 2025-12-04T10:04:14.9976825Z 2025-12-04T10:04:14.9979038Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:04:14.9979435Z config.json: 1.37kB [00:00, 7.94MB/s] 2025-12-04T10:04:15.1191763Z 2025-12-04T10:04:15.1192277Z preprocessor_config.json: 0% 0.00/350 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:04:24.2510439Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:24.2538960Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:24.2546167Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:24.2553483Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:24.2561282Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:24.2569022Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:24.2572012Z (EngineCore_DP0 pid=18187) INFO 12-04 10:04:24 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:04:24.5403808Z (EngineCore_DP0 pid=18187) INFO 12-04 10:04:24 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:04:24.5552079Z (EngineCore_DP0 pid=18187) WARNING 12-04 10:04:24 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-12-04T10:04:26.1255361Z (EngineCore_DP0 pid=18187) 2025-12-04T10:04:26.1256559Z chat_template.json: 0.00B [00:00, ?B/s] 2025-12-04T10:04:26.1256896Z chat_template.json: 1.05kB [00:00, 3.76MB/s] 2025-12-04T10:04:26.6729419Z (EngineCore_DP0 pid=18187) INFO 12-04 10:04:26 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen2.5-VL-7B-Instruct... 2025-12-04T10:04:27.0243746Z (EngineCore_DP0 pid=18187) INFO 12-04 10:04:27 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:04:27.0893537Z (EngineCore_DP0 pid=18187) INFO 12-04 10:04:27 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:04:27.1548197Z (EngineCore_DP0 pid=18187) INFO 12-04 10:04:27 [gpu_model_runner.py:2879] Loading drafter model... 2025-12-04T10:04:27.1683310Z (EngineCore_DP0 pid=18187) WARNING 12-04 10:04:27 [vllm.py:821] `torch.compile` is turned on, but the model Qwen/Qwen2.5-VL-7B-Instruct does not support it. Please open an issue on GitHub if you want it to be supported. 2025-12-04T10:04:27.1918828Z (EngineCore_DP0 pid=18187) INFO 12-04 10:04:27 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-12-04T10:04:27.1920286Z (EngineCore_DP0 pid=18187) INFO 12-04 10:04:27 [eagle.py:1043] The EAGLE head's lm_head will be loaded separately from the target model. 2025-12-04T10:04:27.7421371Z (EngineCore_DP0 pid=18187) INFO 12-04 10:04:27 [gpu_model_runner.py:2902] Model loading took 4.6155 GiB and 0.167878 seconds 2025-12-04T10:04:27.7425718Z (EngineCore_DP0 pid=18187) INFO 12-04 10:04:27 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:04:27.7426628Z (EngineCore_DP0 pid=18187) INFO 12-04 10:04:27 [kv_cache_utils.py:1204] Maximum concurrency for 128,000 tokens per request: 20.48x 2025-12-04T10:04:28.4664783Z (EngineCore_DP0 pid=18187) INFO 12-04 10:04:28 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:04:28.4681548Z INFO 12-04 10:04:28 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:04:28.9716602Z PASSED 2025-12-04T10:04:28.9866028Z models/test_initialization.py::test_can_initialize_large_subset[Phi3ForCausalLM] Fork a new process to run a test 18257 2025-12-04T10:04:28.9876971Z Fork a new process to run a test 0 2025-12-04T10:04:29.0145618Z INFO 12-04 10:04:29 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Phi3ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'microsoft/Phi-3-mini-4k-instruct'} 2025-12-04T10:04:29.0861531Z 2025-12-04T10:04:29.0862715Z config.json: 0% 0.00/967 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:04:37.7784599Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:37.7811638Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:37.7818605Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:37.7826180Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:37.7833530Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:37.7841388Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:37.7844250Z (EngineCore_DP0 pid=18335) INFO 12-04 10:04:37 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:04:38.0658334Z (EngineCore_DP0 pid=18335) INFO 12-04 10:04:38 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:04:38.1301812Z (EngineCore_DP0 pid=18335) INFO 12-04 10:04:38 [gpu_model_runner.py:2840] Starting to load model microsoft/Phi-3-mini-4k-instruct... 2025-12-04T10:04:38.4028671Z (EngineCore_DP0 pid=18335) INFO 12-04 10:04:38 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:04:38.4687148Z (EngineCore_DP0 pid=18335) INFO 12-04 10:04:38 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:04:39.0815656Z (EngineCore_DP0 pid=18335) INFO 12-04 10:04:39 [gpu_model_runner.py:2902] Model loading took 0.5789 GiB and 0.125394 seconds 2025-12-04T10:04:39.0819979Z (EngineCore_DP0 pid=18335) INFO 12-04 10:04:39 [kv_cache_utils.py:1199] GPU KV cache size: 873,808 tokens 2025-12-04T10:04:39.0821148Z (EngineCore_DP0 pid=18335) INFO 12-04 10:04:39 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 212.50x 2025-12-04T10:04:39.6151642Z (EngineCore_DP0 pid=18335) INFO 12-04 10:04:39 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:04:39.6174315Z INFO 12-04 10:04:39 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:04:40.0601219Z PASSED 2025-12-04T10:04:40.0746288Z models/test_initialization.py::test_can_initialize_large_subset[InternVLForConditionalGeneration] Fork a new process to run a test 18405 2025-12-04T10:04:40.0756669Z Fork a new process to run a test 0 2025-12-04T10:04:40.1031055Z INFO 12-04 10:04:40 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='InternVLForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'OpenGVLab/InternVL3-1B-hf'} 2025-12-04T10:04:40.2293848Z 2025-12-04T10:04:40.2295834Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:04:40.2296133Z config.json: 1.96kB [00:00, 11.9MB/s] 2025-12-04T10:04:40.3372273Z 2025-12-04T10:04:40.3373800Z preprocessor_config.json: 0% 0.00/666 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:04:49.8461734Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:49.8491448Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:49.8499229Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:49.8506518Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:49.8513928Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:49.8521932Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:04:49.8525045Z (EngineCore_DP0 pid=18483) INFO 12-04 10:04:49 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:04:50.1316971Z (EngineCore_DP0 pid=18483) INFO 12-04 10:04:50 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:04:50.2641924Z (EngineCore_DP0 pid=18483) 2025-12-04T10:04:50.2642949Z processor_config.json: 0% 0.00/72.0 [00:00 8192). Running this sequence through the model will result in indexing errors 2025-12-04T10:04:54.8619992Z (EngineCore_DP0 pid=18483) INFO 12-04 10:04:54 [gpu_model_runner.py:2840] Starting to load model OpenGVLab/InternVL3-1B-hf... 2025-12-04T10:04:55.1399677Z (EngineCore_DP0 pid=18483) INFO 12-04 10:04:55 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:04:55.1440176Z (EngineCore_DP0 pid=18483) INFO 12-04 10:04:55 [layer.py:497] MultiHeadAttention attn_backend: _Backend.FLASH_ATTN, use_upstream_fa: False 2025-12-04T10:04:55.1706533Z (EngineCore_DP0 pid=18483) INFO 12-04 10:04:55 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:04:55.7673565Z (EngineCore_DP0 pid=18483) INFO 12-04 10:04:55 [gpu_model_runner.py:2902] Model loading took 0.5808 GiB and 0.090677 seconds 2025-12-04T10:04:55.7677554Z (EngineCore_DP0 pid=18483) INFO 12-04 10:04:55 [kv_cache_utils.py:1199] GPU KV cache size: 20,971,520 tokens 2025-12-04T10:04:55.7678604Z (EngineCore_DP0 pid=18483) INFO 12-04 10:04:55 [kv_cache_utils.py:1204] Maximum concurrency for 65,536 tokens per request: 320.00x 2025-12-04T10:05:00.1869781Z (EngineCore_DP0 pid=18483) INFO 12-04 10:05:00 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:05:00.2122063Z INFO 12-04 10:05:00 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:05:00.7039000Z PASSED 2025-12-04T10:05:00.7187466Z models/test_initialization.py::test_can_initialize_large_subset[AquilaForCausalLM] Fork a new process to run a test 18569 2025-12-04T10:05:00.7196615Z Fork a new process to run a test 0 2025-12-04T10:05:00.7465700Z INFO 12-04 10:05:00 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='AquilaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'BAAI/AquilaChat2-7B'} 2025-12-04T10:05:00.8811821Z 2025-12-04T10:05:00.8813342Z config.json: 0% 0.00/678 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:05:03.3296839Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:03.3324319Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:03.3332023Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:03.3340185Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:03.3348649Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:03.3356785Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:03.3359893Z (EngineCore_DP0 pid=18577) INFO 12-04 10:05:03 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:05:03.6199918Z (EngineCore_DP0 pid=18577) INFO 12-04 10:05:03 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:05:03.6865699Z (EngineCore_DP0 pid=18577) INFO 12-04 10:05:03 [gpu_model_runner.py:2840] Starting to load model BAAI/AquilaChat2-7B... 2025-12-04T10:05:03.9664680Z (EngineCore_DP0 pid=18577) INFO 12-04 10:05:03 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:05:04.0319539Z (EngineCore_DP0 pid=18577) INFO 12-04 10:05:04 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:05:04.6358100Z (EngineCore_DP0 pid=18577) INFO 12-04 10:05:04 [gpu_model_runner.py:2902] Model loading took 1.9048 GiB and 0.125210 seconds 2025-12-04T10:05:04.6362992Z (EngineCore_DP0 pid=18577) INFO 12-04 10:05:04 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-12-04T10:05:04.6363909Z (EngineCore_DP0 pid=18577) INFO 12-04 10:05:04 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 320.00x 2025-12-04T10:05:05.2630431Z (EngineCore_DP0 pid=18577) INFO 12-04 10:05:05 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:05:05.2652781Z INFO 12-04 10:05:05 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:05:05.7191124Z PASSED 2025-12-04T10:05:05.7337802Z models/test_initialization.py::test_can_initialize_large_subset[GLM4VForCausalLM] Fork a new process to run a test 18647 2025-12-04T10:05:05.7348394Z Fork a new process to run a test 0 2025-12-04T10:05:05.7626898Z INFO 12-04 10:05:05 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GLM4VForCausalLM', exist_overrides={'architectures': ['GLM4VForCausalLM']}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'zai-org/glm-4v-9b'} 2025-12-04T10:05:05.8634810Z 2025-12-04T10:05:05.8636952Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:05:05.8637279Z config.json: 1.77kB [00:00, 10.9MB/s] 2025-12-04T10:05:12.4605603Z INFO 12-04 10:05:12 [model.py:653] Resolved architecture: GLM4VForCausalLM 2025-12-04T10:05:12.4606072Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:05:12.4858866Z INFO 12-04 10:05:12 [model.py:1714] Using max model len 8192 2025-12-04T10:05:12.4861521Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T10:05:12.7639031Z INFO 12-04 10:05:12 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:05:12.7981946Z 2025-12-04T10:05:12.7983606Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:05:12.7984019Z tokenizer_config.json: 3.22kB [00:00, 12.7MB/s] 2025-12-04T10:05:12.8362425Z 2025-12-04T10:05:12.8364629Z tokenization_chatglm.py: 0.00B [00:00, ?B/s] 2025-12-04T10:05:12.8365113Z tokenization_chatglm.py: 17.7kB [00:00, 86.9MB/s] 2025-12-04T10:05:12.8509189Z A new version of the following files was downloaded from https://huggingface.co/zai-org/glm-4v-9b: 2025-12-04T10:05:12.8509703Z - tokenization_chatglm.py 2025-12-04T10:05:12.8510340Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:05:12.9745394Z 2025-12-04T10:05:13.2583158Z tokenizer.model: 0% 0.00/2.62M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:05:15.4670235Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:15.4700088Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:15.4706820Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:15.4714016Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:15.4721505Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:15.4729331Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:15.4731730Z (EngineCore_DP0 pid=18725) INFO 12-04 10:05:15 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:05:15.7581472Z (EngineCore_DP0 pid=18725) INFO 12-04 10:05:15 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:05:15.8351816Z (EngineCore_DP0 pid=18725) WARNING 12-04 10:05:15 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T10:05:15.8358335Z (EngineCore_DP0 pid=18725) WARNING 12-04 10:05:15 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T10:05:15.8481355Z (EngineCore_DP0 pid=18725) INFO 12-04 10:05:15 [gpu_model_runner.py:2840] Starting to load model zai-org/glm-4v-9b... 2025-12-04T10:05:16.1261186Z (EngineCore_DP0 pid=18725) INFO 12-04 10:05:16 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:05:16.1835383Z (EngineCore_DP0 pid=18725) INFO 12-04 10:05:16 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:05:16.2544698Z (EngineCore_DP0 pid=18725) INFO 12-04 10:05:16 [layer.py:497] MultiHeadAttention attn_backend: _Backend.XFORMERS, use_upstream_fa: False 2025-12-04T10:05:16.8064411Z (EngineCore_DP0 pid=18725) INFO 12-04 10:05:16 [gpu_model_runner.py:2902] Model loading took 3.2448 GiB and 0.133097 seconds 2025-12-04T10:05:16.8069030Z (EngineCore_DP0 pid=18725) INFO 12-04 10:05:16 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T10:05:16.8069923Z (EngineCore_DP0 pid=18725) INFO 12-04 10:05:16 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 1280.00x 2025-12-04T10:05:17.1624179Z (EngineCore_DP0 pid=18725) WARNING 12-04 10:05:17 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-12-04T10:05:17.1918930Z (EngineCore_DP0 pid=18725) WARNING 12-04 10:05:17 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T10:05:18.4617500Z (EngineCore_DP0 pid=18725) INFO 12-04 10:05:18 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:05:18.4633446Z INFO 12-04 10:05:18 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:05:18.9401658Z PASSED 2025-12-04T10:05:18.9550151Z models/test_initialization.py::test_can_initialize_large_subset[Idefics3ForConditionalGeneration] Fork a new process to run a test 18795 2025-12-04T10:05:18.9560395Z Fork a new process to run a test 0 2025-12-04T10:05:18.9830990Z INFO 12-04 10:05:18 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Idefics3ForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'HuggingFaceM4/Idefics3-8B-Llama3'} 2025-12-04T10:05:19.0738292Z 2025-12-04T10:05:19.0741289Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:05:19.0741718Z config.json: 7.18kB [00:00, 30.1MB/s] 2025-12-04T10:05:19.2008025Z 2025-12-04T10:05:19.2009052Z preprocessor_config.json: 0% 0.00/435 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:05:28.4012547Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:28.4041755Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:28.4049070Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:28.4056827Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:28.4065008Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:28.4072865Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:28.4075681Z (EngineCore_DP0 pid=18853) INFO 12-04 10:05:28 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:05:28.6929844Z (EngineCore_DP0 pid=18853) INFO 12-04 10:05:28 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:05:30.4386659Z (EngineCore_DP0 pid=18853) 2025-12-04T10:05:30.4388950Z chat_template.json: 0% 0.00/434 [00:00, model_arch='CLIPModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'openai/clip-vit-base-patch32'} 2025-12-04T10:05:33.8599569Z 2025-12-04T10:05:33.8601480Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:05:33.8601808Z config.json: 4.19kB [00:00, 19.2MB/s] 2025-12-04T10:05:33.9776878Z 2025-12-04T10:05:33.9779736Z preprocessor_config.json: 0% 0.00/316 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:05:44.7863369Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:44.7890360Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:44.7898060Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:44.7906222Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:44.7914573Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:44.7921812Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:05:44.7924800Z (EngineCore_DP0 pid=18998) INFO 12-04 10:05:44 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:05:45.0786742Z (EngineCore_DP0 pid=18998) INFO 12-04 10:05:45 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:05:45.0933620Z (EngineCore_DP0 pid=18998) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-12-04T10:05:45.9327542Z (EngineCore_DP0 pid=18998) WARNING 12-04 10:05:45 [processing.py:1091] CLIPProcessor did not return `BatchFeature`. Make sure to match the behaviour of `ProcessorMixin` when implementing custom processors. 2025-12-04T10:05:45.9572514Z (EngineCore_DP0 pid=18998) INFO 12-04 10:05:45 [gpu_model_runner.py:2840] Starting to load model openai/clip-vit-base-patch32... 2025-12-04T10:05:46.2401645Z (EngineCore_DP0 pid=18998) INFO 12-04 10:05:46 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:05:46.2424244Z (EngineCore_DP0 pid=18998) INFO 12-04 10:05:46 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:05:46.2486360Z (EngineCore_DP0 pid=18998) INFO 12-04 10:05:46 [layer.py:497] MultiHeadAttention attn_backend: _Backend.FLASH_ATTN, use_upstream_fa: False 2025-12-04T10:05:46.2806144Z (EngineCore_DP0 pid=18998) WARNING 12-04 10:05:46 [vllm.py:821] `torch.compile` is turned on, but the model openai/clip-vit-base-patch32 does not support it. Please open an issue on GitHub if you want it to be supported. 2025-12-04T10:05:46.8308779Z (EngineCore_DP0 pid=18998) INFO 12-04 10:05:46 [gpu_model_runner.py:2902] Model loading took 0.0724 GiB and 0.042451 seconds 2025-12-04T10:05:46.8313153Z (EngineCore_DP0 pid=18998) INFO 12-04 10:05:46 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:05:46.8314528Z (EngineCore_DP0 pid=18998) INFO 12-04 10:05:46 [kv_cache_utils.py:1204] Maximum concurrency for 77 tokens per request: 65536.00x 2025-12-04T10:05:47.7620760Z (EngineCore_DP0 pid=18998) INFO 12-04 10:05:47 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:05:47.7639045Z INFO 12-04 10:05:47 [llm.py:337] Supported tasks: ['embed'] 2025-12-04T10:05:48.1999474Z PASSED 2025-12-04T10:05:48.2146451Z models/test_initialization.py::test_can_initialize_large_subset[LlavaOnevisionForConditionalGeneration] Fork a new process to run a test 19084 2025-12-04T10:05:48.2156879Z Fork a new process to run a test 0 2025-12-04T10:05:48.2422695Z INFO 12-04 10:05:48 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='LlavaOnevisionForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'llava-hf/llava-onevision-qwen2-0.5b-ov-hf'} 2025-12-04T10:05:48.3197293Z 2025-12-04T10:05:48.3199461Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:05:48.3199779Z config.json: 2.59kB [00:00, 15.3MB/s] 2025-12-04T10:05:48.4396166Z 2025-12-04T10:05:48.4397136Z preprocessor_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:05:48.4397525Z preprocessor_config.json: 1.73kB [00:00, 22.3MB/s] 2025-12-04T10:05:55.2055147Z INFO 12-04 10:05:55 [model.py:653] Resolved architecture: LlavaOnevisionForConditionalGeneration 2025-12-04T10:05:55.2055998Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:05:55.2307908Z INFO 12-04 10:05:55 [model.py:1714] Using max model len 32768 2025-12-04T10:05:55.4015361Z INFO 12-04 10:05:55 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:05:55.4422806Z 2025-12-04T10:05:55.4423852Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:05:55.4424250Z tokenizer_config.json: 1.80kB [00:00, 18.5MB/s] 2025-12-04T10:05:55.5210416Z 2025-12-04T10:05:55.5363736Z vocab.json: 0.00B [00:00, ?B/s] 2025-12-04T10:05:55.5364050Z vocab.json: 2.78MB [00:00, 181MB/s] 2025-12-04T10:05:55.5738628Z 2025-12-04T10:05:55.5817296Z merges.txt: 0.00B [00:00, ?B/s] 2025-12-04T10:05:55.5817580Z merges.txt: 1.67MB [00:00, 210MB/s] 2025-12-04T10:05:55.6262993Z 2025-12-04T10:05:55.6501508Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T10:05:55.6501833Z tokenizer.json: 7.03MB [00:00, 295MB/s] 2025-12-04T10:05:55.7090531Z 2025-12-04T10:05:55.7091647Z added_tokens.json: 0% 0.00/122 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:06:04.3716139Z (EngineCore_DP0 pid=19144) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:357: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-12-04T10:06:04.3717472Z (EngineCore_DP0 pid=19144) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-12-04T10:06:04.3719279Z (EngineCore_DP0 pid=19144) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-12-04T10:06:04.3720835Z (EngineCore_DP0 pid=19144) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 2025-12-04T10:06:04.3721619Z (EngineCore_DP0 pid=19144) dispatch key: ADInplaceOrView 2025-12-04T10:06:04.3722548Z (EngineCore_DP0 pid=19144) previous kernel: no debug info 2025-12-04T10:06:04.3723864Z (EngineCore_DP0 pid=19144) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-12-04T10:06:04.3724993Z (EngineCore_DP0 pid=19144) self.m.impl( 2025-12-04T10:06:04.9714669Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:04.9743896Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:04.9750135Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:04.9757318Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:04.9764409Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:04.9771951Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:04.9773909Z (EngineCore_DP0 pid=19144) INFO 12-04 10:06:04 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:06:05.2350270Z (EngineCore_DP0 pid=19144) INFO 12-04 10:06:05 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:06:05.3462819Z (EngineCore_DP0 pid=19144) 2025-12-04T10:06:05.3463254Z processor_config.json: 0% 0.00/178 [00:00, model_arch='PhiForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'microsoft/phi-2'} 2025-12-04T10:06:20.2622358Z 2025-12-04T10:06:20.2623140Z config.json: 0% 0.00/735 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:06:28.8465960Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:28.8493461Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:28.8501037Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:28.8508468Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:28.8515711Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:28.8522957Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:28.8525645Z (EngineCore_DP0 pid=19472) INFO 12-04 10:06:28 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:06:29.1325904Z (EngineCore_DP0 pid=19472) INFO 12-04 10:06:29 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:06:29.1960549Z (EngineCore_DP0 pid=19472) INFO 12-04 10:06:29 [gpu_model_runner.py:2840] Starting to load model microsoft/phi-2... 2025-12-04T10:06:29.4679476Z (EngineCore_DP0 pid=19472) INFO 12-04 10:06:29 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:06:29.5336260Z (EngineCore_DP0 pid=19472) INFO 12-04 10:06:29 [cuda.py:420] Using FlexAttention backend for head_size=80 on V1 engine. 2025-12-04T10:06:30.1927817Z (EngineCore_DP0 pid=19472) INFO 12-04 10:06:30 [gpu_model_runner.py:2902] Model loading took 0.6355 GiB and 0.188021 seconds 2025-12-04T10:06:30.1932694Z (EngineCore_DP0 pid=19472) INFO 12-04 10:06:30 [kv_cache_utils.py:1199] GPU KV cache size: 1,048,576 tokens 2025-12-04T10:06:30.1933620Z (EngineCore_DP0 pid=19472) INFO 12-04 10:06:30 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 512.00x 2025-12-04T10:06:30.6649836Z (EngineCore_DP0 pid=19472) INFO 12-04 10:06:30 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:06:30.6672621Z INFO 12-04 10:06:30 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:06:31.1089946Z PASSED 2025-12-04T10:06:31.1237795Z models/test_initialization.py::test_can_initialize_large_subset[LlamaForCausalLMEagle3] Fork a new process to run a test 19542 2025-12-04T10:06:31.1247592Z Fork a new process to run a test 0 2025-12-04T10:06:31.1526428Z INFO 12-04 10:06:31 [utils.py:239] non-default args: {'tokenizer': 'Qwen/Qwen3-8B', 'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='LlamaForCausalLMEagle3', exist_overrides={}, use_original_num_layers=True), 'speculative_config': {'model': 'AngelSlim/Qwen3-8B_eagle3', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'Qwen/Qwen3-8B'} 2025-12-04T10:06:31.2295757Z 2025-12-04T10:06:31.2297380Z config.json: 0% 0.00/728 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:06:40.3581324Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:40.3608119Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:40.3614556Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:40.3621979Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:40.3629315Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:40.3637023Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:40.3639608Z (EngineCore_DP0 pid=19620) INFO 12-04 10:06:40 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:06:40.6493337Z (EngineCore_DP0 pid=19620) INFO 12-04 10:06:40 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:06:40.6681377Z (EngineCore_DP0 pid=19620) WARNING 12-04 10:06:40 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-12-04T10:06:40.7223875Z (EngineCore_DP0 pid=19620) INFO 12-04 10:06:40 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen3-8B... 2025-12-04T10:06:41.0016271Z (EngineCore_DP0 pid=19620) INFO 12-04 10:06:41 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:06:41.0607102Z (EngineCore_DP0 pid=19620) INFO 12-04 10:06:41 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:06:41.1876014Z (EngineCore_DP0 pid=19620) INFO 12-04 10:06:41 [gpu_model_runner.py:2879] Loading drafter model... 2025-12-04T10:06:41.2550215Z (EngineCore_DP0 pid=19620) INFO 12-04 10:06:41 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-12-04T10:06:41.2551518Z (EngineCore_DP0 pid=19620) INFO 12-04 10:06:41 [eagle.py:1043] The EAGLE head's lm_head will be loaded separately from the target model. 2025-12-04T10:06:41.7981305Z (EngineCore_DP0 pid=19620) INFO 12-04 10:06:41 [gpu_model_runner.py:2902] Model loading took 16.0127 GiB and 0.253670 seconds 2025-12-04T10:06:41.7989353Z (EngineCore_DP0 pid=19620) INFO 12-04 10:06:41 [kv_cache_utils.py:1199] GPU KV cache size: 70,848 tokens 2025-12-04T10:06:41.7990272Z (EngineCore_DP0 pid=19620) INFO 12-04 10:06:41 [kv_cache_utils.py:1204] Maximum concurrency for 40,960 tokens per request: 1.73x 2025-12-04T10:06:42.5904986Z (EngineCore_DP0 pid=19620) INFO 12-04 10:06:42 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:06:42.5919517Z INFO 12-04 10:06:42 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:06:43.0861180Z PASSED 2025-12-04T10:06:43.1013488Z models/test_initialization.py::test_can_initialize_large_subset[NemotronForCausalLM] Fork a new process to run a test 19690 2025-12-04T10:06:43.1022411Z Fork a new process to run a test 0 2025-12-04T10:06:43.1291732Z INFO 12-04 10:06:43 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='NemotronForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'nvidia/Minitron-8B-Base'} 2025-12-04T10:06:43.2252337Z 2025-12-04T10:06:43.2253328Z config.json: 0% 0.00/635 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:06:53.2144801Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:53.2172303Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:53.2180064Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:53.2187983Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:53.2195559Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:53.2203322Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:53.2206167Z (EngineCore_DP0 pid=19769) INFO 12-04 10:06:53 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:06:53.5029437Z (EngineCore_DP0 pid=19769) INFO 12-04 10:06:53 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:06:53.5768446Z (EngineCore_DP0 pid=19769) INFO 12-04 10:06:53 [gpu_model_runner.py:2840] Starting to load model nvidia/Minitron-8B-Base... 2025-12-04T10:06:53.8525885Z (EngineCore_DP0 pid=19769) INFO 12-04 10:06:53 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:06:53.9223942Z (EngineCore_DP0 pid=19769) INFO 12-04 10:06:53 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:06:54.5288855Z (EngineCore_DP0 pid=19769) INFO 12-04 10:06:54 [gpu_model_runner.py:2902] Model loading took 4.2662 GiB and 0.128164 seconds 2025-12-04T10:06:54.5293230Z (EngineCore_DP0 pid=19769) INFO 12-04 10:06:54 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:06:54.5294508Z (EngineCore_DP0 pid=19769) INFO 12-04 10:06:54 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 640.00x 2025-12-04T10:06:55.8557926Z (EngineCore_DP0 pid=19769) INFO 12-04 10:06:55 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:06:55.8572094Z INFO 12-04 10:06:55 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:06:56.3401049Z PASSED 2025-12-04T10:06:56.3550186Z models/test_initialization.py::test_can_initialize_large_subset[CwmForCausalLM] Fork a new process to run a test 19839 2025-12-04T10:06:56.3560230Z Fork a new process to run a test 0 2025-12-04T10:06:56.3562104Z Model is not available online 2025-12-04T10:06:56.6385453Z PASSED 2025-12-04T10:06:56.6534589Z models/test_initialization.py::test_can_initialize_large_subset[RobertaForSequenceClassification] Fork a new process to run a test 19840 2025-12-04T10:06:56.6543936Z Fork a new process to run a test 0 2025-12-04T10:06:56.6817230Z INFO 12-04 10:06:56 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='RobertaForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'cross-encoder/quora-roberta-base'} 2025-12-04T10:06:56.8166129Z 2025-12-04T10:06:56.8167350Z config.json: 0% 0.00/608 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:06:59.3368741Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:59.3399584Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:59.3406390Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:59.3413975Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:59.3421850Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:59.3429553Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:06:59.3432287Z (EngineCore_DP0 pid=19848) INFO 12-04 10:06:59 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:06:59.6237135Z (EngineCore_DP0 pid=19848) INFO 12-04 10:06:59 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:06:59.6379074Z (EngineCore_DP0 pid=19848) INFO 12-04 10:06:59 [gpu_model_runner.py:2840] Starting to load model cross-encoder/quora-roberta-base... 2025-12-04T10:06:59.9020495Z (EngineCore_DP0 pid=19848) INFO 12-04 10:06:59 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:06:59.9076278Z (EngineCore_DP0 pid=19848) INFO 12-04 10:06:59 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:07:00.5084778Z (EngineCore_DP0 pid=19848) INFO 12-04 10:07:00 [gpu_model_runner.py:2902] Model loading took 0.0884 GiB and 0.066830 seconds 2025-12-04T10:07:00.6878089Z (EngineCore_DP0 pid=19848) INFO 12-04 10:07:00 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T10:07:00.7278233Z (EngineCore_DP0 pid=19848) INFO 12-04 10:07:00 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T10:07:01.0041825Z (EngineCore_DP0 pid=19848) INFO 12-04 10:07:01 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:07:01.0057669Z INFO 12-04 10:07:01 [llm.py:337] Supported tasks: ['encode', 'classify', 'score'] 2025-12-04T10:07:01.4317830Z PASSED 2025-12-04T10:07:01.4467625Z models/test_initialization.py::test_can_initialize_large_subset[MambaForCausalLM] Fork a new process to run a test 19918 2025-12-04T10:07:01.4477412Z Fork a new process to run a test 0 2025-12-04T10:07:01.4750583Z INFO 12-04 10:07:01 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MambaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'state-spaces/mamba-130m-hf'} 2025-12-04T10:07:01.5429186Z 2025-12-04T10:07:01.5430606Z config.json: 0% 0.00/895 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:07:10.0459898Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:10.0486967Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:10.0494419Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:10.0501729Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:10.0509040Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:10.0516655Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:10.0519374Z (EngineCore_DP0 pid=19976) INFO 12-04 10:07:10 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:07:10.3305155Z (EngineCore_DP0 pid=19976) INFO 12-04 10:07:10 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:07:10.4996566Z (EngineCore_DP0 pid=19976) INFO 12-04 10:07:10 [gpu_model_runner.py:2840] Starting to load model state-spaces/mamba-130m-hf... 2025-12-04T10:07:10.7701614Z (EngineCore_DP0 pid=19976) INFO 12-04 10:07:10 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:07:11.3679605Z (EngineCore_DP0 pid=19976) INFO 12-04 10:07:11 [gpu_model_runner.py:2902] Model loading took 0.0793 GiB and 0.060775 seconds 2025-12-04T10:07:11.3685066Z (EngineCore_DP0 pid=19976) INFO 12-04 10:07:11 [kv_cache_utils.py:1199] GPU KV cache size: 376,750,080 tokens 2025-12-04T10:07:11.3686026Z (EngineCore_DP0 pid=19976) INFO 12-04 10:07:11 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 183960.00x 2025-12-04T10:07:11.8902176Z (EngineCore_DP0 pid=19976) INFO 12-04 10:07:11 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:07:11.8918465Z INFO 12-04 10:07:11 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:07:12.3175497Z PASSED 2025-12-04T10:07:12.3325435Z models/test_initialization.py::test_can_initialize_large_subset[MiMoForCausalLM] Fork a new process to run a test 20046 2025-12-04T10:07:12.3335081Z Fork a new process to run a test 0 2025-12-04T10:07:12.3606379Z INFO 12-04 10:07:12 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MiMoForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'XiaomiMiMo/MiMo-7B-RL'} 2025-12-04T10:07:12.4709891Z 2025-12-04T10:07:12.4711773Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:07:12.4712081Z config.json: 1.04kB [00:00, 6.78MB/s] 2025-12-04T10:07:12.5405333Z 2025-12-04T10:07:12.5407180Z configuration_mimo.py: 0% 0.00/376 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:07:21.2812413Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:21.2839453Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:21.2846817Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:21.2854425Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:21.2861884Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:21.2869541Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:21.2872994Z (EngineCore_DP0 pid=20104) INFO 12-04 10:07:21 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:07:21.5711874Z (EngineCore_DP0 pid=20104) INFO 12-04 10:07:21 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:07:21.6425542Z (EngineCore_DP0 pid=20104) INFO 12-04 10:07:21 [gpu_model_runner.py:2840] Starting to load model XiaomiMiMo/MiMo-7B-RL... 2025-12-04T10:07:21.9192169Z (EngineCore_DP0 pid=20104) INFO 12-04 10:07:21 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:07:21.9854943Z (EngineCore_DP0 pid=20104) INFO 12-04 10:07:21 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:07:22.5911742Z (EngineCore_DP0 pid=20104) INFO 12-04 10:07:22 [gpu_model_runner.py:2902] Model loading took 2.6543 GiB and 0.126518 seconds 2025-12-04T10:07:22.5915993Z (EngineCore_DP0 pid=20104) INFO 12-04 10:07:22 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:07:22.5916949Z (EngineCore_DP0 pid=20104) INFO 12-04 10:07:22 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 80.00x 2025-12-04T10:07:23.3357616Z (EngineCore_DP0 pid=20104) INFO 12-04 10:07:23 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:07:23.3378483Z INFO 12-04 10:07:23 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:07:23.8068435Z PASSED 2025-12-04T10:07:23.8217114Z models/test_initialization.py::test_can_initialize_large_subset[Glm4vForConditionalGeneration] Fork a new process to run a test 20174 2025-12-04T10:07:23.8227369Z Fork a new process to run a test 0 2025-12-04T10:07:23.8494921Z INFO 12-04 10:07:23 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Glm4vForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'zai-org/GLM-4.1V-9B-Thinking'} 2025-12-04T10:07:23.9408697Z 2025-12-04T10:07:23.9410772Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:07:23.9411204Z config.json: 1.52kB [00:00, 8.10MB/s] 2025-12-04T10:07:24.0309830Z 2025-12-04T10:07:24.0311272Z preprocessor_config.json: 0% 0.00/364 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:07:33.6590836Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:33.6629504Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:33.6637031Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:33.6644603Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:33.6652414Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:33.6660548Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:07:33.6663145Z (EngineCore_DP0 pid=20257) INFO 12-04 10:07:33 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:07:33.9477427Z (EngineCore_DP0 pid=20257) INFO 12-04 10:07:33 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:07:34.1976974Z (EngineCore_DP0 pid=20257) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-12-04T10:07:35.4315029Z (EngineCore_DP0 pid=20257) 2025-12-04T10:07:35.4315702Z video_preprocessor_config.json: 0% 0.00/365 [00:00, model_arch='ErnieMTPModel', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'baidu/ERNIE-4.5-21B-A3B-PT', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'baidu/ERNIE-4.5-21B-A3B-PT'} 2025-12-04T10:07:59.4752128Z INFO 12-04 10:07:59 [model.py:653] Resolved architecture: Ernie4_5_MoeForCausalLM 2025-12-04T10:07:59.4752598Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:07:59.4995379Z INFO 12-04 10:07:59 [model.py:1714] Using max model len 131072 2025-12-04T10:07:59.4996872Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T10:08:06.1791438Z INFO 12-04 10:08:06 [model.py:653] Resolved architecture: ErnieMTPModel 2025-12-04T10:08:06.1793157Z INFO 12-04 10:08:06 [model.py:1714] Using max model len 131072 2025-12-04T10:08:06.1796630Z INFO 12-04 10:08:06 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:08:07.3674632Z (EngineCore_DP0 pid=20400) INFO 12-04 10:08:07 [core.py:727] Waiting for init message from front-end. 2025-12-04T10:08:07.3717629Z (EngineCore_DP0 pid=20400) INFO 12-04 10:08:07 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='baidu/ERNIE-4.5-21B-A3B-PT', speculative_config=SpeculativeConfig(method='mtp', model='baidu/ERNIE-4.5-21B-A3B-PT', num_spec_tokens=1), tokenizer='baidu/ERNIE-4.5-21B-A3B-PT', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=131072, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=baidu/ERNIE-4.5-21B-A3B-PT, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:08:08.7348859Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:08.7377011Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:08.7384257Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:08.7391576Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:08.7399184Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:08.7406928Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:08.7410283Z (EngineCore_DP0 pid=20400) INFO 12-04 10:08:08 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:08:09.0238976Z (EngineCore_DP0 pid=20400) INFO 12-04 10:08:09 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:08:09.0386543Z (EngineCore_DP0 pid=20400) WARNING 12-04 10:08:09 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-12-04T10:08:09.0943066Z (EngineCore_DP0 pid=20400) INFO 12-04 10:08:09 [gpu_model_runner.py:2840] Starting to load model baidu/ERNIE-4.5-21B-A3B-PT... 2025-12-04T10:08:09.3709903Z (EngineCore_DP0 pid=20400) INFO 12-04 10:08:09 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:08:09.4277190Z (EngineCore_DP0 pid=20400) INFO 12-04 10:08:09 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:08:09.4874010Z (EngineCore_DP0 pid=20400) INFO 12-04 10:08:09 [gpu_model_runner.py:2879] Loading drafter model... 2025-12-04T10:08:09.4950703Z (EngineCore_DP0 pid=20400) WARNING 12-04 10:08:09 [vllm.py:821] `torch.compile` is turned on, but the model baidu/ERNIE-4.5-21B-A3B-PT does not support it. Please open an issue on GitHub if you want it to be supported. 2025-12-04T10:08:09.4956549Z (EngineCore_DP0 pid=20400) INFO 12-04 10:08:09 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-12-04T10:08:09.4957499Z (EngineCore_DP0 pid=20400) INFO 12-04 10:08:09 [eagle.py:1027] Loading EAGLE LM head weights from the target model. 2025-12-04T10:08:10.0428969Z (EngineCore_DP0 pid=20400) INFO 12-04 10:08:10 [gpu_model_runner.py:2902] Model loading took 0.9917 GiB and 0.125026 seconds 2025-12-04T10:08:10.0433684Z (EngineCore_DP0 pid=20400) INFO 12-04 10:08:10 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:08:10.0434479Z (EngineCore_DP0 pid=20400) INFO 12-04 10:08:10 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 20.00x 2025-12-04T10:08:11.7984089Z (EngineCore_DP0 pid=20400) INFO 12-04 10:08:11 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:08:11.7997689Z INFO 12-04 10:08:11 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:08:12.2756273Z PASSED 2025-12-04T10:08:12.2906111Z models/test_initialization.py::test_can_initialize_large_subset[GemmaForCausalLM] Fork a new process to run a test 20470 2025-12-04T10:08:12.2916560Z Fork a new process to run a test 0 2025-12-04T10:08:12.3183667Z INFO 12-04 10:08:12 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GemmaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'google/gemma-1.1-2b-it'} 2025-12-04T10:08:12.4769821Z 2025-12-04T10:08:12.4770211Z config.json: 0% 0.00/618 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:08:22.2652674Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:22.2679458Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:22.2686241Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:22.2693399Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:22.2701314Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:22.2708535Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:22.2711412Z (EngineCore_DP0 pid=20550) INFO 12-04 10:08:22 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:08:22.5507687Z (EngineCore_DP0 pid=20550) INFO 12-04 10:08:22 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:08:22.6105636Z (EngineCore_DP0 pid=20550) INFO 12-04 10:08:22 [gpu_model_runner.py:2840] Starting to load model google/gemma-1.1-2b-it... 2025-12-04T10:08:22.8790255Z (EngineCore_DP0 pid=20550) INFO 12-04 10:08:22 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:08:22.9485187Z (EngineCore_DP0 pid=20550) INFO 12-04 10:08:22 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:08:23.5777354Z (EngineCore_DP0 pid=20550) INFO 12-04 10:08:23 [gpu_model_runner.py:2902] Model loading took 1.1856 GiB and 0.149536 seconds 2025-12-04T10:08:23.5781902Z (EngineCore_DP0 pid=20550) INFO 12-04 10:08:23 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T10:08:23.5782825Z (EngineCore_DP0 pid=20550) INFO 12-04 10:08:23 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 1280.00x 2025-12-04T10:08:25.8048910Z (EngineCore_DP0 pid=20550) INFO 12-04 10:08:25 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:08:25.8062660Z INFO 12-04 10:08:25 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:08:26.2893775Z PASSED 2025-12-04T10:08:26.3042535Z models/test_initialization.py::test_can_initialize_large_subset[TarsierForConditionalGeneration] Fork a new process to run a test 20620 2025-12-04T10:08:26.3052598Z Fork a new process to run a test 0 2025-12-04T10:08:26.3327960Z INFO 12-04 10:08:26 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='TarsierForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'omni-research/Tarsier-7b'} 2025-12-04T10:08:26.4424304Z 2025-12-04T10:08:26.4426168Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:08:26.4426479Z config.json: 1.17kB [00:00, 6.87MB/s] 2025-12-04T10:08:26.5627010Z 2025-12-04T10:08:26.5628095Z preprocessor_config.json: 0% 0.00/505 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:08:42.2880195Z (EngineCore_DP0 pid=20700) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:357: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-12-04T10:08:42.2881498Z (EngineCore_DP0 pid=20700) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-12-04T10:08:42.2883288Z (EngineCore_DP0 pid=20700) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-12-04T10:08:42.2884821Z (EngineCore_DP0 pid=20700) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 2025-12-04T10:08:42.2885604Z (EngineCore_DP0 pid=20700) dispatch key: ADInplaceOrView 2025-12-04T10:08:42.2886144Z (EngineCore_DP0 pid=20700) previous kernel: no debug info 2025-12-04T10:08:42.2887458Z (EngineCore_DP0 pid=20700) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-12-04T10:08:42.2888729Z (EngineCore_DP0 pid=20700) self.m.impl( 2025-12-04T10:08:42.6807106Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:42.6834902Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:42.6841810Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:42.6849182Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:42.6855940Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:42.6862739Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:42.6865107Z (EngineCore_DP0 pid=20700) INFO 12-04 10:08:42 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:08:42.9448906Z (EngineCore_DP0 pid=20700) INFO 12-04 10:08:42 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:08:43.0650092Z (EngineCore_DP0 pid=20700) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-12-04T10:08:43.6588174Z (EngineCore_DP0 pid=20700) INFO 12-04 10:08:43 [gpu_model_runner.py:2840] Starting to load model omni-research/Tarsier-7b... 2025-12-04T10:08:43.9151201Z (EngineCore_DP0 pid=20700) INFO 12-04 10:08:43 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:08:43.9659922Z (EngineCore_DP0 pid=20700) INFO 12-04 10:08:43 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:08:44.5061948Z (EngineCore_DP0 pid=20700) INFO 12-04 10:08:44 [gpu_model_runner.py:2902] Model loading took 0.9085 GiB and 0.101697 seconds 2025-12-04T10:08:44.7441969Z (EngineCore_DP0 pid=20700) INFO 12-04 10:08:44 [gpu_model_runner.py:3647] Encoder cache will be initialized with a budget of 8192 tokens, and profiled with 13 image items of the maximum feature size. 2025-12-04T10:08:46.2293014Z (EngineCore_DP0 pid=20700) INFO 12-04 10:08:46 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/083d4441d8/rank_0_0/backbone for vLLM's torch.compile 2025-12-04T10:08:46.2296358Z (EngineCore_DP0 pid=20700) INFO 12-04 10:08:46 [backends.py:608] Dynamo bytecode transform time: 1.22 s 2025-12-04T10:08:46.8706903Z (EngineCore_DP0 pid=20700) [rank0]:W1204 10:08:46.869000 20700 site-packages/torch/_inductor/utils.py:1703] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:08:47.9179367Z (EngineCore_DP0 pid=20700) INFO 12-04 10:08:47 [backends.py:214] Cache the graph for dynamic shape for later use 2025-12-04T10:08:48.8152408Z (EngineCore_DP0 pid=20700) INFO 12-04 10:08:48 [backends.py:241] Compiling a graph for dynamic shape takes 2.38 s 2025-12-04T10:08:49.2479316Z (EngineCore_DP0 pid=20700) INFO 12-04 10:08:49 [monitor.py:33] torch.compile takes 3.59 s in total 2025-12-04T10:08:50.3275672Z (EngineCore_DP0 pid=20700) INFO 12-04 10:08:50 [gpu_worker.py:314] Available KV cache memory: 15.12 GiB 2025-12-04T10:08:50.6078840Z (EngineCore_DP0 pid=20700) INFO 12-04 10:08:50 [kv_cache_utils.py:1199] GPU KV cache size: 990,624 tokens 2025-12-04T10:08:50.6079748Z (EngineCore_DP0 pid=20700) INFO 12-04 10:08:50 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 241.85x 2025-12-04T10:08:50.8934224Z (EngineCore_DP0 pid=20700) 2025-12-04T10:08:50.9942959Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/67 [00:00, model_arch='Qwen3ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen3-8B'} 2025-12-04T10:08:53.1014245Z INFO 12-04 10:08:53 [model.py:653] Resolved architecture: Qwen3ForCausalLM 2025-12-04T10:08:53.1014737Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:08:53.1258820Z INFO 12-04 10:08:53 [model.py:1714] Using max model len 40960 2025-12-04T10:08:53.2935206Z INFO 12-04 10:08:53 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:08:53.8034941Z (EngineCore_DP0 pid=20972) INFO 12-04 10:08:53 [core.py:727] Waiting for init message from front-end. 2025-12-04T10:08:53.8069670Z (EngineCore_DP0 pid=20972) INFO 12-04 10:08:53 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='Qwen/Qwen3-8B', speculative_config=None, tokenizer='Qwen/Qwen3-8B', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.bfloat16, max_seq_len=40960, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=Qwen/Qwen3-8B, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:08:55.1176717Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:55.1205086Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:55.1212282Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:55.1220026Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:55.1227365Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:55.1234798Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:08:55.1237784Z (EngineCore_DP0 pid=20972) INFO 12-04 10:08:55 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:08:55.4040120Z (EngineCore_DP0 pid=20972) INFO 12-04 10:08:55 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:08:55.4832751Z (EngineCore_DP0 pid=20972) INFO 12-04 10:08:55 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen3-8B... 2025-12-04T10:08:55.7576897Z (EngineCore_DP0 pid=20972) INFO 12-04 10:08:55 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:08:55.8245515Z (EngineCore_DP0 pid=20972) INFO 12-04 10:08:55 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:08:56.4234558Z (EngineCore_DP0 pid=20972) INFO 12-04 10:08:56 [gpu_model_runner.py:2902] Model loading took 2.6895 GiB and 0.126606 seconds 2025-12-04T10:08:56.4239110Z (EngineCore_DP0 pid=20972) INFO 12-04 10:08:56 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:08:56.4240003Z (EngineCore_DP0 pid=20972) INFO 12-04 10:08:56 [kv_cache_utils.py:1204] Maximum concurrency for 40,960 tokens per request: 64.00x 2025-12-04T10:08:57.2374587Z (EngineCore_DP0 pid=20972) INFO 12-04 10:08:57 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:08:57.2394686Z INFO 12-04 10:08:57 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:08:57.7177915Z PASSED 2025-12-04T10:08:57.7326969Z models/test_initialization.py::test_can_initialize_large_subset[BailingMoeV2ForCausalLM] Fork a new process to run a test 21042 2025-12-04T10:08:57.7338077Z Fork a new process to run a test 0 2025-12-04T10:08:57.7606948Z INFO 12-04 10:08:57 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='BailingMoeV2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'inclusionAI/Ling-mini-2.0'} 2025-12-04T10:08:57.8600708Z 2025-12-04T10:08:57.8602649Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:08:57.8602948Z config.json: 1.52kB [00:00, 8.21MB/s] 2025-12-04T10:08:57.9264575Z 2025-12-04T10:08:57.9266307Z configuration_bailing_moe_v2.py: 0.00B [00:00, ?B/s] 2025-12-04T10:08:57.9266747Z configuration_bailing_moe_v2.py: 3.16kB [00:00, 27.9MB/s] 2025-12-04T10:08:57.9359792Z A new version of the following files was downloaded from https://huggingface.co/inclusionAI/Ling-mini-2.0: 2025-12-04T10:08:57.9360361Z - configuration_bailing_moe_v2.py 2025-12-04T10:08:57.9360997Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:08:57.9700346Z You are using a model of type bailing_moe to instantiate a model of type . This is not supported for all configurations of models and can yield errors. 2025-12-04T10:09:04.3990606Z INFO 12-04 10:09:04 [model.py:653] Resolved architecture: BailingMoeV2ForCausalLM 2025-12-04T10:09:04.3991123Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:09:04.4248052Z INFO 12-04 10:09:04 [model.py:1714] Using max model len 32768 2025-12-04T10:09:04.4250042Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T10:09:04.4808806Z INFO 12-04 10:09:04 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:09:04.5241908Z 2025-12-04T10:09:04.5243529Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:09:04.5243866Z tokenizer_config.json: 4.59kB [00:00, 33.3MB/s] 2025-12-04T10:09:04.5840529Z 2025-12-04T10:09:04.6172049Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T10:09:04.6172357Z tokenizer.json: 7.66MB [00:00, 232MB/s] 2025-12-04T10:09:04.7290008Z 2025-12-04T10:09:04.7290762Z special_tokens_map.json: 0% 0.00/152 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:09:06.6330283Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:06.6357571Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:06.6364822Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:06.6372746Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:06.6380799Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:06.6387954Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:06.6390936Z (EngineCore_DP0 pid=21100) INFO 12-04 10:09:06 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:09:06.9195892Z (EngineCore_DP0 pid=21100) INFO 12-04 10:09:06 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:09:06.9797379Z (EngineCore_DP0 pid=21100) INFO 12-04 10:09:06 [gpu_model_runner.py:2840] Starting to load model inclusionAI/Ling-mini-2.0... 2025-12-04T10:09:07.2526223Z (EngineCore_DP0 pid=21100) INFO 12-04 10:09:07 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:09:07.3181572Z (EngineCore_DP0 pid=21100) INFO 12-04 10:09:07 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:09:07.9204548Z (EngineCore_DP0 pid=21100) INFO 12-04 10:09:07 [gpu_model_runner.py:2902] Model loading took 1.3224 GiB and 0.125655 seconds 2025-12-04T10:09:07.9208461Z (EngineCore_DP0 pid=21100) INFO 12-04 10:09:07 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:09:07.9209382Z (EngineCore_DP0 pid=21100) INFO 12-04 10:09:07 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 160.00x 2025-12-04T10:09:09.1515942Z (EngineCore_DP0 pid=21100) INFO 12-04 10:09:09 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:09:09.1529311Z INFO 12-04 10:09:09 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:09:09.6151135Z PASSED 2025-12-04T10:09:09.6300004Z models/test_initialization.py::test_can_initialize_large_subset[SolarForCausalLM] Fork a new process to run a test 21170 2025-12-04T10:09:09.6310227Z Fork a new process to run a test 0 2025-12-04T10:09:09.6581258Z INFO 12-04 10:09:09 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='SolarForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'upstage/solar-pro-preview-instruct'} 2025-12-04T10:09:09.7722217Z 2025-12-04T10:09:09.7725041Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:09:09.7725346Z config.json: 1.03kB [00:00, 3.23MB/s] 2025-12-04T10:09:09.8443987Z 2025-12-04T10:09:09.8446096Z configuration_solar.py: 0.00B [00:00, ?B/s] 2025-12-04T10:09:09.8446464Z configuration_solar.py: 10.1kB [00:00, 39.3MB/s] 2025-12-04T10:09:09.8544305Z A new version of the following files was downloaded from https://huggingface.co/upstage/solar-pro-preview-instruct: 2025-12-04T10:09:09.8544896Z - configuration_solar.py 2025-12-04T10:09:09.8545503Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:09:16.3941424Z INFO 12-04 10:09:16 [model.py:653] Resolved architecture: SolarForCausalLM 2025-12-04T10:09:16.3941910Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:09:16.4199645Z INFO 12-04 10:09:16 [model.py:1714] Using max model len 4096 2025-12-04T10:09:16.4201860Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T10:09:16.4805396Z INFO 12-04 10:09:16 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:09:16.5208523Z 2025-12-04T10:09:16.5210875Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:09:16.5211224Z tokenizer_config.json: 24.6kB [00:00, 130MB/s] 2025-12-04T10:09:16.6573509Z 2025-12-04T10:09:16.9022172Z tokenizer.model: 0% 0.00/500k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:09:18.5332705Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:18.5360282Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:18.5367913Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:18.5375424Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:18.5383362Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:18.5391052Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:18.5393927Z (EngineCore_DP0 pid=21248) INFO 12-04 10:09:18 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:09:18.8295800Z (EngineCore_DP0 pid=21248) INFO 12-04 10:09:18 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:09:18.9270222Z (EngineCore_DP0 pid=21248) INFO 12-04 10:09:18 [gpu_model_runner.py:2840] Starting to load model upstage/solar-pro-preview-instruct... 2025-12-04T10:09:19.2113234Z (EngineCore_DP0 pid=21248) INFO 12-04 10:09:19 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:09:19.2771919Z (EngineCore_DP0 pid=21248) INFO 12-04 10:09:19 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:09:19.8899998Z (EngineCore_DP0 pid=21248) INFO 12-04 10:09:19 [gpu_model_runner.py:2902] Model loading took 1.2510 GiB and 0.123764 seconds 2025-12-04T10:09:19.8904036Z (EngineCore_DP0 pid=21248) INFO 12-04 10:09:19 [kv_cache_utils.py:1199] GPU KV cache size: 2,097,152 tokens 2025-12-04T10:09:19.8904805Z (EngineCore_DP0 pid=21248) INFO 12-04 10:09:19 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 510.01x 2025-12-04T10:09:20.3936604Z (EngineCore_DP0 pid=21248) INFO 12-04 10:09:20 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:09:20.3958747Z INFO 12-04 10:09:20 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:09:20.8613634Z PASSED 2025-12-04T10:09:20.8763778Z models/test_initialization.py::test_can_initialize_large_subset[GPTNeoXForCausalLM] Fork a new process to run a test 21318 2025-12-04T10:09:20.8772780Z Fork a new process to run a test 0 2025-12-04T10:09:20.9043156Z INFO 12-04 10:09:20 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GPTNeoXForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'EleutherAI/pythia-70m'} 2025-12-04T10:09:20.9842722Z 2025-12-04T10:09:20.9843919Z config.json: 0% 0.00/567 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:09:29.4962528Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:29.4989538Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:29.4996952Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:29.5004175Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:29.5011859Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:29.5019413Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:29.5022185Z (EngineCore_DP0 pid=21376) INFO 12-04 10:09:29 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:09:29.7851879Z (EngineCore_DP0 pid=21376) INFO 12-04 10:09:29 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:09:29.8205336Z (EngineCore_DP0 pid=21376) INFO 12-04 10:09:29 [gpu_model_runner.py:2840] Starting to load model EleutherAI/pythia-70m... 2025-12-04T10:09:30.0980165Z (EngineCore_DP0 pid=21376) INFO 12-04 10:09:30 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:09:30.1633634Z (EngineCore_DP0 pid=21376) INFO 12-04 10:09:30 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:09:30.7707806Z (EngineCore_DP0 pid=21376) INFO 12-04 10:09:30 [gpu_model_runner.py:2902] Model loading took 0.1036 GiB and 0.123607 seconds 2025-12-04T10:09:30.7711899Z (EngineCore_DP0 pid=21376) INFO 12-04 10:09:30 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:09:30.7712790Z (EngineCore_DP0 pid=21376) INFO 12-04 10:09:30 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 2560.00x 2025-12-04T10:09:31.7086607Z (EngineCore_DP0 pid=21376) INFO 12-04 10:09:31 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:09:31.7101411Z INFO 12-04 10:09:31 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:09:32.1502993Z PASSED 2025-12-04T10:09:32.1652038Z models/test_initialization.py::test_can_initialize_large_subset[EagleLlamaForCausalLM] Fork a new process to run a test 21446 2025-12-04T10:09:32.1661439Z Fork a new process to run a test 0 2025-12-04T10:09:32.1937434Z INFO 12-04 10:09:32 [utils.py:239] non-default args: {'tokenizer': 'meta-llama/Meta-Llama-3-8B-Instruct', 'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='EagleLlamaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'yuhuili/EAGLE-LLaMA3-Instruct-8B', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'meta-llama/Meta-Llama-3-8B-Instruct'} 2025-12-04T10:09:32.3024121Z 2025-12-04T10:09:32.3025300Z config.json: 0% 0.00/654 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:09:35.4431846Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:35.4458353Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:35.4465743Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:35.4473853Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:35.4481182Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:35.4488854Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:35.4491791Z (EngineCore_DP0 pid=21454) INFO 12-04 10:09:35 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:09:35.7298469Z (EngineCore_DP0 pid=21454) INFO 12-04 10:09:35 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:09:35.7450688Z (EngineCore_DP0 pid=21454) WARNING 12-04 10:09:35 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-12-04T10:09:35.7827303Z (EngineCore_DP0 pid=21454) INFO 12-04 10:09:35 [gpu_model_runner.py:2840] Starting to load model meta-llama/Meta-Llama-3-8B-Instruct... 2025-12-04T10:09:36.0586112Z (EngineCore_DP0 pid=21454) INFO 12-04 10:09:36 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:09:36.1139729Z (EngineCore_DP0 pid=21454) INFO 12-04 10:09:36 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:09:36.1730969Z (EngineCore_DP0 pid=21454) INFO 12-04 10:09:36 [gpu_model_runner.py:2879] Loading drafter model... 2025-12-04T10:09:36.1852795Z (EngineCore_DP0 pid=21454) INFO 12-04 10:09:36 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-12-04T10:09:36.1854089Z (EngineCore_DP0 pid=21454) INFO 12-04 10:09:36 [eagle.py:1027] Loading EAGLE LM head weights from the target model. 2025-12-04T10:09:36.7264031Z (EngineCore_DP0 pid=21454) INFO 12-04 10:09:36 [gpu_model_runner.py:2902] Model loading took 2.8345 GiB and 0.126961 seconds 2025-12-04T10:09:36.7268512Z (EngineCore_DP0 pid=21454) INFO 12-04 10:09:36 [kv_cache_utils.py:1199] GPU KV cache size: 1,310,720 tokens 2025-12-04T10:09:36.7269267Z (EngineCore_DP0 pid=21454) INFO 12-04 10:09:36 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 160.00x 2025-12-04T10:09:37.4535231Z (EngineCore_DP0 pid=21454) INFO 12-04 10:09:37 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:09:37.4558282Z INFO 12-04 10:09:37 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:09:37.9244426Z PASSED 2025-12-04T10:09:37.9394309Z models/test_initialization.py::test_can_initialize_large_subset[RForConditionalGeneration] Fork a new process to run a test 21524 2025-12-04T10:09:37.9404233Z Fork a new process to run a test 0 2025-12-04T10:09:37.9678500Z INFO 12-04 10:09:37 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='RForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'YannQi/R-4B'} 2025-12-04T10:09:38.0537782Z 2025-12-04T10:09:38.0539982Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:09:38.0540304Z config.json: 2.19kB [00:00, 13.5MB/s] 2025-12-04T10:09:38.1342353Z 2025-12-04T10:09:38.1343424Z configuration_r.py: 0.00B [00:00, ?B/s] 2025-12-04T10:09:38.1343784Z configuration_r.py: 3.59kB [00:00, 36.8MB/s] 2025-12-04T10:09:38.1441005Z A new version of the following files was downloaded from https://huggingface.co/YannQi/R-4B: 2025-12-04T10:09:38.1441482Z - configuration_r.py 2025-12-04T10:09:38.1442096Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:09:38.3727932Z 2025-12-04T10:09:38.3729017Z preprocessor_config.json: 0% 0.00/745 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:09:55.1587183Z (EngineCore_DP0 pid=21605) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:357: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-12-04T10:09:55.1588462Z (EngineCore_DP0 pid=21605) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-12-04T10:09:55.1590212Z (EngineCore_DP0 pid=21605) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-12-04T10:09:55.1592095Z (EngineCore_DP0 pid=21605) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 2025-12-04T10:09:55.1592892Z (EngineCore_DP0 pid=21605) dispatch key: ADInplaceOrView 2025-12-04T10:09:55.1593393Z (EngineCore_DP0 pid=21605) previous kernel: no debug info 2025-12-04T10:09:55.1594451Z (EngineCore_DP0 pid=21605) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-12-04T10:09:55.1595346Z (EngineCore_DP0 pid=21605) self.m.impl( 2025-12-04T10:09:55.8810714Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:55.8840365Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:55.8847221Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:55.8854439Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:55.8861859Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:55.8868790Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:09:55.8871838Z (EngineCore_DP0 pid=21605) INFO 12-04 10:09:55 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:09:56.1531825Z (EngineCore_DP0 pid=21605) INFO 12-04 10:09:56 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:09:56.2885074Z (EngineCore_DP0 pid=21605) 2025-12-04T10:09:56.2885536Z processor_config.json: 0% 0.00/330 [00:00, model_arch='LongcatFlashForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'meituan-longcat/LongCat-Flash-Chat'} 2025-12-04T10:10:11.5297483Z INFO 12-04 10:10:11 [model.py:653] Resolved architecture: LongcatFlashForCausalLM 2025-12-04T10:10:11.5298379Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:10:11.7318156Z 2025-12-04T10:10:11.9845726Z Parse safetensors files: 0% 0/75 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:10:14.0211207Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:10:14.0240431Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:10:14.0248458Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:10:14.0257264Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:10:14.0265771Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:10:14.0274578Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:10:14.0277649Z (EngineCore_DP0 pid=22006) INFO 12-04 10:10:14 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:10:14.3167081Z (EngineCore_DP0 pid=22006) INFO 12-04 10:10:14 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:10:14.4289780Z (EngineCore_DP0 pid=22006) INFO 12-04 10:10:14 [gpu_model_runner.py:2840] Starting to load model meituan-longcat/LongCat-Flash-Chat... 2025-12-04T10:10:14.7165668Z (EngineCore_DP0 pid=22006) INFO 12-04 10:10:14 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:10:14.7835576Z (EngineCore_DP0 pid=22006) INFO 12-04 10:10:14 [cuda.py:328] Using Triton MLA backend on V1 engine. 2025-12-04T10:10:15.4139781Z (EngineCore_DP0 pid=22006) INFO 12-04 10:10:15 [gpu_model_runner.py:2902] Model loading took 4.3434 GiB and 0.143760 seconds 2025-12-04T10:10:15.4144434Z (EngineCore_DP0 pid=22006) INFO 12-04 10:10:15 [kv_cache_utils.py:1199] GPU KV cache size: 4,660,336 tokens 2025-12-04T10:10:15.4145431Z (EngineCore_DP0 pid=22006) INFO 12-04 10:10:15 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 35.56x 2025-12-04T10:10:16.5844683Z (EngineCore_DP0 pid=22006) INFO 12-04 10:10:16 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:10:16.5860809Z INFO 12-04 10:10:16 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:10:17.0897013Z PASSED 2025-12-04T10:10:17.1046002Z models/test_initialization.py::test_can_initialize_large_subset[MantisForConditionalGeneration] Fork a new process to run a test 22076 2025-12-04T10:10:17.1056208Z Fork a new process to run a test 0 2025-12-04T10:10:17.1060290Z `transformers==4.56.2` installed, but `transformers<=4.48` is required to run this model. Reason: HF model is not compatible. 2025-12-04T10:10:17.3913095Z PASSED 2025-12-04T10:10:17.4063240Z models/test_initialization.py::test_can_initialize_large_subset[PersimmonForCausalLM] Fork a new process to run a test 22077 2025-12-04T10:10:17.4073401Z Fork a new process to run a test 0 2025-12-04T10:10:17.4350602Z INFO 12-04 10:10:17 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='PersimmonForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'adept/persimmon-8b-chat'} 2025-12-04T10:10:17.5877813Z 2025-12-04T10:10:17.5879242Z config.json: 0% 0.00/635 [00:00. This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message 2025-12-04T10:10:26.1039958Z You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message. 2025-12-04T10:10:47.7673255Z 2025-12-04T10:10:47.7673855Z generation_config.json: 0% 0.00/124 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:10:48.8039372Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:10:48.8066610Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:10:48.8073692Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:10:48.8081376Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:10:48.8088895Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:10:48.8096346Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:10:48.8099094Z (EngineCore_DP0 pid=22155) INFO 12-04 10:10:48 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:10:49.1663142Z (EngineCore_DP0 pid=22155) INFO 12-04 10:10:49 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:10:49.2474252Z (EngineCore_DP0 pid=22155) INFO 12-04 10:10:49 [gpu_model_runner.py:2840] Starting to load model adept/persimmon-8b-chat... 2025-12-04T10:10:49.5266903Z (EngineCore_DP0 pid=22155) INFO 12-04 10:10:49 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:10:49.5955401Z (EngineCore_DP0 pid=22155) INFO 12-04 10:10:49 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:10:50.1996075Z (EngineCore_DP0 pid=22155) INFO 12-04 10:10:50 [gpu_model_runner.py:2902] Model loading took 4.3761 GiB and 0.128841 seconds 2025-12-04T10:10:50.2000625Z (EngineCore_DP0 pid=22155) INFO 12-04 10:10:50 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-12-04T10:10:50.2001544Z (EngineCore_DP0 pid=22155) INFO 12-04 10:10:50 [kv_cache_utils.py:1204] Maximum concurrency for 16,384 tokens per request: 40.00x 2025-12-04T10:11:13.6679462Z (EngineCore_DP0 pid=22155) INFO 12-04 10:11:13 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:11:13.6693574Z INFO 12-04 10:11:13 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:11:14.2291723Z PASSED 2025-12-04T10:11:14.2443468Z models/test_initialization.py::test_can_initialize_large_subset[Step3TextForCausalLM] Fork a new process to run a test 22225 2025-12-04T10:11:14.2453579Z Fork a new process to run a test 0 2025-12-04T10:11:14.2732403Z INFO 12-04 10:11:14 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Step3TextForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'stepfun-ai/step3'} 2025-12-04T10:11:14.5826700Z INFO 12-04 10:11:14 [model.py:653] Resolved architecture: Step3VLForConditionalGeneration 2025-12-04T10:11:14.5827569Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:11:14.6070874Z INFO 12-04 10:11:14 [model.py:1714] Using max model len 65536 2025-12-04T10:11:14.6073485Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T10:11:14.7774412Z INFO 12-04 10:11:14 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:11:15.7237181Z (EngineCore_DP0 pid=22232) INFO 12-04 10:11:15 [core.py:727] Waiting for init message from front-end. 2025-12-04T10:11:15.7274118Z (EngineCore_DP0 pid=22232) INFO 12-04 10:11:15 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='stepfun-ai/step3', speculative_config=None, tokenizer='stepfun-ai/step3', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=65536, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=stepfun-ai/step3, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:11:17.0657814Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:17.0685667Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:17.0693312Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:17.0701034Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:17.0708745Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:17.0716648Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:17.0719325Z (EngineCore_DP0 pid=22232) INFO 12-04 10:11:17 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:11:17.3521397Z (EngineCore_DP0 pid=22232) INFO 12-04 10:11:17 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:11:17.4597532Z (EngineCore_DP0 pid=22232) INFO 12-04 10:11:17 [gpu_model_runner.py:2840] Starting to load model stepfun-ai/step3... 2025-12-04T10:11:17.7356747Z (EngineCore_DP0 pid=22232) INFO 12-04 10:11:17 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:11:17.7532534Z (EngineCore_DP0 pid=22232) INFO 12-04 10:11:17 [layer.py:497] MultiHeadAttention attn_backend: _Backend.XFORMERS, use_upstream_fa: False 2025-12-04T10:11:17.8207089Z (EngineCore_DP0 pid=22232) INFO 12-04 10:11:17 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:11:18.4205898Z (EngineCore_DP0 pid=22232) INFO 12-04 10:11:18 [gpu_model_runner.py:2902] Model loading took 5.3915 GiB and 0.145162 seconds 2025-12-04T10:11:18.4209997Z (EngineCore_DP0 pid=22232) INFO 12-04 10:11:18 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T10:11:18.4211343Z (EngineCore_DP0 pid=22232) INFO 12-04 10:11:18 [kv_cache_utils.py:1204] Maximum concurrency for 65,536 tokens per request: 160.00x 2025-12-04T10:11:20.0936059Z (EngineCore_DP0 pid=22232) INFO 12-04 10:11:20 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:11:20.0950151Z INFO 12-04 10:11:20 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:11:20.5952011Z PASSED 2025-12-04T10:11:20.6101859Z models/test_initialization.py::test_can_initialize_large_subset[InternLM2ForCausalLM] Fork a new process to run a test 22302 2025-12-04T10:11:20.6111784Z Fork a new process to run a test 0 2025-12-04T10:11:20.6387233Z INFO 12-04 10:11:20 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='InternLM2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'internlm/internlm2-chat-7b'} 2025-12-04T10:11:20.7472698Z 2025-12-04T10:11:20.7474381Z config.json: 0% 0.00/916 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:11:30.2235025Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:30.2262237Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:30.2269271Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:30.2277559Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:30.2284703Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:30.2292215Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:30.2295054Z (EngineCore_DP0 pid=22381) INFO 12-04 10:11:30 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:11:30.5159050Z (EngineCore_DP0 pid=22381) INFO 12-04 10:11:30 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:11:30.5924349Z (EngineCore_DP0 pid=22381) INFO 12-04 10:11:30 [gpu_model_runner.py:2840] Starting to load model internlm/internlm2-chat-7b... 2025-12-04T10:11:30.8740362Z (EngineCore_DP0 pid=22381) INFO 12-04 10:11:30 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:11:30.9397141Z (EngineCore_DP0 pid=22381) INFO 12-04 10:11:30 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:11:31.5570541Z (EngineCore_DP0 pid=22381) INFO 12-04 10:11:31 [gpu_model_runner.py:2902] Model loading took 1.8360 GiB and 0.124580 seconds 2025-12-04T10:11:31.5575435Z (EngineCore_DP0 pid=22381) INFO 12-04 10:11:31 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:11:31.5576363Z (EngineCore_DP0 pid=22381) INFO 12-04 10:11:31 [kv_cache_utils.py:1204] Maximum concurrency for 65,536 tokens per request: 40.00x 2025-12-04T10:11:33.0217844Z (EngineCore_DP0 pid=22381) INFO 12-04 10:11:33 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:11:33.0231819Z INFO 12-04 10:11:33 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:11:33.4918507Z PASSED 2025-12-04T10:11:33.5067377Z models/test_initialization.py::test_can_initialize_large_subset[LlamaModel] Fork a new process to run a test 22451 2025-12-04T10:11:33.5077987Z Fork a new process to run a test 0 2025-12-04T10:11:33.5079870Z Model is not available online 2025-12-04T10:11:33.7902290Z PASSED 2025-12-04T10:11:33.8051274Z models/test_initialization.py::test_can_initialize_large_subset[Qwen3MoeForCausalLM] Fork a new process to run a test 22452 2025-12-04T10:11:33.8061432Z Fork a new process to run a test 0 2025-12-04T10:11:33.8338441Z INFO 12-04 10:11:33 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen3MoeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen3-30B-A3B'} 2025-12-04T10:11:33.9136170Z 2025-12-04T10:11:33.9136552Z config.json: 0% 0.00/963 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:11:42.8917385Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:42.8945514Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:42.8952845Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:42.8960440Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:42.8968295Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:42.8975896Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:42.8978857Z (EngineCore_DP0 pid=22531) INFO 12-04 10:11:42 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:11:43.1799586Z (EngineCore_DP0 pid=22531) INFO 12-04 10:11:43 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:11:43.2415800Z (EngineCore_DP0 pid=22531) INFO 12-04 10:11:43 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen3-30B-A3B... 2025-12-04T10:11:43.5130437Z (EngineCore_DP0 pid=22531) INFO 12-04 10:11:43 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:11:43.5806313Z (EngineCore_DP0 pid=22531) INFO 12-04 10:11:43 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:11:44.1818981Z (EngineCore_DP0 pid=22531) INFO 12-04 10:11:44 [gpu_model_runner.py:2902] Model loading took 1.2227 GiB and 0.127317 seconds 2025-12-04T10:11:44.1823551Z (EngineCore_DP0 pid=22531) INFO 12-04 10:11:44 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:11:44.1824375Z (EngineCore_DP0 pid=22531) INFO 12-04 10:11:44 [kv_cache_utils.py:1204] Maximum concurrency for 40,960 tokens per request: 128.00x 2025-12-04T10:11:45.4264770Z (EngineCore_DP0 pid=22531) INFO 12-04 10:11:45 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:11:45.4279185Z INFO 12-04 10:11:45 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:11:45.8988192Z PASSED 2025-12-04T10:11:45.9137829Z models/test_initialization.py::test_can_initialize_large_subset[ArceeForCausalLM] Fork a new process to run a test 22601 2025-12-04T10:11:45.9148677Z Fork a new process to run a test 0 2025-12-04T10:11:45.9425634Z INFO 12-04 10:11:45 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='ArceeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'arcee-ai/AFM-4.5B-Base'} 2025-12-04T10:11:46.0794704Z 2025-12-04T10:11:46.0795776Z config.json: 0% 0.00/843 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:11:55.4098806Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:55.4126470Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:55.4133431Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:55.4141296Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:55.4148400Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:55.4156161Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:11:55.4158940Z (EngineCore_DP0 pid=22679) INFO 12-04 10:11:55 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:11:55.6988694Z (EngineCore_DP0 pid=22679) INFO 12-04 10:11:55 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:11:55.7851066Z (EngineCore_DP0 pid=22679) INFO 12-04 10:11:55 [gpu_model_runner.py:2840] Starting to load model arcee-ai/AFM-4.5B-Base... 2025-12-04T10:11:56.0585263Z (EngineCore_DP0 pid=22679) INFO 12-04 10:11:56 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:11:56.1625039Z (EngineCore_DP0 pid=22679) INFO 12-04 10:11:56 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:11:56.7606644Z (EngineCore_DP0 pid=22679) INFO 12-04 10:11:56 [gpu_model_runner.py:2902] Model loading took 1.4478 GiB and 0.162373 seconds 2025-12-04T10:11:56.7611624Z (EngineCore_DP0 pid=22679) INFO 12-04 10:11:56 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:11:56.7612520Z (EngineCore_DP0 pid=22679) INFO 12-04 10:11:56 [kv_cache_utils.py:1204] Maximum concurrency for 81,920 tokens per request: 64.00x 2025-12-04T10:11:58.0804637Z (EngineCore_DP0 pid=22679) INFO 12-04 10:11:58 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:11:58.0818363Z INFO 12-04 10:11:58 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:11:58.5830535Z PASSED 2025-12-04T10:11:58.5979429Z models/test_initialization.py::test_can_initialize_large_subset[NVLM_D] Fork a new process to run a test 22749 2025-12-04T10:11:58.5990896Z Fork a new process to run a test 0 2025-12-04T10:11:58.6267146Z INFO 12-04 10:11:58 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='NVLM_D', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'nvidia/NVLM-D-72B'} 2025-12-04T10:11:58.7974693Z 2025-12-04T10:11:58.7977268Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:11:58.7977581Z config.json: 3.71kB [00:00, 16.1MB/s] 2025-12-04T10:11:58.8648642Z 2025-12-04T10:11:58.8650490Z configuration_nvlm_d.py: 0.00B [00:00, ?B/s] 2025-12-04T10:11:58.8650851Z configuration_nvlm_d.py: 3.80kB [00:00, 27.6MB/s] 2025-12-04T10:11:58.9248477Z 2025-12-04T10:11:58.9250632Z configuration_intern_vit.py: 0.00B [00:00, ?B/s] 2025-12-04T10:11:58.9251233Z configuration_intern_vit.py: 5.55kB [00:00, 35.8MB/s] 2025-12-04T10:11:58.9273422Z A new version of the following files was downloaded from https://huggingface.co/nvidia/NVLM-D-72B: 2025-12-04T10:11:58.9274211Z - configuration_intern_vit.py 2025-12-04T10:11:58.9275125Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:11:58.9276365Z A new version of the following files was downloaded from https://huggingface.co/nvidia/NVLM-D-72B: 2025-12-04T10:11:58.9277058Z - configuration_nvlm_d.py 2025-12-04T10:11:58.9277618Z - configuration_intern_vit.py 2025-12-04T10:11:58.9278510Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:11:59.1567492Z 2025-12-04T10:11:59.1568029Z preprocessor_config.json: 0% 0.00/287 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:12:08.1041862Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:08.1071454Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:08.1078683Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:08.1086111Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:08.1093713Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:08.1101383Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:08.1104508Z (EngineCore_DP0 pid=22807) INFO 12-04 10:12:08 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:12:08.3915376Z (EngineCore_DP0 pid=22807) INFO 12-04 10:12:08 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:12:08.5104358Z (EngineCore_DP0 pid=22807) WARNING 12-04 10:12:08 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T10:12:08.5122579Z (EngineCore_DP0 pid=22807) WARNING 12-04 10:12:08 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T10:12:08.5494585Z (EngineCore_DP0 pid=22807) INFO 12-04 10:12:08 [gpu_model_runner.py:2840] Starting to load model nvidia/NVLM-D-72B... 2025-12-04T10:12:08.8295114Z (EngineCore_DP0 pid=22807) INFO 12-04 10:12:08 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:12:08.8343086Z (EngineCore_DP0 pid=22807) INFO 12-04 10:12:08 [layer.py:497] MultiHeadAttention attn_backend: _Backend.FLASH_ATTN, use_upstream_fa: False 2025-12-04T10:12:08.8931396Z (EngineCore_DP0 pid=22807) INFO 12-04 10:12:08 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:12:09.4963647Z (EngineCore_DP0 pid=22807) INFO 12-04 10:12:09 [gpu_model_runner.py:2902] Model loading took 7.7171 GiB and 0.125085 seconds 2025-12-04T10:12:09.4967843Z (EngineCore_DP0 pid=22807) INFO 12-04 10:12:09 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:12:09.4968811Z (EngineCore_DP0 pid=22807) INFO 12-04 10:12:09 [kv_cache_utils.py:1204] Maximum concurrency for 98,304 tokens per request: 26.67x 2025-12-04T10:12:09.8683365Z (EngineCore_DP0 pid=22807) WARNING 12-04 10:12:09 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T10:12:10.2738829Z (EngineCore_DP0 pid=22807) INFO 12-04 10:12:10 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:12:10.2755712Z INFO 12-04 10:12:10 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:12:10.7768268Z PASSED 2025-12-04T10:12:10.7922524Z models/test_initialization.py::test_can_initialize_large_subset[Ovis2_5] Fork a new process to run a test 22893 2025-12-04T10:12:10.7932832Z Fork a new process to run a test 0 2025-12-04T10:12:10.8208593Z INFO 12-04 10:12:10 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Ovis2_5', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'AIDC-AI/Ovis2.5-2B'} 2025-12-04T10:12:10.9264347Z 2025-12-04T10:12:10.9266304Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:12:10.9266612Z config.json: 2.00kB [00:00, 10.6MB/s] 2025-12-04T10:12:10.9907129Z 2025-12-04T10:12:10.9908822Z configuration_ovis2_5.py: 0.00B [00:00, ?B/s] 2025-12-04T10:12:10.9909221Z configuration_ovis2_5.py: 3.94kB [00:00, 22.0MB/s] 2025-12-04T10:12:11.0013406Z A new version of the following files was downloaded from https://huggingface.co/AIDC-AI/Ovis2.5-2B: 2025-12-04T10:12:11.0014130Z - configuration_ovis2_5.py 2025-12-04T10:12:11.2087288Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:12:11.2087970Z 2025-12-04T10:12:11.2088137Z preprocessor_config.json: 0% 0.00/394 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:12:20.3017254Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:20.3045008Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:20.3051922Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:20.3060182Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:20.3068137Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:20.3076070Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:20.3078787Z (EngineCore_DP0 pid=22971) INFO 12-04 10:12:20 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:12:20.5894099Z (EngineCore_DP0 pid=22971) INFO 12-04 10:12:20 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:12:20.8033543Z (EngineCore_DP0 pid=22971) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-12-04T10:12:22.1339998Z (EngineCore_DP0 pid=22971) 2025-12-04T10:12:22.1340548Z chat_template.json: 0.00B [00:00, ?B/s] 2025-12-04T10:12:22.1340870Z chat_template.json: 1.69kB [00:00, 13.8MB/s] 2025-12-04T10:12:23.2063681Z (EngineCore_DP0 pid=22971) INFO 12-04 10:12:23 [gpu_model_runner.py:2840] Starting to load model AIDC-AI/Ovis2.5-2B... 2025-12-04T10:12:23.5715541Z (EngineCore_DP0 pid=22971) INFO 12-04 10:12:23 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:12:23.5980626Z (EngineCore_DP0 pid=22971) INFO 12-04 10:12:23 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:12:24.2419278Z (EngineCore_DP0 pid=22971) INFO 12-04 10:12:24 [gpu_model_runner.py:2902] Model loading took 2.2769 GiB and 0.126294 seconds 2025-12-04T10:12:24.2423676Z (EngineCore_DP0 pid=22971) INFO 12-04 10:12:24 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:12:24.2424591Z (EngineCore_DP0 pid=22971) INFO 12-04 10:12:24 [kv_cache_utils.py:1204] Maximum concurrency for 40,960 tokens per request: 64.00x 2025-12-04T10:12:25.6108670Z (EngineCore_DP0 pid=22971) INFO 12-04 10:12:25 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:12:25.6171245Z INFO 12-04 10:12:25 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:12:26.1070321Z PASSED 2025-12-04T10:12:26.1218896Z models/test_initialization.py::test_can_initialize_large_subset[FalconH1ForCausalLM] Fork a new process to run a test 23057 2025-12-04T10:12:26.1228287Z Fork a new process to run a test 0 2025-12-04T10:12:26.1494233Z INFO 12-04 10:12:26 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='FalconH1ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'tiiuae/Falcon-H1-0.5B-Base'} 2025-12-04T10:12:26.4020008Z 2025-12-04T10:12:26.4022423Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:12:26.4022823Z config.json: 1.58kB [00:00, 7.16MB/s] 2025-12-04T10:12:32.8471365Z INFO 12-04 10:12:32 [model.py:653] Resolved architecture: FalconH1ForCausalLM 2025-12-04T10:12:32.8471910Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:12:32.8724110Z INFO 12-04 10:12:32 [model.py:1714] Using max model len 16384 2025-12-04T10:12:33.0402075Z INFO 12-04 10:12:33 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:12:33.0402972Z INFO 12-04 10:12:33 [config.py:313] Disabling cascade attention since it is not supported for hybrid models. 2025-12-04T10:12:33.0979701Z INFO 12-04 10:12:33 [config.py:429] Setting attention block size to 800 tokens to ensure that attention page size is >= mamba page size. 2025-12-04T10:12:33.0981401Z INFO 12-04 10:12:33 [config.py:453] Padding mamba page size by 1.39% to ensure that mamba page size and attention page size are exactly equal. 2025-12-04T10:12:33.1337709Z 2025-12-04T10:12:33.1343123Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:12:33.1343593Z tokenizer_config.json: 99.7kB [00:00, 200MB/s] 2025-12-04T10:12:33.2104920Z 2025-12-04T10:12:33.2199865Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T10:12:33.2200538Z tokenizer.json: 2.35MB [00:00, 244MB/s] 2025-12-04T10:12:33.3146620Z 2025-12-04T10:12:33.3148532Z special_tokens_map.json: 0.00B [00:00, ?B/s] 2025-12-04T10:12:33.3149007Z special_tokens_map.json: 7.42kB [00:00, 27.8MB/s] 2025-12-04T10:12:33.4567741Z 2025-12-04T10:12:33.4569241Z generation_config.json: 0% 0.00/138 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:12:34.7967030Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:34.7994325Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:34.8001360Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:34.8008747Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:34.8016570Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:34.8024297Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:34.8027086Z (EngineCore_DP0 pid=23115) INFO 12-04 10:12:34 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:12:35.0882633Z (EngineCore_DP0 pid=23115) INFO 12-04 10:12:35 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:12:35.1285268Z (EngineCore_DP0 pid=23115) INFO 12-04 10:12:35 [gpu_model_runner.py:2840] Starting to load model tiiuae/Falcon-H1-0.5B-Base... 2025-12-04T10:12:35.4006657Z (EngineCore_DP0 pid=23115) INFO 12-04 10:12:35 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:12:35.4663512Z (EngineCore_DP0 pid=23115) INFO 12-04 10:12:35 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:12:36.0687751Z (EngineCore_DP0 pid=23115) INFO 12-04 10:12:36 [gpu_model_runner.py:2902] Model loading took 0.1515 GiB and 0.127058 seconds 2025-12-04T10:12:36.0693664Z (EngineCore_DP0 pid=23115) INFO 12-04 10:12:36 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,600 tokens 2025-12-04T10:12:36.0694617Z (EngineCore_DP0 pid=23115) INFO 12-04 10:12:36 [kv_cache_utils.py:1204] Maximum concurrency for 16,384 tokens per request: 1191.55x 2025-12-04T10:12:36.5127362Z (EngineCore_DP0 pid=23115) INFO 12-04 10:12:36 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:12:36.5141799Z INFO 12-04 10:12:36 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:12:36.9543687Z PASSED 2025-12-04T10:12:36.9692530Z models/test_initialization.py::test_can_initialize_large_subset[MistralForCausalLM] Fork a new process to run a test 23185 2025-12-04T10:12:36.9703189Z Fork a new process to run a test 0 2025-12-04T10:12:36.9982606Z INFO 12-04 10:12:36 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MistralForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'mistralai/Mistral-7B-Instruct-v0.1'} 2025-12-04T10:12:37.0628262Z 2025-12-04T10:12:37.0629203Z config.json: 0% 0.00/571 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:12:39.2615695Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:39.2642646Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:39.2649531Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:39.2657208Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:39.2664882Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:39.2672866Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:39.2675512Z (EngineCore_DP0 pid=23213) INFO 12-04 10:12:39 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:12:39.5504421Z (EngineCore_DP0 pid=23213) INFO 12-04 10:12:39 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:12:39.6187440Z (EngineCore_DP0 pid=23213) INFO 12-04 10:12:39 [gpu_model_runner.py:2840] Starting to load model mistralai/Mistral-7B-Instruct-v0.1... 2025-12-04T10:12:39.8951982Z (EngineCore_DP0 pid=23213) INFO 12-04 10:12:39 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:12:39.9608794Z (EngineCore_DP0 pid=23213) INFO 12-04 10:12:39 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:12:40.5624721Z (EngineCore_DP0 pid=23213) INFO 12-04 10:12:40 [gpu_model_runner.py:2902] Model loading took 0.9024 GiB and 0.124534 seconds 2025-12-04T10:12:40.5629244Z (EngineCore_DP0 pid=23213) INFO 12-04 10:12:40 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:12:40.5630152Z (EngineCore_DP0 pid=23213) INFO 12-04 10:12:40 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 213.06x 2025-12-04T10:12:41.1750140Z (EngineCore_DP0 pid=23213) INFO 12-04 10:12:41 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:12:41.1772905Z INFO 12-04 10:12:41 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:12:41.6323057Z PASSED 2025-12-04T10:12:41.6471772Z models/test_initialization.py::test_can_initialize_large_subset[QwenVLForConditionalGeneration] Fork a new process to run a test 23283 2025-12-04T10:12:41.6480949Z Fork a new process to run a test 0 2025-12-04T10:12:41.6756470Z INFO 12-04 10:12:41 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='QwenVLForConditionalGeneration', exist_overrides={'architectures': ['QwenVLForConditionalGeneration']}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen-VL'} 2025-12-04T10:12:41.7694231Z 2025-12-04T10:12:41.7696428Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:12:41.7696806Z config.json: 1.16kB [00:00, 5.93MB/s] 2025-12-04T10:12:41.8402094Z 2025-12-04T10:12:41.8404038Z configuration_qwen.py: 0.00B [00:00, ?B/s] 2025-12-04T10:12:41.8404428Z configuration_qwen.py: 2.09kB [00:00, 9.41MB/s] 2025-12-04T10:12:41.8497108Z A new version of the following files was downloaded from https://huggingface.co/Qwen/Qwen-VL: 2025-12-04T10:12:41.8497734Z - configuration_qwen.py 2025-12-04T10:12:41.8498644Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:12:48.5580081Z INFO 12-04 10:12:48 [model.py:653] Resolved architecture: QwenVLForConditionalGeneration 2025-12-04T10:12:48.5580614Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:12:48.5837121Z INFO 12-04 10:12:48 [model.py:1714] Using max model len 2048 2025-12-04T10:12:48.5839524Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T10:12:48.6532234Z INFO 12-04 10:12:48 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:12:48.7274123Z 2025-12-04T10:12:48.7275214Z tokenizer_config.json: 0% 0.00/173 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:12:54.4270478Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:54.4312249Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:54.4319273Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:54.4326983Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:54.4334997Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:54.4343368Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:12:54.4346122Z (EngineCore_DP0 pid=23341) INFO 12-04 10:12:54 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:12:54.7248819Z (EngineCore_DP0 pid=23341) INFO 12-04 10:12:54 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:12:55.1363451Z (EngineCore_DP0 pid=23341) WARNING 12-04 10:12:55 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T10:12:55.1372922Z (EngineCore_DP0 pid=23341) WARNING 12-04 10:12:55 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T10:12:55.1418966Z (EngineCore_DP0 pid=23341) INFO 12-04 10:12:55 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen-VL... 2025-12-04T10:12:55.4272023Z (EngineCore_DP0 pid=23341) INFO 12-04 10:12:55 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:12:55.4863130Z (EngineCore_DP0 pid=23341) INFO 12-04 10:12:55 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:12:56.1861019Z (EngineCore_DP0 pid=23341) INFO 12-04 10:12:56 [gpu_model_runner.py:2902] Model loading took 6.3133 GiB and 0.194334 seconds 2025-12-04T10:12:56.1865689Z (EngineCore_DP0 pid=23341) INFO 12-04 10:12:56 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-12-04T10:12:56.1866462Z (EngineCore_DP0 pid=23341) INFO 12-04 10:12:56 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 320.00x 2025-12-04T10:12:56.4960957Z (EngineCore_DP0 pid=23341) WARNING 12-04 10:12:56 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-12-04T10:12:56.5820341Z (EngineCore_DP0 pid=23341) WARNING 12-04 10:12:56 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T10:12:56.9028328Z (EngineCore_DP0 pid=23341) INFO 12-04 10:12:56 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:12:56.9045195Z INFO 12-04 10:12:56 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:12:57.4330188Z PASSED 2025-12-04T10:12:57.4480051Z models/test_initialization.py::test_can_initialize_large_subset[Mistral3ForConditionalGeneration] Fork a new process to run a test 23411 2025-12-04T10:12:57.4489375Z Fork a new process to run a test 0 2025-12-04T10:12:57.4766495Z INFO 12-04 10:12:57 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Mistral3ForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'mistralai/Mistral-Small-3.1-24B-Instruct-2503'} 2025-12-04T10:12:57.5560020Z 2025-12-04T10:12:57.5561891Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:12:57.5562214Z config.json: 1.18kB [00:00, 8.05MB/s] 2025-12-04T10:12:57.6699764Z 2025-12-04T10:12:57.6700850Z preprocessor_config.json: 0% 0.00/634 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:13:14.8935910Z (EngineCore_DP0 pid=23490) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:357: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-12-04T10:13:14.8937210Z (EngineCore_DP0 pid=23490) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-12-04T10:13:14.8939065Z (EngineCore_DP0 pid=23490) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-12-04T10:13:14.8940766Z (EngineCore_DP0 pid=23490) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 2025-12-04T10:13:14.8941849Z (EngineCore_DP0 pid=23490) dispatch key: ADInplaceOrView 2025-12-04T10:13:14.8942337Z (EngineCore_DP0 pid=23490) previous kernel: no debug info 2025-12-04T10:13:14.8943383Z (EngineCore_DP0 pid=23490) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-12-04T10:13:14.8944318Z (EngineCore_DP0 pid=23490) self.m.impl( 2025-12-04T10:13:14.9112372Z (EngineCore_DP0 pid=23490) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/transformers_utils/tokenizer.py:271: FutureWarning: It is strongly recommended to run mistral models with `--tokenizer-mode "mistral"` to ensure correct encoding and decoding. 2025-12-04T10:13:14.9113622Z (EngineCore_DP0 pid=23490) return cached_get_tokenizer( 2025-12-04T10:13:15.8317654Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:15.8346138Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:15.8353369Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:15.8360734Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:15.8367981Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:15.8375203Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:15.8377973Z (EngineCore_DP0 pid=23490) INFO 12-04 10:13:15 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:13:16.0897151Z (EngineCore_DP0 pid=23490) INFO 12-04 10:13:16 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:13:16.2601557Z (EngineCore_DP0 pid=23490) 2025-12-04T10:13:16.2601983Z processor_config.json: 0% 0.00/189 [00:00, model_arch='Olmo2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'allenai/OLMo-2-0425-1B'} 2025-12-04T10:13:29.9956638Z 2025-12-04T10:13:29.9958465Z config.json: 0% 0.00/623 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:13:38.9320458Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:38.9347062Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:38.9354435Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:38.9362431Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:38.9370366Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:38.9378598Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:38.9381546Z (EngineCore_DP0 pid=23839) INFO 12-04 10:13:38 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:13:39.2224284Z (EngineCore_DP0 pid=23839) INFO 12-04 10:13:39 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:13:39.2834839Z (EngineCore_DP0 pid=23839) INFO 12-04 10:13:39 [gpu_model_runner.py:2840] Starting to load model allenai/OLMo-2-0425-1B... 2025-12-04T10:13:39.5588212Z (EngineCore_DP0 pid=23839) INFO 12-04 10:13:39 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:13:39.5609482Z (EngineCore_DP0 pid=23839) INFO 12-04 10:13:39 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:13:40.2294519Z (EngineCore_DP0 pid=23839) INFO 12-04 10:13:40 [gpu_model_runner.py:2902] Model loading took 0.8916 GiB and 0.125379 seconds 2025-12-04T10:13:40.2298905Z (EngineCore_DP0 pid=23839) INFO 12-04 10:13:40 [kv_cache_utils.py:1199] GPU KV cache size: 1,310,720 tokens 2025-12-04T10:13:40.2299924Z (EngineCore_DP0 pid=23839) INFO 12-04 10:13:40 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 320.00x 2025-12-04T10:13:40.8863987Z (EngineCore_DP0 pid=23839) INFO 12-04 10:13:40 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:13:40.8882127Z INFO 12-04 10:13:40 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:13:41.3402307Z PASSED 2025-12-04T10:13:41.3552385Z models/test_initialization.py::test_can_initialize_large_subset[MiMoMTPModel] Fork a new process to run a test 23909 2025-12-04T10:13:41.3562518Z Fork a new process to run a test 0 2025-12-04T10:13:41.3839911Z INFO 12-04 10:13:41 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MiMoMTPModel', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'XiaomiMiMo/MiMo-7B-RL', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'XiaomiMiMo/MiMo-7B-RL'} 2025-12-04T10:13:41.7723177Z INFO 12-04 10:13:41 [model.py:653] Resolved architecture: MiMoForCausalLM 2025-12-04T10:13:41.7723702Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:13:41.7969158Z INFO 12-04 10:13:41 [model.py:1714] Using max model len 32768 2025-12-04T10:13:41.7970514Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T10:13:48.3972391Z INFO 12-04 10:13:48 [model.py:653] Resolved architecture: MiMoMTPModel 2025-12-04T10:13:48.3973148Z INFO 12-04 10:13:48 [model.py:1714] Using max model len 32768 2025-12-04T10:13:48.3975297Z INFO 12-04 10:13:48 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:13:48.8312680Z (EngineCore_DP0 pid=23966) INFO 12-04 10:13:48 [core.py:727] Waiting for init message from front-end. 2025-12-04T10:13:48.8383947Z (EngineCore_DP0 pid=23966) INFO 12-04 10:13:48 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='XiaomiMiMo/MiMo-7B-RL', speculative_config=SpeculativeConfig(method='mtp', model='XiaomiMiMo/MiMo-7B-RL', num_spec_tokens=1), tokenizer='XiaomiMiMo/MiMo-7B-RL', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=32768, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=XiaomiMiMo/MiMo-7B-RL, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:13:50.1717672Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:50.1744474Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:50.1751613Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:50.1758886Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:50.1766629Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:50.1775275Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:13:50.1777927Z (EngineCore_DP0 pid=23966) INFO 12-04 10:13:50 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:13:50.4633753Z (EngineCore_DP0 pid=23966) INFO 12-04 10:13:50 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:13:50.4781967Z (EngineCore_DP0 pid=23966) WARNING 12-04 10:13:50 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-12-04T10:13:50.5215250Z (EngineCore_DP0 pid=23966) INFO 12-04 10:13:50 [gpu_model_runner.py:2840] Starting to load model XiaomiMiMo/MiMo-7B-RL... 2025-12-04T10:13:50.7957745Z (EngineCore_DP0 pid=23966) INFO 12-04 10:13:50 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:13:50.8513690Z (EngineCore_DP0 pid=23966) INFO 12-04 10:13:50 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:13:50.9118990Z (EngineCore_DP0 pid=23966) INFO 12-04 10:13:50 [gpu_model_runner.py:2879] Loading drafter model... 2025-12-04T10:13:50.9248510Z (EngineCore_DP0 pid=23966) WARNING 12-04 10:13:50 [vllm.py:821] `torch.compile` is turned on, but the model XiaomiMiMo/MiMo-7B-RL does not support it. Please open an issue on GitHub if you want it to be supported. 2025-12-04T10:13:50.9253741Z (EngineCore_DP0 pid=23966) INFO 12-04 10:13:50 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-12-04T10:13:50.9254715Z (EngineCore_DP0 pid=23966) INFO 12-04 10:13:50 [eagle.py:1027] Loading EAGLE LM head weights from the target model. 2025-12-04T10:13:51.4712796Z (EngineCore_DP0 pid=23966) INFO 12-04 10:13:51 [gpu_model_runner.py:2902] Model loading took 3.0470 GiB and 0.129979 seconds 2025-12-04T10:13:51.4717176Z (EngineCore_DP0 pid=23966) INFO 12-04 10:13:51 [kv_cache_utils.py:1199] GPU KV cache size: 1,310,720 tokens 2025-12-04T10:13:51.4718261Z (EngineCore_DP0 pid=23966) INFO 12-04 10:13:51 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 40.00x 2025-12-04T10:13:52.1583576Z (EngineCore_DP0 pid=23966) INFO 12-04 10:13:52 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:13:52.1606460Z INFO 12-04 10:13:52 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:13:52.6291099Z PASSED 2025-12-04T10:13:52.6440857Z models/test_initialization.py::test_can_initialize_large_subset[GptOssForCausalLM] Fork a new process to run a test 24036 2025-12-04T10:13:52.6451286Z Fork a new process to run a test 0 2025-12-04T10:13:52.6724655Z INFO 12-04 10:13:52 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GptOssForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'lmsys/gpt-oss-20b-bf16'} 2025-12-04T10:13:52.7874352Z 2025-12-04T10:13:52.7876603Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:13:52.7877195Z config.json: 1.59kB [00:00, 9.08MB/s] 2025-12-04T10:13:59.1635213Z INFO 12-04 10:13:59 [model.py:653] Resolved architecture: GptOssForCausalLM 2025-12-04T10:13:59.1635722Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:13:59.1888679Z INFO 12-04 10:13:59 [model.py:1714] Using max model len 131072 2025-12-04T10:13:59.4209583Z INFO 12-04 10:13:59 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:13:59.4210628Z INFO 12-04 10:13:59 [config.py:263] Overriding max cuda graph capture size to 992 for performance. 2025-12-04T10:13:59.4586256Z 2025-12-04T10:13:59.4588149Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:13:59.4588542Z tokenizer_config.json: 4.20kB [00:00, 28.8MB/s] 2025-12-04T10:13:59.5655179Z 2025-12-04T10:13:59.8029168Z tokenizer.json: 0% 0.00/27.9M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [992, 976, 960, 944, 928, 912, 896, 880, 864, 848, 832, 816, 800, 784, 768, 752, 736, 720, 704, 688, 672, 656, 640, 624, 608, 592, 576, 560, 544, 528, 512, 496, 480, 464, 448, 432, 416, 400, 384, 368, 352, 336, 320, 304, 288, 272, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 992, 'local_cache_dir': None} 2025-12-04T10:14:02.1946346Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:02.1973113Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:02.1980035Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:02.1987232Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:02.1995544Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:02.2002827Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:02.2005840Z (EngineCore_DP0 pid=24114) INFO 12-04 10:14:02 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:14:02.4854089Z (EngineCore_DP0 pid=24114) INFO 12-04 10:14:02 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:14:02.5826399Z (EngineCore_DP0 pid=24114) INFO 12-04 10:14:02 [gpu_model_runner.py:2840] Starting to load model lmsys/gpt-oss-20b-bf16... 2025-12-04T10:14:02.8626207Z (EngineCore_DP0 pid=24114) INFO 12-04 10:14:02 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:14:02.9707975Z (EngineCore_DP0 pid=24114) INFO 12-04 10:14:02 [cuda.py:359] Using Triton backend on V1 engine. 2025-12-04T10:14:03.5812515Z (EngineCore_DP0 pid=24114) INFO 12-04 10:14:03 [gpu_model_runner.py:2902] Model loading took 2.3320 GiB and 0.168331 seconds 2025-12-04T10:14:03.5816719Z (EngineCore_DP0 pid=24114) INFO 12-04 10:14:03 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:14:03.5817619Z (EngineCore_DP0 pid=24114) INFO 12-04 10:14:03 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 628.94x 2025-12-04T10:14:05.2736569Z (EngineCore_DP0 pid=24114) INFO 12-04 10:14:05 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:14:05.2750727Z INFO 12-04 10:14:05 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:14:05.8032265Z PASSED 2025-12-04T10:14:05.8181482Z models/test_initialization.py::test_can_initialize_large_subset[FalconMambaForCausalLM] Fork a new process to run a test 24200 2025-12-04T10:14:05.8193101Z Fork a new process to run a test 0 2025-12-04T10:14:05.8467831Z INFO 12-04 10:14:05 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='FalconMambaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'tiiuae/falcon-mamba-7b-instruct'} 2025-12-04T10:14:05.9782162Z 2025-12-04T10:14:05.9784052Z config.json: 0% 0.00/870 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:14:08.1168240Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:08.1195915Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:08.1203121Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:08.1210729Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:08.1218914Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:08.1226649Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:08.1229958Z (EngineCore_DP0 pid=24208) INFO 12-04 10:14:08 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:14:08.4033604Z (EngineCore_DP0 pid=24208) INFO 12-04 10:14:08 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:14:08.5983110Z (EngineCore_DP0 pid=24208) INFO 12-04 10:14:08 [gpu_model_runner.py:2840] Starting to load model tiiuae/falcon-mamba-7b-instruct... 2025-12-04T10:14:08.8749032Z (EngineCore_DP0 pid=24208) INFO 12-04 10:14:08 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:14:09.4754554Z (EngineCore_DP0 pid=24208) INFO 12-04 10:14:09 [gpu_model_runner.py:2902] Model loading took 1.1886 GiB and 0.061524 seconds 2025-12-04T10:14:09.4759508Z (EngineCore_DP0 pid=24208) INFO 12-04 10:14:09 [kv_cache_utils.py:1199] GPU KV cache size: 70,639,616 tokens 2025-12-04T10:14:09.4760817Z (EngineCore_DP0 pid=24208) INFO 12-04 10:14:09 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 34492.00x 2025-12-04T10:14:09.9533236Z (EngineCore_DP0 pid=24208) INFO 12-04 10:14:09 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:14:09.9547531Z INFO 12-04 10:14:09 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:14:10.4034825Z PASSED 2025-12-04T10:14:10.4183851Z models/test_initialization.py::test_can_initialize_large_subset[GPTJForCausalLM] Fork a new process to run a test 24278 2025-12-04T10:14:10.4194661Z Fork a new process to run a test 0 2025-12-04T10:14:10.4477663Z INFO 12-04 10:14:10 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GPTJForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Milos/slovak-gpt-j-405M'} 2025-12-04T10:14:10.5473994Z 2025-12-04T10:14:10.5475449Z config.json: 0% 0.00/836 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:14:19.2485633Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:19.2512848Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:19.2520158Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:19.2527370Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:19.2535371Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:19.2543320Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:19.2546285Z (EngineCore_DP0 pid=24336) INFO 12-04 10:14:19 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:14:19.5351807Z (EngineCore_DP0 pid=24336) INFO 12-04 10:14:19 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:14:19.5747519Z (EngineCore_DP0 pid=24336) INFO 12-04 10:14:19 [gpu_model_runner.py:2840] Starting to load model Milos/slovak-gpt-j-405M... 2025-12-04T10:14:19.8425325Z (EngineCore_DP0 pid=24336) INFO 12-04 10:14:19 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:14:19.9080651Z (EngineCore_DP0 pid=24336) INFO 12-04 10:14:19 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:14:20.5035602Z (EngineCore_DP0 pid=24336) INFO 12-04 10:14:20 [gpu_model_runner.py:2902] Model loading took 0.2157 GiB and 0.124645 seconds 2025-12-04T10:14:20.5040344Z (EngineCore_DP0 pid=24336) INFO 12-04 10:14:20 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:14:20.5041697Z (EngineCore_DP0 pid=24336) INFO 12-04 10:14:20 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1280.00x 2025-12-04T10:14:21.1280850Z (EngineCore_DP0 pid=24336) INFO 12-04 10:14:21 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:14:21.1296726Z INFO 12-04 10:14:21 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:14:21.5669216Z PASSED 2025-12-04T10:14:21.5818209Z models/test_initialization.py::test_can_initialize_large_subset[Qwen3NextForCausalLM] Fork a new process to run a test 24406 2025-12-04T10:14:21.5828494Z Fork a new process to run a test 0 2025-12-04T10:14:21.5832356Z `transformers==4.56.2` installed, but `transformers>=4.56.3` is required to run this model. 2025-12-04T10:14:21.8666771Z PASSED 2025-12-04T10:14:21.8816108Z models/test_initialization.py::test_can_initialize_large_subset[Olmo3ForCausalLM] Fork a new process to run a test 24407 2025-12-04T10:14:21.8826791Z Fork a new process to run a test 0 2025-12-04T10:14:21.9100940Z INFO 12-04 10:14:21 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Olmo3ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'shanearora/2025-sep-a-base-model'} 2025-12-04T10:14:22.1411868Z 2025-12-04T10:14:22.1413724Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:14:22.1414030Z config.json: 1.46kB [00:00, 8.29MB/s] 2025-12-04T10:14:22.2416721Z INFO 12-04 10:14:22 [model.py:653] Resolved architecture: Olmo2ForCausalLM 2025-12-04T10:14:22.2417187Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:14:22.2662339Z INFO 12-04 10:14:22 [model.py:1714] Using max model len 8192 2025-12-04T10:14:22.4521550Z INFO 12-04 10:14:22 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:14:22.4878594Z 2025-12-04T10:14:22.4880427Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:14:22.4880776Z tokenizer_config.json: 4.34kB [00:00, 26.3MB/s] 2025-12-04T10:14:22.5645745Z 2025-12-04T10:14:22.5748151Z vocab.json: 0.00B [00:00, ?B/s] 2025-12-04T10:14:22.5748450Z vocab.json: 1.61MB [00:00, 158MB/s] 2025-12-04T10:14:22.6148053Z 2025-12-04T10:14:22.6202389Z merges.txt: 0.00B [00:00, ?B/s] 2025-12-04T10:14:22.6202684Z merges.txt: 917kB [00:00, 164MB/s] 2025-12-04T10:14:22.6825136Z 2025-12-04T10:14:22.7030305Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T10:14:22.7030738Z tokenizer.json: 7.14MB [00:00, 349MB/s] 2025-12-04T10:14:22.7666684Z 2025-12-04T10:14:22.7668009Z special_tokens_map.json: 0% 0.00/125 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:14:24.4711515Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:24.4738297Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:24.4744767Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:24.4751728Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:24.4758883Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:24.4766547Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:24.4769498Z (EngineCore_DP0 pid=24415) INFO 12-04 10:14:24 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:14:24.7639481Z (EngineCore_DP0 pid=24415) INFO 12-04 10:14:24 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:14:24.8301638Z (EngineCore_DP0 pid=24415) INFO 12-04 10:14:24 [gpu_model_runner.py:2840] Starting to load model shanearora/2025-sep-a-base-model... 2025-12-04T10:14:25.1025625Z (EngineCore_DP0 pid=24415) INFO 12-04 10:14:25 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:14:25.1049592Z (EngineCore_DP0 pid=24415) INFO 12-04 10:14:25 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:14:25.7647990Z (EngineCore_DP0 pid=24415) INFO 12-04 10:14:25 [gpu_model_runner.py:2902] Model loading took 1.9102 GiB and 0.126439 seconds 2025-12-04T10:14:25.7652870Z (EngineCore_DP0 pid=24415) INFO 12-04 10:14:25 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-12-04T10:14:25.7653770Z (EngineCore_DP0 pid=24415) INFO 12-04 10:14:25 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 79.84x 2025-12-04T10:14:26.3599410Z (EngineCore_DP0 pid=24415) INFO 12-04 10:14:26 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:14:26.3622036Z INFO 12-04 10:14:26 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:14:26.8260419Z PASSED 2025-12-04T10:14:26.8408087Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2Model] Fork a new process to run a test 24485 2025-12-04T10:14:26.8417363Z Fork a new process to run a test 0 2025-12-04T10:14:26.8689631Z INFO 12-04 10:14:26 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2Model', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ssmits/Qwen2-7B-Instruct-embed-base'} 2025-12-04T10:14:26.9720781Z 2025-12-04T10:14:26.9721524Z config.json: 0% 0.00/706 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:14:36.4416343Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:36.4443501Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:36.4450736Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:36.4458045Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:36.4466038Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:36.4473832Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:36.4476411Z (EngineCore_DP0 pid=24543) INFO 12-04 10:14:36 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:14:36.7271728Z (EngineCore_DP0 pid=24543) INFO 12-04 10:14:36 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:14:36.7954489Z (EngineCore_DP0 pid=24543) INFO 12-04 10:14:36 [gpu_model_runner.py:2840] Starting to load model ssmits/Qwen2-7B-Instruct-embed-base... 2025-12-04T10:14:37.0692998Z (EngineCore_DP0 pid=24543) INFO 12-04 10:14:37 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:14:37.1348110Z (EngineCore_DP0 pid=24543) INFO 12-04 10:14:37 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:14:37.7324057Z (EngineCore_DP0 pid=24543) INFO 12-04 10:14:37 [gpu_model_runner.py:2902] Model loading took 2.4751 GiB and 0.124709 seconds 2025-12-04T10:14:37.7328327Z (EngineCore_DP0 pid=24543) INFO 12-04 10:14:37 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:14:37.7329128Z (EngineCore_DP0 pid=24543) INFO 12-04 10:14:37 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 160.00x 2025-12-04T10:14:38.8381323Z (EngineCore_DP0 pid=24543) INFO 12-04 10:14:38 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:14:38.8400109Z INFO 12-04 10:14:38 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:14:39.3109437Z PASSED 2025-12-04T10:14:39.3258059Z models/test_initialization.py::test_can_initialize_large_subset[EagleLlama4ForCausalLM] Fork a new process to run a test 24613 2025-12-04T10:14:39.3268188Z Fork a new process to run a test 0 2025-12-04T10:14:39.3546054Z INFO 12-04 10:14:39 [utils.py:239] non-default args: {'tokenizer': 'meta-llama/Llama-4-Scout-17B-16E-Instruct', 'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='EagleLlama4ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'morgendave/EAGLE-Llama-4-Scout-17B-16E-Instruct', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'morgendave/EAGLE-Llama-4-Scout-17B-16E-Instruct'} 2025-12-04T10:14:39.4599595Z 2025-12-04T10:14:39.4601678Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:14:39.4602006Z config.json: 1.51kB [00:00, 9.63MB/s] 2025-12-04T10:14:39.5694812Z 2025-12-04T10:14:39.5696832Z preprocessor_config.json: 0% 0.00/636 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:14:48.6259439Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:48.6286594Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:48.6293468Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:48.6300545Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:48.6308778Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:48.6316261Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:48.6319337Z (EngineCore_DP0 pid=24671) INFO 12-04 10:14:48 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:14:48.9135594Z (EngineCore_DP0 pid=24671) INFO 12-04 10:14:48 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:14:48.9326406Z (EngineCore_DP0 pid=24671) WARNING 12-04 10:14:48 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-12-04T10:14:49.0437945Z (EngineCore_DP0 pid=24671) INFO 12-04 10:14:49 [gpu_model_runner.py:2840] Starting to load model morgendave/EAGLE-Llama-4-Scout-17B-16E-Instruct... 2025-12-04T10:14:49.3223565Z (EngineCore_DP0 pid=24671) INFO 12-04 10:14:49 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:14:49.4673733Z (EngineCore_DP0 pid=24671) INFO 12-04 10:14:49 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:14:49.5279264Z (EngineCore_DP0 pid=24671) INFO 12-04 10:14:49 [gpu_model_runner.py:2879] Loading drafter model... 2025-12-04T10:14:49.5529637Z (EngineCore_DP0 pid=24671) INFO 12-04 10:14:49 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-12-04T10:14:49.5530897Z (EngineCore_DP0 pid=24671) INFO 12-04 10:14:49 [eagle.py:1027] Loading EAGLE LM head weights from the target model. 2025-12-04T10:14:50.0971863Z (EngineCore_DP0 pid=24671) INFO 12-04 10:14:50 [gpu_model_runner.py:2902] Model loading took 6.3595 GiB and 0.230925 seconds 2025-12-04T10:14:50.0977342Z (EngineCore_DP0 pid=24671) INFO 12-04 10:14:50 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-12-04T10:14:50.0978303Z (EngineCore_DP0 pid=24671) INFO 12-04 10:14:50 [kv_cache_utils.py:1204] Maximum concurrency for 262,144 tokens per request: 2.50x 2025-12-04T10:14:51.6177569Z (EngineCore_DP0 pid=24671) INFO 12-04 10:14:51 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:14:51.6192008Z INFO 12-04 10:14:51 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:14:52.1679170Z PASSED 2025-12-04T10:14:52.1830095Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2_5OmniModel] Fork a new process to run a test 24741 2025-12-04T10:14:52.1839262Z Fork a new process to run a test 0 2025-12-04T10:14:52.2107583Z INFO 12-04 10:14:52 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2_5OmniModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen2.5-Omni-3B'} 2025-12-04T10:14:52.3149750Z 2025-12-04T10:14:52.3152372Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:14:52.3152685Z config.json: 13.2kB [00:00, 58.0MB/s] 2025-12-04T10:14:52.4706412Z Unrecognized keys in `rope_scaling` for 'rope_type'='default': {'mrope_section'} 2025-12-04T10:14:52.6061716Z 2025-12-04T10:14:52.6062894Z preprocessor_config.json: 0% 0.00/667 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:14:56.5324546Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:56.5354439Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:56.5362073Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:56.5369903Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:56.5377497Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:56.5386013Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:14:56.5388880Z (EngineCore_DP0 pid=24769) INFO 12-04 10:14:56 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:14:56.8229919Z (EngineCore_DP0 pid=24769) INFO 12-04 10:14:56 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:15:05.9037134Z (EngineCore_DP0 pid=24769) 2025-12-04T10:15:05.9037790Z chat_template.json: 0.00B [00:00, ?B/s] 2025-12-04T10:15:05.9038314Z chat_template.json: 1.31kB [00:00, 6.85MB/s] 2025-12-04T10:15:07.0714513Z (EngineCore_DP0 pid=24769) INFO 12-04 10:15:07 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen2.5-Omni-3B... 2025-12-04T10:15:07.4284805Z (EngineCore_DP0 pid=24769) INFO 12-04 10:15:07 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:15:07.4286500Z (EngineCore_DP0 pid=24769) WARNING 12-04 10:15:07 [qwen2_5_omni_thinker.py:865] flash_attn is not available, the model may not yield the exactly same result as the transformers implementation in the audio tower part. 2025-12-04T10:15:07.5504369Z (EngineCore_DP0 pid=24769) INFO 12-04 10:15:07 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:15:08.1795057Z (EngineCore_DP0 pid=24769) INFO 12-04 10:15:08 [gpu_model_runner.py:2902] Model loading took 3.7809 GiB and 0.195219 seconds 2025-12-04T10:15:08.1799894Z (EngineCore_DP0 pid=24769) INFO 12-04 10:15:08 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T10:15:08.1800848Z (EngineCore_DP0 pid=24769) INFO 12-04 10:15:08 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 320.00x 2025-12-04T10:15:09.9989154Z (EngineCore_DP0 pid=24769) INFO 12-04 10:15:09 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:15:10.0005027Z INFO 12-04 10:15:10 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:15:10.5003172Z PASSED 2025-12-04T10:15:10.5152717Z models/test_initialization.py::test_can_initialize_large_subset[SkyworkR1VChatModel] Fork a new process to run a test 24839 2025-12-04T10:15:10.5162819Z Fork a new process to run a test 0 2025-12-04T10:15:10.5441945Z INFO 12-04 10:15:10 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='SkyworkR1VChatModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Skywork/Skywork-R1V-38B'} 2025-12-04T10:15:10.6204715Z 2025-12-04T10:15:10.6206757Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:15:10.6207088Z config.json: 3.71kB [00:00, 20.7MB/s] 2025-12-04T10:15:10.7078838Z 2025-12-04T10:15:10.7079928Z configuration_skywork_chat.py: 0.00B [00:00, ?B/s] 2025-12-04T10:15:10.7080338Z configuration_skywork_chat.py: 3.77kB [00:00, 33.3MB/s] 2025-12-04T10:15:10.8680708Z 2025-12-04T10:15:10.8682427Z configuration_skywork_lm2.py: 0.00B [00:00, ?B/s] 2025-12-04T10:15:10.8682820Z configuration_skywork_lm2.py: 6.40kB [00:00, 47.9MB/s] 2025-12-04T10:15:10.8705960Z A new version of the following files was downloaded from https://huggingface.co/Skywork/Skywork-R1V-38B: 2025-12-04T10:15:10.8706508Z - configuration_skywork_lm2.py 2025-12-04T10:15:10.9160446Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:15:10.9161005Z 2025-12-04T10:15:10.9161167Z configuration_skywork_vit.py: 0.00B [00:00, ?B/s] 2025-12-04T10:15:10.9161535Z configuration_skywork_vit.py: 4.52kB [00:00, 41.5MB/s] 2025-12-04T10:15:10.9181519Z A new version of the following files was downloaded from https://huggingface.co/Skywork/Skywork-R1V-38B: 2025-12-04T10:15:10.9182017Z - configuration_skywork_vit.py 2025-12-04T10:15:10.9182632Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:15:10.9183676Z A new version of the following files was downloaded from https://huggingface.co/Skywork/Skywork-R1V-38B: 2025-12-04T10:15:10.9184178Z - configuration_skywork_chat.py 2025-12-04T10:15:10.9184440Z - configuration_skywork_lm2.py 2025-12-04T10:15:10.9184692Z - configuration_skywork_vit.py 2025-12-04T10:15:10.9185289Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:15:11.1727690Z 2025-12-04T10:15:11.1728262Z preprocessor_config.json: 0% 0.00/287 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:15:20.0169785Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:20.0197688Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:20.0204709Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:20.0212434Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:20.0220866Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:20.0229106Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:20.0231737Z (EngineCore_DP0 pid=24897) INFO 12-04 10:15:20 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:15:20.3146304Z (EngineCore_DP0 pid=24897) INFO 12-04 10:15:20 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:15:20.4345954Z (EngineCore_DP0 pid=24897) WARNING 12-04 10:15:20 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T10:15:20.4365382Z (EngineCore_DP0 pid=24897) WARNING 12-04 10:15:20 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T10:15:20.4621138Z (EngineCore_DP0 pid=24897) INFO 12-04 10:15:20 [gpu_model_runner.py:2840] Starting to load model Skywork/Skywork-R1V-38B... 2025-12-04T10:15:20.7478184Z (EngineCore_DP0 pid=24897) INFO 12-04 10:15:20 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:15:20.7526570Z (EngineCore_DP0 pid=24897) INFO 12-04 10:15:20 [layer.py:497] MultiHeadAttention attn_backend: _Backend.FLASH_ATTN, use_upstream_fa: False 2025-12-04T10:15:20.8142141Z (EngineCore_DP0 pid=24897) INFO 12-04 10:15:20 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:15:21.4326260Z (EngineCore_DP0 pid=24897) INFO 12-04 10:15:21 [gpu_model_runner.py:2902] Model loading took 4.2528 GiB and 0.128265 seconds 2025-12-04T10:15:21.4330700Z (EngineCore_DP0 pid=24897) INFO 12-04 10:15:21 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:15:21.4331693Z (EngineCore_DP0 pid=24897) INFO 12-04 10:15:21 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 20.00x 2025-12-04T10:15:21.8987348Z (EngineCore_DP0 pid=24897) WARNING 12-04 10:15:21 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T10:15:22.3039181Z (EngineCore_DP0 pid=24897) INFO 12-04 10:15:22 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:15:22.3054070Z INFO 12-04 10:15:22 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:15:22.8324185Z PASSED 2025-12-04T10:15:22.8473501Z models/test_initialization.py::test_can_initialize_large_subset[Phi4MMForCausalLM] Fork a new process to run a test 24983 2025-12-04T10:15:22.8483487Z Fork a new process to run a test 0 2025-12-04T10:15:22.8753808Z INFO 12-04 10:15:22 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Phi4MMForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'microsoft/Phi-4-multimodal-instruct'} 2025-12-04T10:15:22.9490791Z 2025-12-04T10:15:22.9493125Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:15:22.9493567Z config.json: 4.63kB [00:00, 21.3MB/s] 2025-12-04T10:15:23.0188920Z 2025-12-04T10:15:23.0190884Z configuration_phi4mm.py: 0.00B [00:00, ?B/s] 2025-12-04T10:15:23.0191413Z configuration_phi4mm.py: 11.0kB [00:00, 43.2MB/s] 2025-12-04T10:15:23.0293908Z A new version of the following files was downloaded from https://huggingface.co/microsoft/Phi-4-multimodal-instruct: 2025-12-04T10:15:23.0294740Z - configuration_phi4mm.py 2025-12-04T10:15:23.0295574Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:15:23.0669623Z INFO 12-04 10:15:23 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T10:15:23.2537892Z 2025-12-04T10:15:23.2538516Z preprocessor_config.json: 0% 0.00/482 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:15:33.1189294Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:33.1217828Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:33.1225265Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:33.1232714Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:33.1240114Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:33.1247755Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:33.1250687Z (EngineCore_DP0 pid=25062) INFO 12-04 10:15:33 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:15:33.4034482Z (EngineCore_DP0 pid=25062) INFO 12-04 10:15:33 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:15:33.5261241Z (EngineCore_DP0 pid=25062) 2025-12-04T10:15:33.5261646Z processor_config.json: 0% 0.00/121 [00:00, model_arch='Gemma2Model', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'BAAI/bge-multilingual-gemma2'} 2025-12-04T10:15:40.6994019Z 2025-12-04T10:15:40.6995773Z config.json: 0% 0.00/897 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:15:44.0323612Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:44.0350456Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:44.0357388Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:44.0365367Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:44.0373408Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:44.0381420Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:44.0383974Z (EngineCore_DP0 pid=25192) INFO 12-04 10:15:44 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:15:44.3260647Z (EngineCore_DP0 pid=25192) INFO 12-04 10:15:44 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:15:44.3817229Z (EngineCore_DP0 pid=25192) INFO 12-04 10:15:44 [gpu_model_runner.py:2840] Starting to load model BAAI/bge-multilingual-gemma2... 2025-12-04T10:15:44.6608675Z (EngineCore_DP0 pid=25192) INFO 12-04 10:15:44 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:15:44.7303427Z (EngineCore_DP0 pid=25192) INFO 12-04 10:15:44 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:15:45.3407602Z (EngineCore_DP0 pid=25192) INFO 12-04 10:15:45 [gpu_model_runner.py:2902] Model loading took 2.0825 GiB and 0.130758 seconds 2025-12-04T10:15:45.3412255Z (EngineCore_DP0 pid=25192) INFO 12-04 10:15:45 [kv_cache_utils.py:1199] GPU KV cache size: 1,310,720 tokens 2025-12-04T10:15:45.3413231Z (EngineCore_DP0 pid=25192) INFO 12-04 10:15:45 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 159.69x 2025-12-04T10:15:46.6399146Z (EngineCore_DP0 pid=25192) INFO 12-04 10:15:46 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:15:46.6413993Z INFO 12-04 10:15:46 [llm.py:337] Supported tasks: ['embed'] 2025-12-04T10:15:47.1251466Z PASSED 2025-12-04T10:15:47.1402091Z models/test_initialization.py::test_can_initialize_large_subset[TransformersForCausalLM] Fork a new process to run a test 25262 2025-12-04T10:15:47.1412807Z Fork a new process to run a test 0 2025-12-04T10:15:47.1690382Z INFO 12-04 10:15:47 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='TransformersForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'transformers', 'model': 'hmellor/Ilama-3.2-1B'} 2025-12-04T10:15:48.0944914Z INFO 12-04 10:15:48 [model.py:653] Resolved architecture: TransformersForCausalLM 2025-12-04T10:15:48.0945420Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:15:48.1186665Z INFO 12-04 10:15:48 [model.py:1939] Downcasting torch.float32 to torch.bfloat16. 2025-12-04T10:15:48.1189174Z INFO 12-04 10:15:48 [model.py:1714] Using max model len 131072 2025-12-04T10:15:48.1191365Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T10:15:48.1701846Z INFO 12-04 10:15:48 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:15:48.7672579Z (EngineCore_DP0 pid=25269) INFO 12-04 10:15:48 [core.py:727] Waiting for init message from front-end. 2025-12-04T10:15:48.7709062Z (EngineCore_DP0 pid=25269) INFO 12-04 10:15:48 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='hmellor/Ilama-3.2-1B', speculative_config=None, tokenizer='hmellor/Ilama-3.2-1B', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=131072, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=hmellor/Ilama-3.2-1B, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:15:50.1230341Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:50.1257143Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:50.1264435Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:50.1272341Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:50.1279725Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:50.1287537Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:15:50.1290372Z (EngineCore_DP0 pid=25269) INFO 12-04 10:15:50 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:15:50.4129555Z (EngineCore_DP0 pid=25269) INFO 12-04 10:15:50 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:15:50.6157524Z (EngineCore_DP0 pid=25269) INFO 12-04 10:15:50 [gpu_model_runner.py:2840] Starting to load model hmellor/Ilama-3.2-1B... 2025-12-04T10:15:50.8942777Z (EngineCore_DP0 pid=25269) INFO 12-04 10:15:50 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:15:50.8943840Z (EngineCore_DP0 pid=25269) INFO 12-04 10:15:50 [transformers.py:493] Using Transformers backend. 2025-12-04T10:15:50.9682660Z (EngineCore_DP0 pid=25269) `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:15:51.0839293Z (EngineCore_DP0 pid=25269) INFO 12-04 10:15:51 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:15:51.6835289Z (EngineCore_DP0 pid=25269) INFO 12-04 10:15:51 [gpu_model_runner.py:2902] Model loading took 0.6035 GiB and 0.247289 seconds 2025-12-04T10:15:51.6840183Z (EngineCore_DP0 pid=25269) INFO 12-04 10:15:51 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:15:51.6841131Z (EngineCore_DP0 pid=25269) INFO 12-04 10:15:51 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 40.00x 2025-12-04T10:15:52.9785145Z (EngineCore_DP0 pid=25269) INFO 12-04 10:15:52 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:15:52.9798972Z INFO 12-04 10:15:52 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:15:53.4720326Z PASSED 2025-12-04T10:15:53.4870904Z models/test_initialization.py::test_can_initialize_large_subset[AyaVisionForConditionalGeneration] Fork a new process to run a test 25339 2025-12-04T10:15:53.4881295Z Fork a new process to run a test 0 2025-12-04T10:15:53.5159733Z INFO 12-04 10:15:53 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='AyaVisionForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'CohereForAI/aya-vision-8b'} 2025-12-04T10:15:53.6597070Z 2025-12-04T10:15:53.6597797Z config.json: 0% 0.00/1.17k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:16:04.7317654Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:04.7345495Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:04.7352577Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:04.7360302Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:04.7368022Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:04.7376146Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:04.7379473Z (EngineCore_DP0 pid=25417) INFO 12-04 10:16:04 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:16:05.0257124Z (EngineCore_DP0 pid=25417) INFO 12-04 10:16:05 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:16:05.2620644Z (EngineCore_DP0 pid=25417) 2025-12-04T10:16:05.2621433Z processor_config.json: 0% 0.00/389 [00:00, model_arch='Ernie4_5_VLMoeForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'baidu/ERNIE-4.5-VL-28B-A3B-PT'} 2025-12-04T10:16:13.0042171Z 2025-12-04T10:16:13.0043834Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:16:13.0044170Z config.json: 2.08kB [00:00, 13.9MB/s] 2025-12-04T10:16:13.0684409Z 2025-12-04T10:16:13.0686958Z configuration_ernie4_5_vl.py: 0.00B [00:00, ?B/s] 2025-12-04T10:16:13.0687376Z configuration_ernie4_5_vl.py: 28.3kB [00:00, 103MB/s] 2025-12-04T10:16:13.0816693Z A new version of the following files was downloaded from https://huggingface.co/baidu/ERNIE-4.5-VL-28B-A3B-PT: 2025-12-04T10:16:13.0817243Z - configuration_ernie4_5_vl.py 2025-12-04T10:16:13.0817864Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:16:13.1124595Z INFO 12-04 10:16:13 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T10:16:13.2925627Z 2025-12-04T10:16:13.2926055Z preprocessor_config.json: 0% 0.00/477 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:16:24.6555421Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:24.6583454Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:24.6591073Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:24.6598411Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:24.6605722Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:24.6613098Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:24.6616287Z (EngineCore_DP0 pid=25582) INFO 12-04 10:16:24 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:16:24.9457432Z (EngineCore_DP0 pid=25582) INFO 12-04 10:16:24 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:16:30.3355236Z (EngineCore_DP0 pid=25582) 2025-12-04T10:16:30.3355620Z chat_template.json: 0.00B [00:00, ?B/s] 2025-12-04T10:16:30.3355964Z chat_template.json: 3.19kB [00:00, 25.2MB/s] 2025-12-04T10:16:32.4978628Z (EngineCore_DP0 pid=25582) INFO 12-04 10:16:32 [gpu_model_runner.py:2840] Starting to load model baidu/ERNIE-4.5-VL-28B-A3B-PT... 2025-12-04T10:16:32.8023375Z (EngineCore_DP0 pid=25582) INFO 12-04 10:16:32 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:16:32.8981308Z (EngineCore_DP0 pid=25582) INFO 12-04 10:16:32 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:16:32.9008105Z (EngineCore_DP0 pid=25582) WARNING 12-04 10:16:32 [vllm.py:821] `torch.compile` is turned on, but the model baidu/ERNIE-4.5-VL-28B-A3B-PT does not support it. Please open an issue on GitHub if you want it to be supported. 2025-12-04T10:16:33.4607098Z (EngineCore_DP0 pid=25582) INFO 12-04 10:16:33 [gpu_model_runner.py:2902] Model loading took 2.2676 GiB and 0.108401 seconds 2025-12-04T10:16:33.4611839Z (EngineCore_DP0 pid=25582) INFO 12-04 10:16:33 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:16:33.4612829Z (EngineCore_DP0 pid=25582) INFO 12-04 10:16:33 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 40.00x 2025-12-04T10:16:33.8553319Z (EngineCore_DP0 pid=25582) WARNING 12-04 10:16:33 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-12-04T10:16:34.8165165Z (EngineCore_DP0 pid=25582) INFO 12-04 10:16:34 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:16:34.8180284Z INFO 12-04 10:16:34 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:16:35.2985704Z PASSED 2025-12-04T10:16:35.3134416Z models/test_initialization.py::test_can_initialize_large_subset[MixtralForCausalLM] Fork a new process to run a test 25652 2025-12-04T10:16:35.3145321Z Fork a new process to run a test 0 2025-12-04T10:16:35.3420446Z INFO 12-04 10:16:35 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MixtralForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'mistralai/Mixtral-8x7B-Instruct-v0.1'} 2025-12-04T10:16:35.4124300Z 2025-12-04T10:16:35.4125434Z config.json: 0% 0.00/720 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:16:43.8326421Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:43.8353550Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:43.8360361Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:43.8367851Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:43.8375355Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:43.8383262Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:43.8386284Z (EngineCore_DP0 pid=25730) INFO 12-04 10:16:43 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:16:44.1247586Z (EngineCore_DP0 pid=25730) INFO 12-04 10:16:44 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:16:44.1936118Z (EngineCore_DP0 pid=25730) INFO 12-04 10:16:44 [gpu_model_runner.py:2840] Starting to load model mistralai/Mixtral-8x7B-Instruct-v0.1... 2025-12-04T10:16:44.4719211Z (EngineCore_DP0 pid=25730) INFO 12-04 10:16:44 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:16:44.5404806Z (EngineCore_DP0 pid=25730) INFO 12-04 10:16:44 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:16:45.1651786Z (EngineCore_DP0 pid=25730) INFO 12-04 10:16:45 [gpu_model_runner.py:2902] Model loading took 1.2305 GiB and 0.152431 seconds 2025-12-04T10:16:45.1655839Z (EngineCore_DP0 pid=25730) INFO 12-04 10:16:45 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:16:45.1656769Z (EngineCore_DP0 pid=25730) INFO 12-04 10:16:45 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 80.00x 2025-12-04T10:16:45.7570548Z (EngineCore_DP0 pid=25730) INFO 12-04 10:16:45 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:16:45.7588429Z INFO 12-04 10:16:45 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:16:46.2105541Z PASSED 2025-12-04T10:16:46.2255813Z models/test_initialization.py::test_can_initialize_large_subset[KimiVLForConditionalGeneration] Fork a new process to run a test 25800 2025-12-04T10:16:46.2266452Z Fork a new process to run a test 0 2025-12-04T10:16:46.2533273Z INFO 12-04 10:16:46 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='KimiVLForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'moonshotai/Kimi-VL-A3B-Instruct'} 2025-12-04T10:16:46.3521609Z 2025-12-04T10:16:46.3523599Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:16:46.3523929Z config.json: 2.00kB [00:00, 13.2MB/s] 2025-12-04T10:16:46.4594054Z 2025-12-04T10:16:46.4594760Z preprocessor_config.json: 0% 0.00/401 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:16:55.8669413Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:55.8696680Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:55.8704390Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:55.8712750Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:55.8721527Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:55.8729836Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:16:55.8732740Z (EngineCore_DP0 pid=25878) INFO 12-04 10:16:55 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:16:56.1630493Z (EngineCore_DP0 pid=25878) INFO 12-04 10:16:56 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:16:56.7978440Z (EngineCore_DP0 pid=25878) 2025-12-04T10:16:56.7979806Z configuration_kimi_vl.py: 0.00B [00:00, ?B/s] 2025-12-04T10:16:56.7980167Z configuration_kimi_vl.py: 13.2kB [00:00, 49.9MB/s] 2025-12-04T10:16:57.2201236Z (EngineCore_DP0 pid=25878) 2025-12-04T10:16:57.2203825Z processing_kimi_vl.py: 0.00B [00:00, ?B/s] 2025-12-04T10:16:57.2204196Z processing_kimi_vl.py: 7.54kB [00:00, 17.1MB/s] 2025-12-04T10:16:57.3494225Z (EngineCore_DP0 pid=25878) 2025-12-04T10:16:57.3494926Z image_processing_kimi_vl.py: 0.00B [00:00, ?B/s] 2025-12-04T10:16:57.3495463Z image_processing_kimi_vl.py: 4.68kB [00:00, 17.7MB/s] 2025-12-04T10:17:05.2222320Z (EngineCore_DP0 pid=25878) INFO 12-04 10:17:05 [gpu_model_runner.py:2840] Starting to load model moonshotai/Kimi-VL-A3B-Instruct... 2025-12-04T10:17:05.5036404Z (EngineCore_DP0 pid=25878) INFO 12-04 10:17:05 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:17:05.5380561Z (EngineCore_DP0 pid=25878) INFO 12-04 10:17:05 [cuda.py:328] Using Triton MLA backend on V1 engine. 2025-12-04T10:17:06.1704254Z (EngineCore_DP0 pid=25878) INFO 12-04 10:17:06 [gpu_model_runner.py:2902] Model loading took 1.4523 GiB and 0.114574 seconds 2025-12-04T10:17:06.1708441Z (EngineCore_DP0 pid=25878) INFO 12-04 10:17:06 [kv_cache_utils.py:1199] GPU KV cache size: 9,320,672 tokens 2025-12-04T10:17:06.1709418Z (EngineCore_DP0 pid=25878) INFO 12-04 10:17:06 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 71.11x 2025-12-04T10:17:06.6557443Z (EngineCore_DP0 pid=25878) WARNING 12-04 10:17:06 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-12-04T10:17:07.9012310Z (EngineCore_DP0 pid=25878) INFO 12-04 10:17:07 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:17:07.9027675Z INFO 12-04 10:17:07 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:17:08.3982410Z PASSED 2025-12-04T10:17:08.4133058Z models/test_initialization.py::test_can_initialize_large_subset[MiniMaxVL01ForConditionalGeneration] Fork a new process to run a test 25948 2025-12-04T10:17:08.4143476Z Fork a new process to run a test 0 2025-12-04T10:17:08.7007674Z PASSED 2025-12-04T10:17:08.7157406Z models/test_initialization.py::test_can_initialize_large_subset[MiniCPMForCausalLM] Fork a new process to run a test 25949 2025-12-04T10:17:08.7168100Z Fork a new process to run a test 0 2025-12-04T10:17:08.7443135Z INFO 12-04 10:17:08 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MiniCPMForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'openbmb/MiniCPM-2B-sft-bf16'} 2025-12-04T10:17:08.8904562Z 2025-12-04T10:17:08.8906636Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:17:08.8906951Z config.json: 1.01kB [00:00, 6.59MB/s] 2025-12-04T10:17:09.0200694Z 2025-12-04T10:17:09.0203705Z configuration_minicpm.py: 0.00B [00:00, ?B/s] 2025-12-04T10:17:09.0204114Z configuration_minicpm.py: 9.77kB [00:00, 32.1MB/s] 2025-12-04T10:17:09.0298002Z A new version of the following files was downloaded from https://huggingface.co/openbmb/MiniCPM-2B-sft-bf16: 2025-12-04T10:17:09.0298563Z - configuration_minicpm.py 2025-12-04T10:17:09.0299261Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:17:15.6801817Z INFO 12-04 10:17:15 [model.py:653] Resolved architecture: MiniCPMForCausalLM 2025-12-04T10:17:15.6802328Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:17:15.7057164Z INFO 12-04 10:17:15 [model.py:1714] Using max model len 4096 2025-12-04T10:17:15.7058972Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T10:17:15.7597560Z INFO 12-04 10:17:15 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:17:15.8079639Z 2025-12-04T10:17:15.8081025Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:17:15.8081450Z tokenizer_config.json: 1.12kB [00:00, 5.46MB/s] 2025-12-04T10:17:15.9143577Z 2025-12-04T10:17:16.2292190Z tokenizer.model: 0% 0.00/1.99M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:17:18.1456229Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:18.1484288Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:18.1490926Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:18.1498153Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:18.1505961Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:18.1514037Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:18.1516499Z (EngineCore_DP0 pid=26027) INFO 12-04 10:17:18 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:17:18.4311545Z (EngineCore_DP0 pid=26027) INFO 12-04 10:17:18 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:17:18.4971248Z (EngineCore_DP0 pid=26027) INFO 12-04 10:17:18 [gpu_model_runner.py:2840] Starting to load model openbmb/MiniCPM-2B-sft-bf16... 2025-12-04T10:17:18.7717202Z (EngineCore_DP0 pid=26027) INFO 12-04 10:17:18 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:17:18.8374187Z (EngineCore_DP0 pid=26027) INFO 12-04 10:17:18 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:17:19.4372751Z (EngineCore_DP0 pid=26027) INFO 12-04 10:17:19 [gpu_model_runner.py:2902] Model loading took 0.6422 GiB and 0.125790 seconds 2025-12-04T10:17:19.4377266Z (EngineCore_DP0 pid=26027) INFO 12-04 10:17:19 [kv_cache_utils.py:1199] GPU KV cache size: 1,165,072 tokens 2025-12-04T10:17:19.4378367Z (EngineCore_DP0 pid=26027) INFO 12-04 10:17:19 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 284.44x 2025-12-04T10:17:20.2129190Z (EngineCore_DP0 pid=26027) INFO 12-04 10:17:20 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:17:20.2146321Z INFO 12-04 10:17:20 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:17:20.6788301Z PASSED 2025-12-04T10:17:20.6938258Z models/test_initialization.py::test_can_initialize_large_subset[BaichuanForCausalLM] Fork a new process to run a test 26097 2025-12-04T10:17:20.6949027Z Fork a new process to run a test 0 2025-12-04T10:17:20.7224241Z INFO 12-04 10:17:20 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='BaichuanForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'baichuan-inc/Baichuan2-7B-chat'} 2025-12-04T10:17:20.8837685Z 2025-12-04T10:17:20.8839521Z config.json: 0% 0.00/758 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:17:30.5603631Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:30.5630939Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:30.5637541Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:30.5644570Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:30.5651906Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:30.5659405Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:30.5662009Z (EngineCore_DP0 pid=26175) INFO 12-04 10:17:30 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:17:30.8490086Z (EngineCore_DP0 pid=26175) INFO 12-04 10:17:30 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:17:30.9160323Z (EngineCore_DP0 pid=26175) INFO 12-04 10:17:30 [gpu_model_runner.py:2840] Starting to load model baichuan-inc/Baichuan2-7B-chat... 2025-12-04T10:17:31.1917753Z (EngineCore_DP0 pid=26175) INFO 12-04 10:17:31 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:17:31.2574111Z (EngineCore_DP0 pid=26175) INFO 12-04 10:17:31 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:17:31.8559073Z (EngineCore_DP0 pid=26175) INFO 12-04 10:17:31 [gpu_model_runner.py:2902] Model loading took 2.2959 GiB and 0.125312 seconds 2025-12-04T10:17:31.8563214Z (EngineCore_DP0 pid=26175) INFO 12-04 10:17:31 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-12-04T10:17:31.8563987Z (EngineCore_DP0 pid=26175) INFO 12-04 10:17:31 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 160.00x 2025-12-04T10:17:32.2769426Z (EngineCore_DP0 pid=26175) WARNING 12-04 10:17:32 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-12-04T10:17:32.7916206Z (EngineCore_DP0 pid=26175) INFO 12-04 10:17:32 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:17:32.7930399Z INFO 12-04 10:17:32 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:17:33.2717240Z PASSED 2025-12-04T10:17:33.2866676Z models/test_initialization.py::test_can_initialize_large_subset[MPTForCausalLM] Fork a new process to run a test 26245 2025-12-04T10:17:33.2877327Z Fork a new process to run a test 0 2025-12-04T10:17:33.3149182Z INFO 12-04 10:17:33 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MPTForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'mosaicml/mpt-7b'} 2025-12-04T10:17:33.3941992Z 2025-12-04T10:17:33.3943733Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:17:33.3944035Z config.json: 1.23kB [00:00, 7.77MB/s] 2025-12-04T10:17:39.8208790Z INFO 12-04 10:17:39 [model.py:653] Resolved architecture: MPTForCausalLM 2025-12-04T10:17:39.8209303Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:17:39.8461158Z INFO 12-04 10:17:39 [model.py:1714] Using max model len 2048 2025-12-04T10:17:40.0118920Z INFO 12-04 10:17:40 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:17:40.0483986Z 2025-12-04T10:17:40.0484459Z tokenizer_config.json: 0% 0.00/237 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:17:41.8260906Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:41.8288302Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:41.8295115Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:41.8303282Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:41.8310759Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:41.8318471Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:41.8321255Z (EngineCore_DP0 pid=26303) INFO 12-04 10:17:41 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:17:42.1146770Z (EngineCore_DP0 pid=26303) INFO 12-04 10:17:42 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:17:42.1787555Z (EngineCore_DP0 pid=26303) INFO 12-04 10:17:42 [gpu_model_runner.py:2840] Starting to load model mosaicml/mpt-7b... 2025-12-04T10:17:42.4569406Z (EngineCore_DP0 pid=26303) INFO 12-04 10:17:42 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:17:42.4999015Z (EngineCore_DP0 pid=26303) INFO 12-04 10:17:42 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:17:43.1022529Z (EngineCore_DP0 pid=26303) INFO 12-04 10:17:43 [gpu_model_runner.py:2902] Model loading took 0.7598 GiB and 0.100860 seconds 2025-12-04T10:17:43.1026862Z (EngineCore_DP0 pid=26303) INFO 12-04 10:17:43 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-12-04T10:17:43.1028129Z (EngineCore_DP0 pid=26303) INFO 12-04 10:17:43 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 320.00x 2025-12-04T10:17:43.5775167Z (EngineCore_DP0 pid=26303) INFO 12-04 10:17:43 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:17:43.5789341Z INFO 12-04 10:17:43 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:17:44.0258566Z PASSED 2025-12-04T10:17:44.0408047Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2ForCausalLM] Fork a new process to run a test 26373 2025-12-04T10:17:44.0418077Z Fork a new process to run a test 0 2025-12-04T10:17:44.0689884Z INFO 12-04 10:17:44 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen2-0.5B-Instruct'} 2025-12-04T10:17:44.1477193Z 2025-12-04T10:17:44.1478733Z config.json: 0% 0.00/659 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:17:46.4817538Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:46.4845120Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:46.4852155Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:46.4859846Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:46.4867421Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:46.4875689Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:46.4877822Z (EngineCore_DP0 pid=26381) INFO 12-04 10:17:46 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:17:46.7747193Z (EngineCore_DP0 pid=26381) INFO 12-04 10:17:46 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:17:46.8190390Z (EngineCore_DP0 pid=26381) INFO 12-04 10:17:46 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen2-0.5B-Instruct... 2025-12-04T10:17:47.0962341Z (EngineCore_DP0 pid=26381) INFO 12-04 10:17:47 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:17:47.1615101Z (EngineCore_DP0 pid=26381) INFO 12-04 10:17:47 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:17:47.7695708Z (EngineCore_DP0 pid=26381) INFO 12-04 10:17:47 [gpu_model_runner.py:2902] Model loading took 0.2861 GiB and 0.124130 seconds 2025-12-04T10:17:47.7700590Z (EngineCore_DP0 pid=26381) INFO 12-04 10:17:47 [kv_cache_utils.py:1199] GPU KV cache size: 20,971,520 tokens 2025-12-04T10:17:47.7701428Z (EngineCore_DP0 pid=26381) INFO 12-04 10:17:47 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 640.00x 2025-12-04T10:17:50.4842393Z (EngineCore_DP0 pid=26381) INFO 12-04 10:17:50 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:17:50.4857179Z INFO 12-04 10:17:50 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:17:50.9538915Z PASSED 2025-12-04T10:17:50.9689653Z models/test_initialization.py::test_can_initialize_large_subset[GraniteMoeSharedForCausalLM] Fork a new process to run a test 26451 2025-12-04T10:17:50.9698909Z Fork a new process to run a test 0 2025-12-04T10:17:50.9969118Z INFO 12-04 10:17:50 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GraniteMoeSharedForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ibm-research/moe-7b-1b-active-shared-experts'} 2025-12-04T10:17:51.1240291Z 2025-12-04T10:17:51.1242325Z config.json: 0% 0.00/901 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:17:59.9459505Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:59.9487827Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:59.9495053Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:59.9502713Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:59.9510730Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:59.9518894Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:17:59.9521619Z (EngineCore_DP0 pid=26515) INFO 12-04 10:17:59 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:18:00.2388947Z (EngineCore_DP0 pid=26515) INFO 12-04 10:18:00 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:18:00.2878375Z (EngineCore_DP0 pid=26515) INFO 12-04 10:18:00 [gpu_model_runner.py:2840] Starting to load model ibm-research/moe-7b-1b-active-shared-experts... 2025-12-04T10:18:00.5625685Z (EngineCore_DP0 pid=26515) INFO 12-04 10:18:00 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:18:00.6276897Z (EngineCore_DP0 pid=26515) INFO 12-04 10:18:00 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:18:01.2397598Z (EngineCore_DP0 pid=26515) INFO 12-04 10:18:01 [gpu_model_runner.py:2902] Model loading took 0.1748 GiB and 0.125836 seconds 2025-12-04T10:18:01.2402456Z (EngineCore_DP0 pid=26515) INFO 12-04 10:18:01 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:18:01.2403360Z (EngineCore_DP0 pid=26515) INFO 12-04 10:18:01 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 1280.00x 2025-12-04T10:18:02.1831097Z (EngineCore_DP0 pid=26515) INFO 12-04 10:18:02 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:18:02.1846705Z INFO 12-04 10:18:02 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:18:02.6373931Z PASSED 2025-12-04T10:18:02.6524088Z models/test_initialization.py::test_can_initialize_large_subset[DeciLMForCausalLM] Fork a new process to run a test 26585 2025-12-04T10:18:02.6534228Z Fork a new process to run a test 0 2025-12-04T10:18:02.6808539Z INFO 12-04 10:18:02 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='DeciLMForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'nvidia/Llama-3_3-Nemotron-Super-49B-v1'} 2025-12-04T10:18:02.8313271Z 2025-12-04T10:18:02.8316562Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:18:02.8316907Z config.json: 37.1kB [00:00, 122MB/s] 2025-12-04T10:18:02.9128617Z 2025-12-04T10:18:02.9131197Z configuration_decilm.py: 0.00B [00:00, ?B/s] 2025-12-04T10:18:02.9131600Z configuration_decilm.py: 2.57kB [00:00, 15.8MB/s] 2025-12-04T10:18:02.9644223Z 2025-12-04T10:18:02.9646309Z (…)nsformers_4_44_2__configuration_llama.py: 0.00B [00:00, ?B/s] 2025-12-04T10:18:02.9646891Z (…)nsformers_4_44_2__configuration_llama.py: 11.1kB [00:00, 77.9MB/s] 2025-12-04T10:18:03.0159114Z 2025-12-04T10:18:03.0161267Z (…)nsformers_4_44_2__modeling_rope_utils.py: 0.00B [00:00, ?B/s] 2025-12-04T10:18:03.0161797Z (…)nsformers_4_44_2__modeling_rope_utils.py: 28.1kB [00:00, 162MB/s] 2025-12-04T10:18:03.0245542Z A new version of the following files was downloaded from https://huggingface.co/nvidia/Llama-3_3-Nemotron-Super-49B-v1: 2025-12-04T10:18:03.0246181Z - transformers_4_44_2__modeling_rope_utils.py 2025-12-04T10:18:03.0246853Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:18:03.0248023Z A new version of the following files was downloaded from https://huggingface.co/nvidia/Llama-3_3-Nemotron-Super-49B-v1: 2025-12-04T10:18:03.0248608Z - transformers_4_44_2__configuration_llama.py 2025-12-04T10:18:03.0248935Z - transformers_4_44_2__modeling_rope_utils.py 2025-12-04T10:18:03.0249586Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:18:03.0919792Z 2025-12-04T10:18:03.0921697Z block_config.py: 0.00B [00:00, ?B/s] 2025-12-04T10:18:03.0922038Z block_config.py: 4.35kB [00:00, 21.8MB/s] 2025-12-04T10:18:03.0948596Z A new version of the following files was downloaded from https://huggingface.co/nvidia/Llama-3_3-Nemotron-Super-49B-v1: 2025-12-04T10:18:03.0949165Z - block_config.py 2025-12-04T10:18:03.0949752Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:18:03.0950708Z A new version of the following files was downloaded from https://huggingface.co/nvidia/Llama-3_3-Nemotron-Super-49B-v1: 2025-12-04T10:18:03.0951263Z - configuration_decilm.py 2025-12-04T10:18:03.0951549Z - transformers_4_44_2__configuration_llama.py 2025-12-04T10:18:03.0951848Z - block_config.py 2025-12-04T10:18:03.0952610Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:18:09.6692000Z INFO 12-04 10:18:09 [model.py:653] Resolved architecture: DeciLMForCausalLM 2025-12-04T10:18:09.6692499Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:18:09.6949124Z INFO 12-04 10:18:09 [model.py:1714] Using max model len 131072 2025-12-04T10:18:09.6950898Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T10:18:09.7502094Z INFO 12-04 10:18:09 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:18:09.7900936Z 2025-12-04T10:18:09.7903995Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:18:09.7904346Z tokenizer_config.json: 51.3kB [00:00, 180MB/s] 2025-12-04T10:18:09.9019322Z 2025-12-04T10:18:10.1618790Z tokenizer.json: 0% 0.00/17.2M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:18:12.1756917Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:12.1784125Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:12.1790631Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:12.1798359Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:12.1805831Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:12.1813187Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:12.1815655Z (EngineCore_DP0 pid=26663) INFO 12-04 10:18:12 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:18:12.4640126Z (EngineCore_DP0 pid=26663) INFO 12-04 10:18:12 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:18:12.5952090Z (EngineCore_DP0 pid=26663) INFO 12-04 10:18:12 [gpu_model_runner.py:2840] Starting to load model nvidia/Llama-3_3-Nemotron-Super-49B-v1... 2025-12-04T10:18:12.8724226Z (EngineCore_DP0 pid=26663) INFO 12-04 10:18:12 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:18:12.9787157Z (EngineCore_DP0 pid=26663) INFO 12-04 10:18:12 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:18:13.5844564Z (EngineCore_DP0 pid=26663) INFO 12-04 10:18:13 [gpu_model_runner.py:2902] Model loading took 4.8829 GiB and 0.165694 seconds 2025-12-04T10:18:13.5848859Z (EngineCore_DP0 pid=26663) INFO 12-04 10:18:13 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:18:13.5849776Z (EngineCore_DP0 pid=26663) INFO 12-04 10:18:13 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 20.00x 2025-12-04T10:18:14.5684524Z (EngineCore_DP0 pid=26663) INFO 12-04 10:18:14 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:18:14.5698723Z INFO 12-04 10:18:14 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:18:15.0935480Z PASSED 2025-12-04T10:18:15.1086063Z models/test_initialization.py::test_can_initialize_large_subset[BaiChuanForCausalLM] Fork a new process to run a test 26733 2025-12-04T10:18:15.1096191Z Fork a new process to run a test 0 2025-12-04T10:18:15.1367885Z INFO 12-04 10:18:15 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='BaiChuanForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'baichuan-inc/Baichuan-7B'} 2025-12-04T10:18:15.2243213Z 2025-12-04T10:18:15.2244928Z config.json: 0% 0.00/656 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:18:24.1277220Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:24.1304984Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:24.1312034Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:24.1319577Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:24.1327642Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:24.1335178Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:24.1338128Z (EngineCore_DP0 pid=26811) INFO 12-04 10:18:24 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:18:24.4244554Z (EngineCore_DP0 pid=26811) INFO 12-04 10:18:24 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:18:24.4916587Z (EngineCore_DP0 pid=26811) INFO 12-04 10:18:24 [gpu_model_runner.py:2840] Starting to load model baichuan-inc/Baichuan-7B... 2025-12-04T10:18:24.7770279Z (EngineCore_DP0 pid=26811) INFO 12-04 10:18:24 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:18:24.8428861Z (EngineCore_DP0 pid=26811) INFO 12-04 10:18:24 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:18:25.4701922Z (EngineCore_DP0 pid=26811) INFO 12-04 10:18:25 [gpu_model_runner.py:2902] Model loading took 1.3545 GiB and 0.126059 seconds 2025-12-04T10:18:25.4706597Z (EngineCore_DP0 pid=26811) INFO 12-04 10:18:25 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-12-04T10:18:25.4707542Z (EngineCore_DP0 pid=26811) INFO 12-04 10:18:25 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 160.00x 2025-12-04T10:18:25.6829584Z (EngineCore_DP0 pid=26811) WARNING 12-04 10:18:25 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-12-04T10:18:26.1716885Z (EngineCore_DP0 pid=26811) INFO 12-04 10:18:26 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:18:26.1739385Z INFO 12-04 10:18:26 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:18:26.6292138Z PASSED 2025-12-04T10:18:26.6439935Z models/test_initialization.py::test_can_initialize_large_subset[Terratorch] Fork a new process to run a test 26881 2025-12-04T10:18:26.6450055Z Fork a new process to run a test 0 2025-12-04T10:18:26.6726328Z INFO 12-04 10:18:26 [utils.py:239] non-default args: {'skip_tokenizer_init': True, 'load_format': 'dummy', 'dtype': torch.float16, 'gpu_memory_utilization': 0.8, 'max_num_seqs': 32, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Terratorch', exist_overrides={}, use_original_num_layers=False), 'enforce_eager': True, 'model_impl': 'vllm', 'model': 'ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11'} 2025-12-04T10:18:26.9970655Z INFO 12-04 10:18:26 [model.py:915] Resolved `--runner auto` to `--runner pooling`. Pass the value explicitly to silence this message. 2025-12-04T10:18:26.9971551Z INFO 12-04 10:18:26 [model.py:653] Resolved architecture: Terratorch 2025-12-04T10:18:26.9971990Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:18:27.0533870Z ERROR 12-04 10:18:27 [config.py:298] Error retrieving safetensors: 'ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11' is not a safetensors repo. Couldn't find 'model.safetensors.index.json' or 'model.safetensors' files., retrying 1 of 2 2025-12-04T10:18:29.1044929Z ERROR 12-04 10:18:29 [config.py:296] Error retrieving safetensors: 'ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11' is not a safetensors repo. Couldn't find 'model.safetensors.index.json' or 'model.safetensors' files. 2025-12-04T10:18:29.1046069Z INFO 12-04 10:18:29 [model.py:1939] Downcasting torch.float32 to torch.float16. 2025-12-04T10:18:29.1047201Z WARNING 12-04 10:18:29 [model.py:2042] The model's config.json does not contain any of the following keys to determine the original maximum length of the model: ['max_position_embeddings', 'n_positions', 'max_seq_len', 'seq_length', 'model_max_length', 'max_target_positions', 'max_sequence_length', 'max_seq_length', 'seq_len']. Assuming the model's maximum length is 2048. 2025-12-04T10:18:29.1048262Z INFO 12-04 10:18:29 [model.py:1714] Using max model len 2048 2025-12-04T10:18:29.1316596Z INFO 12-04 10:18:29 [arg_utils.py:1725] (Disabling) chunked prefill by default 2025-12-04T10:18:29.1317103Z INFO 12-04 10:18:29 [arg_utils.py:1728] (Disabling) prefix caching by default 2025-12-04T10:18:29.2843817Z INFO 12-04 10:18:29 [vllm.py:354] Cudagraph is disabled under eager mode 2025-12-04T10:18:29.2844822Z INFO 12-04 10:18:29 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T10:18:32.9854696Z WARNING 12-04 10:18:32 [__init__.py:2879] We must use the `spawn` multiprocessing start method. Overriding VLLM_WORKER_MULTIPROC_METHOD to 'spawn'. See https://docs.vllm.ai/en/latest/usage/troubleshooting.html#python-multiprocessing for more information. Reasons: CUDA is initialized 2025-12-04T10:18:35.8317870Z INFO 12-04 10:18:35 [__init__.py:224] Automatically detected platform cuda. 2025-12-04T10:18:38.3592822Z (EngineCore_DP0 pid=26890) INFO 12-04 10:18:38 [core.py:727] Waiting for init message from front-end. 2025-12-04T10:18:38.3845597Z (EngineCore_DP0 pid=26890) INFO 12-04 10:18:38 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11', speculative_config=None, tokenizer='ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11', skip_tokenizer_init=True, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11, enable_prefix_caching=False, chunked_prefill_enabled=False, pooler_config=PoolerConfig(pooling_type='All', normalize=None, dimensions=None, enable_chunked_processing=None, max_embed_len=None, activation=None, logit_bias=None, softmax=None, step_tag_id=None, returned_token_ids=None), compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-12-04T10:18:42.6185583Z (EngineCore_DP0 pid=26890) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:357: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-12-04T10:18:42.6186749Z (EngineCore_DP0 pid=26890) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-12-04T10:18:42.6188478Z (EngineCore_DP0 pid=26890) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-12-04T10:18:42.6189731Z (EngineCore_DP0 pid=26890) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 2025-12-04T10:18:42.6190356Z (EngineCore_DP0 pid=26890) dispatch key: ADInplaceOrView 2025-12-04T10:18:42.6190785Z (EngineCore_DP0 pid=26890) previous kernel: no debug info 2025-12-04T10:18:42.6191837Z (EngineCore_DP0 pid=26890) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-12-04T10:18:42.6192915Z (EngineCore_DP0 pid=26890) self.m.impl( 2025-12-04T10:18:42.8609730Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:42.8635211Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:42.8641944Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:42.8649138Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:42.8655990Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:42.8663419Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:18:42.8665978Z (EngineCore_DP0 pid=26890) INFO 12-04 10:18:42 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:18:43.1931393Z (EngineCore_DP0 pid=26890) INFO 12-04 10:18:43 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:18:43.2024111Z (EngineCore_DP0 pid=26890) INFO 12-04 10:18:43 [gpu_model_runner.py:2840] Starting to load model ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11... 2025-12-04T10:18:43.5184677Z (EngineCore_DP0 pid=26890) INFO 12-04 10:18:43 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:18:44.5653698Z (EngineCore_DP0 pid=26890) WARNING:root:Decoder UperNetDecoder does not have an `includes_head` attribute. Falling back to the value of the registry. 2025-12-04T10:18:44.5656481Z (EngineCore_DP0 pid=26890) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/terratorch/models/decoders/upernet_decoder.py:37: UserWarning: DeprecationWarning: scale_modules is deprecated and will be removed in future versions. Use LearnedInterpolateToPyramidal neck instead. 2025-12-04T10:18:44.5657988Z (EngineCore_DP0 pid=26890) warnings.warn( 2025-12-04T10:18:45.2310916Z (EngineCore_DP0 pid=26890) INFO 12-04 10:18:45 [gpu_model_runner.py:2902] Model loading took 0.5962 GiB and 1.068891 seconds 2025-12-04T10:18:46.4347528Z (EngineCore_DP0 pid=26890) INFO 12-04 10:18:46 [core.py:240] init engine (profile, create kv cache, warmup model) took 1.20 seconds 2025-12-04T10:18:46.4348882Z (EngineCore_DP0 pid=26890) INFO 12-04 10:18:46 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T10:18:46.4395942Z (EngineCore_DP0 pid=26890) INFO 12-04 10:18:46 [vllm.py:354] Cudagraph is disabled under eager mode 2025-12-04T10:18:46.4396910Z (EngineCore_DP0 pid=26890) INFO 12-04 10:18:46 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T10:18:46.7726331Z (EngineCore_DP0 pid=26890) INFO 12-04 10:18:46 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:18:46.7737430Z INFO 12-04 10:18:46 [llm.py:337] Supported tasks: ['encode'] 2025-12-04T10:18:46.9186772Z PASSED[rank0]:[W1204 10:18:46.861086589 ProcessGroupNCCL.cpp:1553] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator()) 2025-12-04T10:18:47.1755748Z 2025-12-04T10:18:47.1905373Z models/test_initialization.py::test_can_initialize_large_subset[TransformersForSequenceClassification] Fork a new process to run a test 27011 2025-12-04T10:18:47.1914601Z Fork a new process to run a test 0 2025-12-04T10:18:47.1918196Z `transformers==4.56.2` installed, but `transformers>=4.57.0.dev0` is required to run this model. 2025-12-04T10:18:47.4849782Z PASSED 2025-12-04T10:18:47.4997603Z models/test_initialization.py::test_can_initialize_large_subset[LlavaNextForConditionalGeneration] Fork a new process to run a test 27012 2025-12-04T10:18:47.5007335Z Fork a new process to run a test 0 2025-12-04T10:18:47.5276418Z INFO 12-04 10:18:47 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='LlavaNextForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'llava-hf/llava-v1.6-mistral-7b-hf'} 2025-12-04T10:18:47.6040161Z 2025-12-04T10:18:47.6042356Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:18:47.6042704Z config.json: 1.28kB [00:00, 8.15MB/s] 2025-12-04T10:18:47.7043350Z 2025-12-04T10:18:47.7044161Z preprocessor_config.json: 0% 0.00/772 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:19:03.3434893Z (EngineCore_DP0 pid=27092) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:357: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-12-04T10:19:03.3436232Z (EngineCore_DP0 pid=27092) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-12-04T10:19:03.3438361Z (EngineCore_DP0 pid=27092) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-12-04T10:19:03.3439981Z (EngineCore_DP0 pid=27092) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 2025-12-04T10:19:03.3440727Z (EngineCore_DP0 pid=27092) dispatch key: ADInplaceOrView 2025-12-04T10:19:03.3441248Z (EngineCore_DP0 pid=27092) previous kernel: no debug info 2025-12-04T10:19:03.3442543Z (EngineCore_DP0 pid=27092) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-12-04T10:19:03.3443710Z (EngineCore_DP0 pid=27092) self.m.impl( 2025-12-04T10:19:03.7105244Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:03.7133591Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:03.7140525Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:03.7147750Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:03.7155747Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:03.7163465Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:03.7166330Z (EngineCore_DP0 pid=27092) INFO 12-04 10:19:03 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:19:03.9734496Z (EngineCore_DP0 pid=27092) INFO 12-04 10:19:03 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:19:04.1000841Z (EngineCore_DP0 pid=27092) 2025-12-04T10:19:04.1001283Z processor_config.json: 0% 0.00/176 [00:00, model_arch='Tarsier2ForConditionalGeneration', exist_overrides={'architectures': ['Tarsier2ForConditionalGeneration']}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'omni-research/Tarsier2-Recap-7b'} 2025-12-04T10:19:14.9197172Z 2025-12-04T10:19:15.1363227Z config.json: 0% 0.00/5.52k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:19:25.6165531Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:25.6204204Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:25.6212443Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:25.6220583Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:25.6228451Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:25.6237547Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:25.6240565Z (EngineCore_DP0 pid=27470) INFO 12-04 10:19:25 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:19:25.9094256Z (EngineCore_DP0 pid=27470) INFO 12-04 10:19:25 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:19:27.3501553Z (EngineCore_DP0 pid=27470) INFO 12-04 10:19:27 [gpu_model_runner.py:2840] Starting to load model omni-research/Tarsier2-Recap-7b... 2025-12-04T10:19:27.6345120Z (EngineCore_DP0 pid=27470) INFO 12-04 10:19:27 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:19:27.7571554Z (EngineCore_DP0 pid=27470) INFO 12-04 10:19:27 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:19:28.4285441Z (EngineCore_DP0 pid=27470) INFO 12-04 10:19:28 [gpu_model_runner.py:2902] Model loading took 15.5314 GiB and 0.237939 seconds 2025-12-04T10:19:28.4292217Z (EngineCore_DP0 pid=27470) INFO 12-04 10:19:28 [kv_cache_utils.py:1199] GPU KV cache size: 187,232 tokens 2025-12-04T10:19:28.4293347Z (EngineCore_DP0 pid=27470) INFO 12-04 10:19:28 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 5.71x 2025-12-04T10:19:30.2947619Z (EngineCore_DP0 pid=27470) INFO 12-04 10:19:30 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:19:30.2969434Z INFO 12-04 10:19:30 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:19:30.7932382Z PASSED 2025-12-04T10:19:30.8079430Z models/test_initialization.py::test_can_initialize_large_subset[OlmoForCausalLM] Fork a new process to run a test 27540 2025-12-04T10:19:30.8089291Z Fork a new process to run a test 0 2025-12-04T10:19:30.8359371Z INFO 12-04 10:19:30 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='OlmoForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'allenai/OLMo-1B-hf'} 2025-12-04T10:19:30.9120771Z 2025-12-04T10:19:30.9122423Z config.json: 0% 0.00/632 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:19:39.5160433Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:39.5187442Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:39.5194358Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:39.5201530Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:39.5209284Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:39.5216848Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:39.5219580Z (EngineCore_DP0 pid=27598) INFO 12-04 10:19:39 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:19:39.8103194Z (EngineCore_DP0 pid=27598) INFO 12-04 10:19:39 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:19:39.8581913Z (EngineCore_DP0 pid=27598) INFO 12-04 10:19:39 [gpu_model_runner.py:2840] Starting to load model allenai/OLMo-1B-hf... 2025-12-04T10:19:40.1361217Z (EngineCore_DP0 pid=27598) INFO 12-04 10:19:40 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:19:40.2011454Z (EngineCore_DP0 pid=27598) INFO 12-04 10:19:40 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:19:40.8129539Z (EngineCore_DP0 pid=27598) INFO 12-04 10:19:40 [gpu_model_runner.py:2902] Model loading took 0.3174 GiB and 0.123528 seconds 2025-12-04T10:19:40.8134472Z (EngineCore_DP0 pid=27598) INFO 12-04 10:19:40 [kv_cache_utils.py:1199] GPU KV cache size: 1,310,720 tokens 2025-12-04T10:19:40.8135430Z (EngineCore_DP0 pid=27598) INFO 12-04 10:19:40 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 640.00x 2025-12-04T10:19:41.3034928Z (EngineCore_DP0 pid=27598) INFO 12-04 10:19:41 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:19:41.3054944Z INFO 12-04 10:19:41 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:19:41.7460532Z PASSED 2025-12-04T10:19:41.7608504Z models/test_initialization.py::test_can_initialize_large_subset[Fairseq2LlamaForCausalLM] Fork a new process to run a test 27668 2025-12-04T10:19:41.7618234Z Fork a new process to run a test 0 2025-12-04T10:19:41.7896669Z INFO 12-04 10:19:41 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Fairseq2LlamaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'mgleize/fairseq2-dummy-Llama-3.2-1B'} 2025-12-04T10:19:41.9409212Z 2025-12-04T10:19:41.9410960Z config.json: 0% 0.00/591 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:19:53.6978456Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:53.7005746Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:53.7013219Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:53.7022206Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:53.7030741Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:53.7038781Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:19:53.7041433Z (EngineCore_DP0 pid=27746) INFO 12-04 10:19:53 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:19:53.9828633Z (EngineCore_DP0 pid=27746) INFO 12-04 10:19:53 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:19:54.0625982Z (EngineCore_DP0 pid=27746) INFO 12-04 10:19:54 [gpu_model_runner.py:2840] Starting to load model mgleize/fairseq2-dummy-Llama-3.2-1B... 2025-12-04T10:19:54.3378640Z (EngineCore_DP0 pid=27746) INFO 12-04 10:19:54 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:19:54.4424853Z (EngineCore_DP0 pid=27746) INFO 12-04 10:19:54 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:19:55.0392734Z (EngineCore_DP0 pid=27746) INFO 12-04 10:19:55 [gpu_model_runner.py:2902] Model loading took 1.1094 GiB and 0.163987 seconds 2025-12-04T10:19:55.0397207Z (EngineCore_DP0 pid=27746) INFO 12-04 10:19:55 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:19:55.0398152Z (EngineCore_DP0 pid=27746) INFO 12-04 10:19:55 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 40.00x 2025-12-04T10:19:56.4187216Z (EngineCore_DP0 pid=27746) INFO 12-04 10:19:56 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:19:56.4201235Z INFO 12-04 10:19:56 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:19:56.9114709Z PASSED 2025-12-04T10:19:56.9262085Z models/test_initialization.py::test_can_initialize_large_subset[GPT2ForSequenceClassification] Fork a new process to run a test 27816 2025-12-04T10:19:56.9272308Z Fork a new process to run a test 0 2025-12-04T10:19:56.9538417Z INFO 12-04 10:19:56 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GPT2ForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'nie3e/sentiment-polish-gpt2-small'} 2025-12-04T10:19:57.2123328Z 2025-12-04T10:19:57.2125125Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:19:57.2125631Z config.json: 1.11kB [00:00, 6.65MB/s] 2025-12-04T10:20:03.6189684Z INFO 12-04 10:20:03 [model.py:915] Resolved `--runner auto` to `--runner pooling`. Pass the value explicitly to silence this message. 2025-12-04T10:20:03.6190796Z INFO 12-04 10:20:03 [model.py:967] Resolved `--convert auto` to `--convert classify`. Pass the value explicitly to silence this message. 2025-12-04T10:20:03.6191563Z INFO 12-04 10:20:03 [model.py:653] Resolved architecture: GPT2ForSequenceClassification 2025-12-04T10:20:03.6192084Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:20:03.6442126Z INFO 12-04 10:20:03 [model.py:1939] Downcasting torch.float32 to torch.float16. 2025-12-04T10:20:03.6443467Z INFO 12-04 10:20:03 [model.py:1714] Using max model len 2048 2025-12-04T10:20:03.6726694Z INFO 12-04 10:20:03 [arg_utils.py:1725] (Enabling) chunked prefill by default 2025-12-04T10:20:03.6727299Z INFO 12-04 10:20:03 [arg_utils.py:1728] (Enabling) prefix caching by default 2025-12-04T10:20:03.8146108Z INFO 12-04 10:20:03 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:20:03.8780888Z 2025-12-04T10:20:03.8783117Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:20:03.8783514Z tokenizer_config.json: 1.17kB [00:00, 5.04MB/s] 2025-12-04T10:20:04.0552283Z 2025-12-04T10:20:04.0607580Z vocab.json: 0.00B [00:00, ?B/s] 2025-12-04T10:20:04.0608015Z vocab.json: 907kB [00:00, 165MB/s] 2025-12-04T10:20:04.1001983Z 2025-12-04T10:20:04.1031883Z merges.txt: 0.00B [00:00, ?B/s] 2025-12-04T10:20:04.1032291Z merges.txt: 559kB [00:00, 190MB/s] 2025-12-04T10:20:04.1489133Z 2025-12-04T10:20:04.1575599Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T10:20:04.1576058Z tokenizer.json: 2.34MB [00:00, 266MB/s] 2025-12-04T10:20:04.2402923Z 2025-12-04T10:20:04.2404059Z special_tokens_map.json: 0% 0.00/437 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:20:05.8364450Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:05.8392723Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:05.8399170Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:05.8406722Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:05.8414026Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:05.8422130Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:05.8424554Z (EngineCore_DP0 pid=27875) INFO 12-04 10:20:05 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:20:06.1271715Z (EngineCore_DP0 pid=27875) INFO 12-04 10:20:06 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:20:06.1453308Z (EngineCore_DP0 pid=27875) WARNING 12-04 10:20:06 [interfaces_base.py:74] The model () is missing the `get_input_embeddings` method. 2025-12-04T10:20:06.1465229Z (EngineCore_DP0 pid=27875) INFO 12-04 10:20:06 [gpu_model_runner.py:2840] Starting to load model nie3e/sentiment-polish-gpt2-small... 2025-12-04T10:20:06.4196322Z (EngineCore_DP0 pid=27875) INFO 12-04 10:20:06 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:20:06.4235919Z (EngineCore_DP0 pid=27875) INFO 12-04 10:20:06 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:20:07.0401014Z (EngineCore_DP0 pid=27875) INFO 12-04 10:20:07 [gpu_model_runner.py:2902] Model loading took 0.0904 GiB and 0.067343 seconds 2025-12-04T10:20:07.0404217Z (EngineCore_DP0 pid=27875) INFO 12-04 10:20:07 [kv_cache_utils.py:1199] GPU KV cache size: 3,495,248 tokens 2025-12-04T10:20:07.0405270Z (EngineCore_DP0 pid=27875) INFO 12-04 10:20:07 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1706.66x 2025-12-04T10:20:07.6743551Z (EngineCore_DP0 pid=27875) INFO 12-04 10:20:07 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:20:07.6758159Z INFO 12-04 10:20:07 [llm.py:337] Supported tasks: ['classify'] 2025-12-04T10:20:08.1026479Z PASSED 2025-12-04T10:20:08.1174181Z models/test_initialization.py::test_can_initialize_large_subset[Llama_Nemotron_Nano_VL] Fork a new process to run a test 27945 2025-12-04T10:20:08.1183950Z Fork a new process to run a test 0 2025-12-04T10:20:08.1450138Z INFO 12-04 10:20:08 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Llama_Nemotron_Nano_VL', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'nvidia/Llama-3.1-Nemotron-Nano-VL-8B-V1'} 2025-12-04T10:20:08.2524545Z 2025-12-04T10:20:08.4868729Z config.json: 0% 0.00/7.65k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:20:27.0507721Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:27.0535580Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:27.0543144Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:27.0551138Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:27.0558314Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:27.0565984Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:27.0568543Z (EngineCore_DP0 pid=28028) INFO 12-04 10:20:27 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:20:27.3062259Z (EngineCore_DP0 pid=28028) INFO 12-04 10:20:27 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:20:27.4309599Z (EngineCore_DP0 pid=28028) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-12-04T10:20:27.4648138Z (EngineCore_DP0 pid=28028) 2025-12-04T10:20:27.4649109Z image_processing.py: 0.00B [00:00, ?B/s] 2025-12-04T10:20:27.4649432Z image_processing.py: 4.78kB [00:00, 30.7MB/s] 2025-12-04T10:20:27.4698020Z (EngineCore_DP0 pid=28028) A new version of the following files was downloaded from https://huggingface.co/nvidia/Llama-3.1-Nemotron-Nano-VL-8B-V1: 2025-12-04T10:20:27.4698981Z (EngineCore_DP0 pid=28028) - image_processing.py 2025-12-04T10:20:27.4700171Z (EngineCore_DP0 pid=28028) . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:20:27.4793999Z (EngineCore_DP0 pid=28028) WARNING 12-04 10:20:27 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T10:20:27.4798801Z (EngineCore_DP0 pid=28028) WARNING 12-04 10:20:27 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T10:20:27.5283230Z (EngineCore_DP0 pid=28028) INFO 12-04 10:20:27 [gpu_model_runner.py:2840] Starting to load model nvidia/Llama-3.1-Nemotron-Nano-VL-8B-V1... 2025-12-04T10:20:27.7743992Z (EngineCore_DP0 pid=28028) INFO 12-04 10:20:27 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:20:27.9287062Z (EngineCore_DP0 pid=28028) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/timm/models/registry.py:4: FutureWarning: Importing from timm.models.registry is deprecated, please import via timm.models 2025-12-04T10:20:27.9288515Z (EngineCore_DP0 pid=28028) warnings.warn(f"Importing from {__name__} is deprecated, please import via timm.models", FutureWarning) 2025-12-04T10:20:28.1908935Z (EngineCore_DP0 pid=28028) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:357: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-12-04T10:20:28.1910395Z (EngineCore_DP0 pid=28028) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-12-04T10:20:28.1911909Z (EngineCore_DP0 pid=28028) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-12-04T10:20:28.1913173Z (EngineCore_DP0 pid=28028) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 2025-12-04T10:20:28.1913797Z (EngineCore_DP0 pid=28028) dispatch key: ADInplaceOrView 2025-12-04T10:20:28.1914217Z (EngineCore_DP0 pid=28028) previous kernel: no debug info 2025-12-04T10:20:28.1915433Z (EngineCore_DP0 pid=28028) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-12-04T10:20:28.1916345Z (EngineCore_DP0 pid=28028) self.m.impl( 2025-12-04T10:20:28.2099292Z (EngineCore_DP0 pid=28028) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/timm/models/layers/__init__.py:48: FutureWarning: Importing from timm.models.layers is deprecated, please import via timm.layers 2025-12-04T10:20:28.2100641Z (EngineCore_DP0 pid=28028) warnings.warn(f"Importing from {__name__} is deprecated, please import via timm.layers", FutureWarning) 2025-12-04T10:20:28.2138542Z (EngineCore_DP0 pid=28028) No pretrained configuration specified for vit_huge_patch16_224 model. Using a default. Please add a config to the model pretrained_cfg registry or pass explicitly. 2025-12-04T10:20:28.3474554Z (EngineCore_DP0 pid=28028) INFO 12-04 10:20:28 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:20:28.9386092Z (EngineCore_DP0 pid=28028) INFO 12-04 10:20:28 [gpu_model_runner.py:2902] Model loading took 3.6833 GiB and 0.639933 seconds 2025-12-04T10:20:29.1939638Z (EngineCore_DP0 pid=28028) INFO 12-04 10:20:29 [gpu_model_runner.py:3647] Encoder cache will be initialized with a budget of 8192 tokens, and profiled with 2 image items of the maximum feature size. 2025-12-04T10:20:29.2138184Z (EngineCore_DP0 pid=28028) WARNING 12-04 10:20:29 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T10:20:29.2141297Z (EngineCore_DP0 pid=28028) WARNING 12-04 10:20:29 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-12-04T10:20:31.4552738Z (EngineCore_DP0 pid=28028) INFO 12-04 10:20:31 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/5d361cb273/rank_0_0/backbone for vLLM's torch.compile 2025-12-04T10:20:31.4553991Z (EngineCore_DP0 pid=28028) INFO 12-04 10:20:31 [backends.py:608] Dynamo bytecode transform time: 1.23 s 2025-12-04T10:20:32.0883244Z (EngineCore_DP0 pid=28028) [rank0]:W1204 10:20:32.087000 28028 site-packages/torch/_inductor/utils.py:1703] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:20:33.3614895Z (EngineCore_DP0 pid=28028) INFO 12-04 10:20:33 [backends.py:214] Cache the graph for dynamic shape for later use 2025-12-04T10:20:34.2435204Z (EngineCore_DP0 pid=28028) INFO 12-04 10:20:34 [backends.py:241] Compiling a graph for dynamic shape takes 2.58 s 2025-12-04T10:20:34.7866568Z (EngineCore_DP0 pid=28028) INFO 12-04 10:20:34 [monitor.py:33] torch.compile takes 3.82 s in total 2025-12-04T10:20:35.8799818Z (EngineCore_DP0 pid=28028) INFO 12-04 10:20:35 [gpu_worker.py:314] Available KV cache memory: 12.86 GiB 2025-12-04T10:20:36.1730641Z (EngineCore_DP0 pid=28028) INFO 12-04 10:20:36 [kv_cache_utils.py:1199] GPU KV cache size: 3,370,240 tokens 2025-12-04T10:20:36.1731592Z (EngineCore_DP0 pid=28028) INFO 12-04 10:20:36 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 25.71x 2025-12-04T10:20:36.4701565Z (EngineCore_DP0 pid=28028) 2025-12-04T10:20:36.5703902Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/67 [00:00=4.57.0.dev0` is required to run this model. 2025-12-04T10:20:39.3896050Z PASSED 2025-12-04T10:20:39.4044870Z models/test_initialization.py::test_can_initialize_large_subset[RobertaForMaskedLM] Fork a new process to run a test 28308 2025-12-04T10:20:39.4055561Z Fork a new process to run a test 0 2025-12-04T10:20:39.4328131Z INFO 12-04 10:20:39 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='RobertaForMaskedLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'sentence-transformers/all-roberta-large-v1'} 2025-12-04T10:20:39.5491989Z 2025-12-04T10:20:39.5492845Z config.json: 0% 0.00/650 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 256, 'local_cache_dir': None} 2025-12-04T10:20:48.2796693Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:48.2827383Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:48.2834913Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:48.2842078Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:48.2849205Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:48.2856976Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:48.2859791Z (EngineCore_DP0 pid=28367) INFO 12-04 10:20:48 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:20:48.5672936Z (EngineCore_DP0 pid=28367) INFO 12-04 10:20:48 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:20:48.5813386Z (EngineCore_DP0 pid=28367) INFO 12-04 10:20:48 [gpu_model_runner.py:2840] Starting to load model sentence-transformers/all-roberta-large-v1... 2025-12-04T10:20:48.8473044Z (EngineCore_DP0 pid=28367) INFO 12-04 10:20:48 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:20:48.8525938Z (EngineCore_DP0 pid=28367) INFO 12-04 10:20:48 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:20:49.4555706Z (EngineCore_DP0 pid=28367) INFO 12-04 10:20:49 [gpu_model_runner.py:2902] Model loading took 0.1211 GiB and 0.066183 seconds 2025-12-04T10:20:49.6207475Z (EngineCore_DP0 pid=28367) INFO 12-04 10:20:49 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T10:20:49.7078288Z (EngineCore_DP0 pid=28367) INFO 12-04 10:20:49 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T10:20:49.9916597Z (EngineCore_DP0 pid=28367) INFO 12-04 10:20:49 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:20:49.9937238Z INFO 12-04 10:20:49 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-12-04T10:20:50.4135879Z PASSED 2025-12-04T10:20:50.4283554Z models/test_initialization.py::test_can_initialize_large_subset[LLaMAForCausalLM] Fork a new process to run a test 28437 2025-12-04T10:20:50.4294566Z Fork a new process to run a test 0 2025-12-04T10:20:50.4295772Z Model is not available online 2025-12-04T10:20:50.7135783Z PASSED 2025-12-04T10:20:50.7283683Z models/test_initialization.py::test_can_initialize_large_subset[InternS1ForConditionalGeneration] Fork a new process to run a test 28438 2025-12-04T10:20:50.7294407Z Fork a new process to run a test 0 2025-12-04T10:20:50.7564092Z INFO 12-04 10:20:50 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='InternS1ForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'internlm/Intern-S1'} 2025-12-04T10:20:50.8433012Z 2025-12-04T10:20:50.8435014Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:20:50.8435317Z config.json: 2.62kB [00:00, 16.3MB/s] 2025-12-04T10:20:50.9094186Z 2025-12-04T10:20:50.9095873Z configuration_interns1.py: 0.00B [00:00, ?B/s] 2025-12-04T10:20:50.9096263Z configuration_interns1.py: 10.2kB [00:00, 78.9MB/s] 2025-12-04T10:20:50.9192682Z A new version of the following files was downloaded from https://huggingface.co/internlm/Intern-S1: 2025-12-04T10:20:50.9193205Z - configuration_interns1.py 2025-12-04T10:20:50.9193857Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:20:51.1377484Z 2025-12-04T10:20:51.1378162Z preprocessor_config.json: 0% 0.00/694 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:20:54.0858229Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:54.0887376Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:54.0894738Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:54.0902396Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:54.0909595Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:54.0917658Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:20:54.0920492Z (EngineCore_DP0 pid=28467) INFO 12-04 10:20:54 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:20:54.3722370Z (EngineCore_DP0 pid=28467) INFO 12-04 10:20:54 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:20:54.5120239Z (EngineCore_DP0 pid=28467) 2025-12-04T10:20:54.5120953Z processor_config.json: 0% 0.00/154 [00:00 32768). Running this sequence through the model will result in indexing errors 2025-12-04T10:21:04.6151194Z (EngineCore_DP0 pid=28467) INFO 12-04 10:21:04 [gpu_model_runner.py:2840] Starting to load model internlm/Intern-S1... 2025-12-04T10:21:04.8993048Z (EngineCore_DP0 pid=28467) INFO 12-04 10:21:04 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:21:04.9035785Z (EngineCore_DP0 pid=28467) INFO 12-04 10:21:04 [layer.py:497] MultiHeadAttention attn_backend: _Backend.FLASH_ATTN, use_upstream_fa: False 2025-12-04T10:21:04.9528383Z (EngineCore_DP0 pid=28467) INFO 12-04 10:21:04 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:21:05.5735559Z (EngineCore_DP0 pid=28467) INFO 12-04 10:21:05 [gpu_model_runner.py:2902] Model loading took 2.9280 GiB and 0.121633 seconds 2025-12-04T10:21:05.5739763Z (EngineCore_DP0 pid=28467) INFO 12-04 10:21:05 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:21:05.5740676Z (EngineCore_DP0 pid=28467) INFO 12-04 10:21:05 [kv_cache_utils.py:1204] Maximum concurrency for 65,536 tokens per request: 80.00x 2025-12-04T10:21:05.8361067Z (EngineCore_DP0 pid=28467) WARNING 12-04 10:21:05 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-12-04T10:21:08.7915732Z (EngineCore_DP0 pid=28467) INFO 12-04 10:21:08 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:21:08.8155879Z INFO 12-04 10:21:08 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:21:09.2972878Z PASSED 2025-12-04T10:21:09.3120405Z models/test_initialization.py::test_can_initialize_large_subset[XverseForCausalLM] Fork a new process to run a test 28537 2025-12-04T10:21:09.3130328Z Fork a new process to run a test 0 2025-12-04T10:21:09.3403841Z INFO 12-04 10:21:09 [utils.py:239] non-default args: {'tokenizer': 'meta-llama/Llama-2-7b', 'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='XverseForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'xverse/XVERSE-7B-Chat'} 2025-12-04T10:21:09.4639449Z 2025-12-04T10:21:09.4640582Z config.json: 0% 0.00/687 [00:00. This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message 2025-12-04T10:21:10.3560755Z You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message. 2025-12-04T10:21:10.6174257Z 2025-12-04T10:21:10.6175223Z generation_config.json: 0% 0.00/232 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:21:11.9500432Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:11.9527896Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:11.9535225Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:11.9543047Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:11.9550718Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:11.9558558Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:11.9561333Z (EngineCore_DP0 pid=28565) INFO 12-04 10:21:11 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:21:12.2435247Z (EngineCore_DP0 pid=28565) INFO 12-04 10:21:12 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:21:12.3118411Z (EngineCore_DP0 pid=28565) INFO 12-04 10:21:12 [gpu_model_runner.py:2840] Starting to load model xverse/XVERSE-7B-Chat... 2025-12-04T10:21:12.5963618Z (EngineCore_DP0 pid=28565) INFO 12-04 10:21:12 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:21:12.6620669Z (EngineCore_DP0 pid=28565) INFO 12-04 10:21:12 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:21:13.2753474Z (EngineCore_DP0 pid=28565) INFO 12-04 10:21:13 [gpu_model_runner.py:2902] Model loading took 1.9141 GiB and 0.125279 seconds 2025-12-04T10:21:13.2757727Z (EngineCore_DP0 pid=28565) INFO 12-04 10:21:13 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-12-04T10:21:13.2758691Z (EngineCore_DP0 pid=28565) INFO 12-04 10:21:13 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 80.00x 2025-12-04T10:21:14.0353857Z (EngineCore_DP0 pid=28565) INFO 12-04 10:21:14 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:21:14.0370307Z INFO 12-04 10:21:14 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:21:14.4895208Z PASSED 2025-12-04T10:21:14.5043916Z models/test_initialization.py::test_can_initialize_large_subset[OPTForCausalLM] Fork a new process to run a test 28635 2025-12-04T10:21:14.5052878Z Fork a new process to run a test 0 2025-12-04T10:21:14.5326876Z INFO 12-04 10:21:14 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='OPTForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'facebook/opt-125m'} 2025-12-04T10:21:14.6092888Z 2025-12-04T10:21:14.6094122Z config.json: 0% 0.00/651 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:21:23.0858442Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:23.0886985Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:23.0894242Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:23.0902338Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:23.0910164Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:23.0918282Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:23.0920954Z (EngineCore_DP0 pid=28693) INFO 12-04 10:21:23 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:21:23.3728729Z (EngineCore_DP0 pid=28693) INFO 12-04 10:21:23 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:21:23.4125631Z (EngineCore_DP0 pid=28693) INFO 12-04 10:21:23 [gpu_model_runner.py:2840] Starting to load model facebook/opt-125m... 2025-12-04T10:21:23.6799444Z (EngineCore_DP0 pid=28693) INFO 12-04 10:21:23 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:21:23.6838440Z (EngineCore_DP0 pid=28693) INFO 12-04 10:21:23 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:21:24.2826882Z (EngineCore_DP0 pid=28693) INFO 12-04 10:21:24 [gpu_model_runner.py:2902] Model loading took 0.0884 GiB and 0.063431 seconds 2025-12-04T10:21:24.2831860Z (EngineCore_DP0 pid=28693) INFO 12-04 10:21:24 [kv_cache_utils.py:1199] GPU KV cache size: 3,495,248 tokens 2025-12-04T10:21:24.2832814Z (EngineCore_DP0 pid=28693) INFO 12-04 10:21:24 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1706.66x 2025-12-04T10:21:24.8635157Z (EngineCore_DP0 pid=28693) INFO 12-04 10:21:24 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:21:24.8655257Z INFO 12-04 10:21:24 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:21:25.2939301Z PASSED 2025-12-04T10:21:25.3086796Z models/test_initialization.py::test_can_initialize_large_subset[Ernie4_5ForCausalLM] Fork a new process to run a test 28763 2025-12-04T10:21:25.3097332Z Fork a new process to run a test 0 2025-12-04T10:21:25.3371902Z INFO 12-04 10:21:25 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Ernie4_5ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'baidu/ERNIE-4.5-0.3B-PT'} 2025-12-04T10:21:25.6095595Z 2025-12-04T10:21:25.6097484Z config.json: 0% 0.00/633 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:21:35.4945519Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:35.4976248Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:35.4983628Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:35.4991266Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:35.4998978Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:35.5006581Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:35.5009466Z (EngineCore_DP0 pid=28842) INFO 12-04 10:21:35 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:21:35.7888663Z (EngineCore_DP0 pid=28842) INFO 12-04 10:21:35 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:21:35.8481530Z (EngineCore_DP0 pid=28842) INFO 12-04 10:21:35 [gpu_model_runner.py:2840] Starting to load model baidu/ERNIE-4.5-0.3B-PT... 2025-12-04T10:21:36.1290148Z (EngineCore_DP0 pid=28842) INFO 12-04 10:21:36 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:21:36.1960391Z (EngineCore_DP0 pid=28842) INFO 12-04 10:21:36 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:21:36.8042064Z (EngineCore_DP0 pid=28842) INFO 12-04 10:21:36 [gpu_model_runner.py:2902] Model loading took 0.2549 GiB and 0.127079 seconds 2025-12-04T10:21:36.8046007Z (EngineCore_DP0 pid=28842) INFO 12-04 10:21:36 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T10:21:36.8046887Z (EngineCore_DP0 pid=28842) INFO 12-04 10:21:36 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 80.00x 2025-12-04T10:21:39.1878913Z (EngineCore_DP0 pid=28842) INFO 12-04 10:21:39 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:21:39.1893037Z INFO 12-04 10:21:39 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:21:39.6615657Z PASSED 2025-12-04T10:21:39.6763587Z models/test_initialization.py::test_can_initialize_large_subset[VoxtralForConditionalGeneration] Fork a new process to run a test 28912 2025-12-04T10:21:39.6773551Z Fork a new process to run a test 0 2025-12-04T10:21:39.6775190Z Model is not available online 2025-12-04T10:21:39.9613597Z PASSED 2025-12-04T10:21:39.9763236Z models/test_initialization.py::test_can_initialize_large_subset[Blip2ForConditionalGeneration] Fork a new process to run a test 28913 2025-12-04T10:21:39.9773202Z Fork a new process to run a test 0 2025-12-04T10:21:40.0047546Z INFO 12-04 10:21:40 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Blip2ForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Salesforce/blip2-opt-2.7b'} 2025-12-04T10:21:40.0757650Z 2025-12-04T10:21:40.0759612Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:21:40.0759912Z config.json: 1.03kB [00:00, 6.59MB/s] 2025-12-04T10:21:40.1659777Z 2025-12-04T10:21:40.1660843Z preprocessor_config.json: 0% 0.00/432 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:21:48.5992879Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:48.6021010Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:48.6028108Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:48.6035986Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:48.6044116Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:48.6051814Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:21:48.6054636Z (EngineCore_DP0 pid=28971) INFO 12-04 10:21:48 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:21:48.8863784Z (EngineCore_DP0 pid=28971) INFO 12-04 10:21:48 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:21:48.9895292Z (EngineCore_DP0 pid=28971) 2025-12-04T10:21:48.9895708Z processor_config.json: 0% 0.00/68.0 [00:00, model_arch='PaliGemmaForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'google/paligemma-3b-mix-224'} 2025-12-04T10:21:52.0006770Z 2025-12-04T10:21:52.0007891Z config.json: 0% 0.00/1.03k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:22:02.5635158Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:02.5661925Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:02.5668648Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:02.5676672Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:02.5684194Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:02.5691753Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:02.5694431Z (EngineCore_DP0 pid=29137) INFO 12-04 10:22:02 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:22:02.8515526Z (EngineCore_DP0 pid=29137) INFO 12-04 10:22:02 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:22:03.0545146Z (EngineCore_DP0 pid=29137) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-12-04T10:22:08.1168908Z (EngineCore_DP0 pid=29137) You are passing both `text` and `images` to `PaliGemmaProcessor`. The processor expects special image tokens in the text, as many tokens as there are images per each text. It is recommended to add `` tokens in the very beginning of your text. For this call, we will infer how many images each text has and add special tokens. 2025-12-04T10:22:08.5385619Z (EngineCore_DP0 pid=29137) INFO 12-04 10:22:08 [gpu_model_runner.py:2840] Starting to load model google/paligemma-3b-mix-224... 2025-12-04T10:22:08.8195539Z (EngineCore_DP0 pid=29137) INFO 12-04 10:22:08 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:22:08.8243781Z (EngineCore_DP0 pid=29137) INFO 12-04 10:22:08 [layer.py:497] MultiHeadAttention attn_backend: _Backend.XFORMERS, use_upstream_fa: False 2025-12-04T10:22:08.8498915Z (EngineCore_DP0 pid=29137) INFO 12-04 10:22:08 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:22:08.8509040Z (EngineCore_DP0 pid=29137) WARNING 12-04 10:22:08 [gemma.py:67] Gemma's activation function was incorrectly set to exact GeLU in the config JSON file when it was initially released. Changing the activation function to approximate GeLU (`gelu_pytorch_tanh`). If you want to use the legacy `gelu_pytorch_tanh`, edit the config JSON to set `hidden_activation=gelu_pytorch_tanh` instead of `hidden_act`. See https://github.com/huggingface/transformers/pull/29402 for more details. 2025-12-04T10:22:09.4596393Z (EngineCore_DP0 pid=29137) INFO 12-04 10:22:09 [gpu_model_runner.py:2902] Model loading took 1.2249 GiB and 0.086249 seconds 2025-12-04T10:22:09.4599427Z (EngineCore_DP0 pid=29137) INFO 12-04 10:22:09 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T10:22:09.4600758Z (EngineCore_DP0 pid=29137) INFO 12-04 10:22:09 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 1280.00x 2025-12-04T10:22:10.2594167Z (EngineCore_DP0 pid=29137) You are passing both `text` and `images` to `PaliGemmaProcessor`. The processor expects special image tokens in the text, as many tokens as there are images per each text. It is recommended to add `` tokens in the very beginning of your text. For this call, we will infer how many images each text has and add special tokens. 2025-12-04T10:22:11.5649018Z (EngineCore_DP0 pid=29137) INFO 12-04 10:22:11 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:22:11.5663953Z INFO 12-04 10:22:11 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:22:12.0824833Z PASSED 2025-12-04T10:22:12.0975628Z models/test_initialization.py::test_can_initialize_large_subset[MiniMaxText01ForCausalLM] Fork a new process to run a test 29223 2025-12-04T10:22:12.0985003Z Fork a new process to run a test 0 2025-12-04T10:22:12.1265297Z INFO 12-04 10:22:12 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'revision': 'a59aa9cbc53b9fb8742ca4e9e1531b9802b6fdc3', 'hf_overrides': functools.partial(, model_arch='MiniMaxText01ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'MiniMaxAI/MiniMax-Text-01'} 2025-12-04T10:22:12.2678566Z 2025-12-04T10:22:12.2680885Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:22:12.2681209Z config.json: 1.89kB [00:00, 8.97MB/s] 2025-12-04T10:22:12.3123201Z 2025-12-04T10:22:12.3126520Z configuration_minimax_text_01.py: 0.00B [00:00, ?B/s] 2025-12-04T10:22:12.3126969Z configuration_minimax_text_01.py: 7.36kB [00:00, 23.3MB/s] 2025-12-04T10:22:12.3231218Z You are using a model of type minimax_text_01 to instantiate a model of type MiniMaxText01. This is not supported for all configurations of models and can yield errors. 2025-12-04T10:22:12.5189262Z INFO 12-04 10:22:12 [model.py:653] Resolved architecture: MiniMaxText01ForCausalLM 2025-12-04T10:22:12.5189747Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:22:12.6142801Z 2025-12-04T10:22:12.6165532Z model.safetensors.index.json: 0.00B [00:00, ?B/s] 2025-12-04T10:22:12.6165973Z model.safetensors.index.json: 823kB [00:00, 372MB/s] 2025-12-04T10:22:12.6429651Z 2025-12-04T10:22:12.7639577Z Parse safetensors files: 0% 0/413 [00:00= mamba page size. 2025-12-04T10:22:14.1187571Z 2025-12-04T10:22:14.1189789Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:22:14.1190172Z tokenizer_config.json: 1.39kB [00:00, 5.09MB/s] 2025-12-04T10:22:14.2099218Z 2025-12-04T10:22:14.2301731Z vocab.json: 0.00B [00:00, ?B/s] 2025-12-04T10:22:14.2302040Z vocab.json: 4.71MB [00:00, 233MB/s] 2025-12-04T10:22:14.2812586Z 2025-12-04T10:22:14.2938540Z merges.txt: 0.00B [00:00, ?B/s] 2025-12-04T10:22:14.2938852Z merges.txt: 2.41MB [00:00, 191MB/s] 2025-12-04T10:22:14.3762860Z 2025-12-04T10:22:14.4127965Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T10:22:14.4128281Z tokenizer.json: 9.72MB [00:00, 266MB/s] 2025-12-04T10:22:15.0667821Z You are using a model of type minimax_text_01 to instantiate a model of type MiniMaxText01. This is not supported for all configurations of models and can yield errors. 2025-12-04T10:22:15.0903502Z (EngineCore_DP0 pid=29251) INFO 12-04 10:22:15 [core.py:727] Waiting for init message from front-end. 2025-12-04T10:22:15.0939314Z (EngineCore_DP0 pid=29251) INFO 12-04 10:22:15 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='MiniMaxAI/MiniMax-Text-01', speculative_config=None, tokenizer='MiniMaxAI/MiniMax-Text-01', skip_tokenizer_init=False, tokenizer_mode=auto, revision=a59aa9cbc53b9fb8742ca4e9e1531b9802b6fdc3, tokenizer_revision=a59aa9cbc53b9fb8742ca4e9e1531b9802b6fdc3, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=10240000, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=MiniMaxAI/MiniMax-Text-01, enable_prefix_caching=False, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:22:16.4011722Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:16.4040834Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:16.4049521Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:16.4058468Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:16.4067380Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:16.4076248Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:16.4078877Z (EngineCore_DP0 pid=29251) INFO 12-04 10:22:16 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:22:16.6956882Z (EngineCore_DP0 pid=29251) INFO 12-04 10:22:16 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:22:17.6602170Z (EngineCore_DP0 pid=29251) INFO 12-04 10:22:17 [gpu_model_runner.py:2840] Starting to load model MiniMaxAI/MiniMax-Text-01... 2025-12-04T10:22:17.9425436Z (EngineCore_DP0 pid=29251) INFO 12-04 10:22:17 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:22:18.5984800Z (EngineCore_DP0 pid=29251) INFO 12-04 10:22:18 [gpu_model_runner.py:2902] Model loading took 7.9713 GiB and 0.108434 seconds 2025-12-04T10:22:18.5990114Z (EngineCore_DP0 pid=29251) INFO 12-04 10:22:18 [kv_cache_utils.py:1199] GPU KV cache size: 52,428,800,000 tokens 2025-12-04T10:22:18.5991126Z (EngineCore_DP0 pid=29251) INFO 12-04 10:22:18 [kv_cache_utils.py:1204] Maximum concurrency for 10,240,000 tokens per request: 5120.00x 2025-12-04T10:22:19.4794834Z (EngineCore_DP0 pid=29251) INFO 12-04 10:22:19 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:22:19.4808964Z INFO 12-04 10:22:19 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:22:20.8875401Z PASSED 2025-12-04T10:22:20.9023893Z models/test_initialization.py::test_can_initialize_large_subset[LlavaNextVideoForConditionalGeneration] Fork a new process to run a test 29321 2025-12-04T10:22:20.9033530Z Fork a new process to run a test 0 2025-12-04T10:22:20.9309483Z INFO 12-04 10:22:20 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='LlavaNextVideoForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'llava-hf/LLaVA-NeXT-Video-7B-hf'} 2025-12-04T10:22:21.0034801Z 2025-12-04T10:22:21.0036414Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:22:21.0036722Z config.json: 1.41kB [00:00, 9.71MB/s] 2025-12-04T10:22:21.1045297Z 2025-12-04T10:22:21.1046279Z preprocessor_config.json: 0% 0.00/741 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:22:36.9107698Z (EngineCore_DP0 pid=29401) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:357: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-12-04T10:22:36.9109297Z (EngineCore_DP0 pid=29401) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-12-04T10:22:36.9111092Z (EngineCore_DP0 pid=29401) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-12-04T10:22:36.9112667Z (EngineCore_DP0 pid=29401) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 2025-12-04T10:22:36.9113429Z (EngineCore_DP0 pid=29401) dispatch key: ADInplaceOrView 2025-12-04T10:22:36.9113949Z (EngineCore_DP0 pid=29401) previous kernel: no debug info 2025-12-04T10:22:36.9115428Z (EngineCore_DP0 pid=29401) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:926 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-12-04T10:22:36.9116384Z (EngineCore_DP0 pid=29401) self.m.impl( 2025-12-04T10:22:37.7507103Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:37.7535254Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:37.7542046Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:37.7549147Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:37.7555896Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:37.7562529Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:37.7565359Z (EngineCore_DP0 pid=29401) INFO 12-04 10:22:37 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:22:38.0076930Z (EngineCore_DP0 pid=29401) INFO 12-04 10:22:38 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:22:38.1023184Z (EngineCore_DP0 pid=29401) 2025-12-04T10:22:38.1023825Z video_preprocessor_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:22:38.1024223Z video_preprocessor_config.json: 1.08kB [00:00, 8.96MB/s] 2025-12-04T10:22:38.1713601Z (EngineCore_DP0 pid=29401) 2025-12-04T10:22:38.1714922Z processor_config.json: 0% 0.00/209 [00:00, model_arch='NemotronHForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'nvidia/Nemotron-H-8B-Base-8K'} 2025-12-04T10:22:49.0829128Z 2025-12-04T10:22:49.0831046Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:22:49.0831435Z config.json: 1.50kB [00:00, 8.92MB/s] 2025-12-04T10:22:49.1788436Z 2025-12-04T10:22:49.1790432Z configuration_nemotron_h.py: 0.00B [00:00, ?B/s] 2025-12-04T10:22:49.1790919Z configuration_nemotron_h.py: 12.1kB [00:00, 39.6MB/s] 2025-12-04T10:22:49.1888437Z A new version of the following files was downloaded from https://huggingface.co/nvidia/Nemotron-H-8B-Base-8K: 2025-12-04T10:22:49.1888991Z - configuration_nemotron_h.py 2025-12-04T10:22:49.1889623Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:22:55.7448853Z INFO 12-04 10:22:55 [model.py:653] Resolved architecture: NemotronHForCausalLM 2025-12-04T10:22:55.7449377Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:22:55.7705416Z INFO 12-04 10:22:55 [model.py:1714] Using max model len 8192 2025-12-04T10:22:55.7707753Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T10:22:55.8287816Z INFO 12-04 10:22:55 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:22:55.8290123Z INFO 12-04 10:22:55 [config.py:313] Disabling cascade attention since it is not supported for hybrid models. 2025-12-04T10:22:55.8744959Z INFO 12-04 10:22:55 [config.py:429] Setting attention block size to 528 tokens to ensure that attention page size is >= mamba page size. 2025-12-04T10:22:55.8748101Z INFO 12-04 10:22:55 [config.py:453] Padding mamba page size by 0.19% to ensure that mamba page size and attention page size are exactly equal. 2025-12-04T10:22:55.9166583Z 2025-12-04T10:22:55.9172776Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:22:55.9173133Z tokenizer_config.json: 178kB [00:00, 344MB/s] 2025-12-04T10:22:56.0145283Z 2025-12-04T10:22:56.5255737Z tokenizer.json: 0% 0.00/17.1M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:22:58.6119127Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:58.6146067Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:58.6153141Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:58.6160138Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:58.6167497Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:58.6175222Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:22:58.6177886Z (EngineCore_DP0 pid=29744) INFO 12-04 10:22:58 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:22:58.9035792Z (EngineCore_DP0 pid=29744) INFO 12-04 10:22:58 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:22:58.9808491Z (EngineCore_DP0 pid=29744) INFO 12-04 10:22:58 [gpu_model_runner.py:2840] Starting to load model nvidia/Nemotron-H-8B-Base-8K... 2025-12-04T10:22:59.2584887Z (EngineCore_DP0 pid=29744) INFO 12-04 10:22:59 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:22:59.2675037Z (EngineCore_DP0 pid=29744) INFO 12-04 10:22:59 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:22:59.9214761Z (EngineCore_DP0 pid=29744) INFO 12-04 10:22:59 [gpu_model_runner.py:2902] Model loading took 15.1125 GiB and 0.112054 seconds 2025-12-04T10:22:59.9226981Z (EngineCore_DP0 pid=29744) INFO 12-04 10:22:59 [kv_cache_utils.py:1199] GPU KV cache size: 93,456 tokens 2025-12-04T10:22:59.9228275Z (EngineCore_DP0 pid=29744) INFO 12-04 10:22:59 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 56.41x 2025-12-04T10:23:00.9232972Z (EngineCore_DP0 pid=29744) INFO 12-04 10:23:00 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:23:00.9249209Z INFO 12-04 10:23:00 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:23:01.4058406Z PASSED 2025-12-04T10:23:01.4207455Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2MoeForCausalLM] Fork a new process to run a test 29814 2025-12-04T10:23:01.4217157Z Fork a new process to run a test 0 2025-12-04T10:23:01.4490224Z INFO 12-04 10:23:01 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2MoeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen1.5-MoE-A2.7B-Chat'} 2025-12-04T10:23:01.5436068Z 2025-12-04T10:23:01.5437303Z config.json: 0% 0.00/920 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:23:10.3818747Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:10.3845839Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:10.3852806Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:10.3860594Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:10.3868305Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:10.3876320Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:10.3878792Z (EngineCore_DP0 pid=29872) INFO 12-04 10:23:10 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:23:10.6682726Z (EngineCore_DP0 pid=29872) INFO 12-04 10:23:10 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:23:10.7208363Z (EngineCore_DP0 pid=29872) INFO 12-04 10:23:10 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen1.5-MoE-A2.7B-Chat... 2025-12-04T10:23:10.9942101Z (EngineCore_DP0 pid=29872) INFO 12-04 10:23:10 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:23:11.0596628Z (EngineCore_DP0 pid=29872) INFO 12-04 10:23:11 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:23:11.6610699Z (EngineCore_DP0 pid=29872) INFO 12-04 10:23:11 [gpu_model_runner.py:2902] Model loading took 1.2959 GiB and 0.125664 seconds 2025-12-04T10:23:11.6615327Z (EngineCore_DP0 pid=29872) INFO 12-04 10:23:11 [kv_cache_utils.py:1199] GPU KV cache size: 1,310,720 tokens 2025-12-04T10:23:11.6616249Z (EngineCore_DP0 pid=29872) INFO 12-04 10:23:11 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 40.00x 2025-12-04T10:23:12.3550748Z (EngineCore_DP0 pid=29872) INFO 12-04 10:23:12 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:23:12.3571464Z INFO 12-04 10:23:12 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:23:12.8173589Z PASSED 2025-12-04T10:23:12.8323187Z models/test_initialization.py::test_can_initialize_large_subset[Grok1ModelForCausalLM] Fork a new process to run a test 29942 2025-12-04T10:23:12.8333595Z Fork a new process to run a test 0 2025-12-04T10:23:12.8612078Z INFO 12-04 10:23:12 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Grok1ModelForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'hpcai-tech/grok-1'} 2025-12-04T10:23:13.0407709Z 2025-12-04T10:23:13.0409588Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:23:13.0409977Z config.json: 1.00kB [00:00, 6.14MB/s] 2025-12-04T10:23:13.1211893Z 2025-12-04T10:23:13.1212711Z configuration_grok1.py: 0.00B [00:00, ?B/s] 2025-12-04T10:23:13.1213112Z configuration_grok1.py: 2.13kB [00:00, 24.0MB/s] 2025-12-04T10:23:13.1308113Z A new version of the following files was downloaded from https://huggingface.co/hpcai-tech/grok-1: 2025-12-04T10:23:13.1308630Z - configuration_grok1.py 2025-12-04T10:23:13.1309258Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:23:19.5850979Z INFO 12-04 10:23:19 [model.py:653] Resolved architecture: Grok1ModelForCausalLM 2025-12-04T10:23:19.5851500Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:23:19.6107115Z INFO 12-04 10:23:19 [model.py:1714] Using max model len 8192 2025-12-04T10:23:19.6108490Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T10:23:19.6633093Z INFO 12-04 10:23:19 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:23:19.6996263Z 2025-12-04T10:23:19.7008326Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:23:19.7008674Z tokenizer_config.json: 1.02kB [00:00, 872kB/s] 2025-12-04T10:23:19.8318308Z 2025-12-04T10:23:19.8661785Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T10:23:19.8662126Z tokenizer.json: 9.14MB [00:00, 267MB/s] 2025-12-04T10:23:19.9586844Z 2025-12-04T10:23:19.9587636Z special_tokens_map.json: 0% 0.00/555 [00:00. This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message. 2025-12-04T10:23:20.5400731Z (EngineCore_DP0 pid=30000) INFO 12-04 10:23:20 [core.py:727] Waiting for init message from front-end. 2025-12-04T10:23:20.5438715Z (EngineCore_DP0 pid=30000) INFO 12-04 10:23:20 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251204) with config: model='hpcai-tech/grok-1', speculative_config=None, tokenizer='hpcai-tech/grok-1', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=8192, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=hpcai-tech/grok-1, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:23:21.8904526Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:21.8931634Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:21.8939151Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:21.8947835Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:21.8955596Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:21.8962890Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:21.8965984Z (EngineCore_DP0 pid=30000) INFO 12-04 10:23:21 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:23:22.1796280Z (EngineCore_DP0 pid=30000) INFO 12-04 10:23:22 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:23:22.2789771Z (EngineCore_DP0 pid=30000) INFO 12-04 10:23:22 [gpu_model_runner.py:2840] Starting to load model hpcai-tech/grok-1... 2025-12-04T10:23:22.5568310Z (EngineCore_DP0 pid=30000) INFO 12-04 10:23:22 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:23:22.6214286Z (EngineCore_DP0 pid=30000) INFO 12-04 10:23:22 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:23:23.2522564Z (EngineCore_DP0 pid=30000) INFO 12-04 10:23:23 [gpu_model_runner.py:2902] Model loading took 3.9161 GiB and 0.123657 seconds 2025-12-04T10:23:23.2526567Z (EngineCore_DP0 pid=30000) INFO 12-04 10:23:23 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:23:23.2527506Z (EngineCore_DP0 pid=30000) INFO 12-04 10:23:23 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 320.00x 2025-12-04T10:23:24.0785898Z (EngineCore_DP0 pid=30000) INFO 12-04 10:23:24 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:23:24.0802682Z INFO 12-04 10:23:24 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:23:24.5663990Z PASSED 2025-12-04T10:23:24.5813268Z models/test_initialization.py::test_can_initialize_large_subset[MiniCPMO] Fork a new process to run a test 30070 2025-12-04T10:23:24.5823951Z Fork a new process to run a test 0 2025-12-04T10:23:24.6095378Z INFO 12-04 10:23:24 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MiniCPMO', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'openbmb/MiniCPM-o-2_6'} 2025-12-04T10:23:24.7006640Z 2025-12-04T10:23:24.7008699Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:23:24.7009000Z config.json: 3.44kB [00:00, 20.2MB/s] 2025-12-04T10:23:24.7610847Z 2025-12-04T10:23:24.7612561Z configuration_minicpm.py: 0.00B [00:00, ?B/s] 2025-12-04T10:23:24.7613116Z configuration_minicpm.py: 7.55kB [00:00, 48.6MB/s] 2025-12-04T10:23:24.8031459Z 2025-12-04T10:23:24.8034248Z modeling_navit_siglip.py: 0.00B [00:00, ?B/s] 2025-12-04T10:23:24.8034598Z modeling_navit_siglip.py: 42.1kB [00:00, 166MB/s] 2025-12-04T10:23:24.8144710Z A new version of the following files was downloaded from https://huggingface.co/openbmb/MiniCPM-o-2_6: 2025-12-04T10:23:24.8145356Z - modeling_navit_siglip.py 2025-12-04T10:23:24.8146206Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:23:24.8147095Z A new version of the following files was downloaded from https://huggingface.co/openbmb/MiniCPM-o-2_6: 2025-12-04T10:23:24.8147583Z - configuration_minicpm.py 2025-12-04T10:23:24.8147827Z - modeling_navit_siglip.py 2025-12-04T10:23:24.8155896Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:23:25.0634597Z 2025-12-04T10:23:25.0635077Z preprocessor_config.json: 0% 0.00/714 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:23:34.0996506Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:34.1025050Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:34.1032177Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:34.1039588Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:34.1048031Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:34.1056062Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:34.1059720Z (EngineCore_DP0 pid=30128) INFO 12-04 10:23:34 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:23:34.3869885Z (EngineCore_DP0 pid=30128) INFO 12-04 10:23:34 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:23:34.5993389Z (EngineCore_DP0 pid=30128) 2025-12-04T10:23:34.5994642Z processing_minicpmo.py: 0.00B [00:00, ?B/s] 2025-12-04T10:23:34.5995132Z processing_minicpmo.py: 20.0kB [00:00, 78.4MB/s] 2025-12-04T10:23:34.6842193Z (EngineCore_DP0 pid=30128) 2025-12-04T10:23:34.6844005Z image_processing_minicpmv.py: 0.00B [00:00, ?B/s] 2025-12-04T10:23:34.6844522Z image_processing_minicpmv.py: 16.7kB [00:00, 93.3MB/s] 2025-12-04T10:23:34.7548338Z (EngineCore_DP0 pid=30128) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-12-04T10:23:41.0613309Z (EngineCore_DP0 pid=30128) INFO 12-04 10:23:41 [gpu_model_runner.py:2840] Starting to load model openbmb/MiniCPM-o-2_6... 2025-12-04T10:23:41.4198110Z (EngineCore_DP0 pid=30128) INFO 12-04 10:23:41 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:23:41.4436040Z (EngineCore_DP0 pid=30128) INFO 12-04 10:23:41 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:23:41.5199510Z (EngineCore_DP0 pid=30128) INFO 12-04 10:23:41 [layer.py:497] MultiHeadAttention attn_backend: _Backend.XFORMERS, use_upstream_fa: False 2025-12-04T10:23:42.1528619Z (EngineCore_DP0 pid=30128) INFO 12-04 10:23:42 [gpu_model_runner.py:2902] Model loading took 2.7360 GiB and 0.184721 seconds 2025-12-04T10:23:42.1533012Z (EngineCore_DP0 pid=30128) INFO 12-04 10:23:42 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:23:42.1534137Z (EngineCore_DP0 pid=30128) INFO 12-04 10:23:42 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 160.00x 2025-12-04T10:23:43.4517962Z (EngineCore_DP0 pid=30128) INFO 12-04 10:23:43 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:23:43.4533943Z INFO 12-04 10:23:43 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:23:43.9481589Z PASSED 2025-12-04T10:23:43.9630506Z models/test_initialization.py::test_can_initialize_large_subset[MiniCPM3ForCausalLM] Fork a new process to run a test 30214 2025-12-04T10:23:43.9641153Z Fork a new process to run a test 0 2025-12-04T10:23:43.9911746Z INFO 12-04 10:23:43 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MiniCPM3ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'openbmb/MiniCPM3-4B'} 2025-12-04T10:23:44.1575372Z 2025-12-04T10:23:44.1577365Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:23:44.1577756Z config.json: 1.93kB [00:00, 10.4MB/s] 2025-12-04T10:23:44.2268977Z 2025-12-04T10:23:44.2270773Z configuration_minicpm.py: 0.00B [00:00, ?B/s] 2025-12-04T10:23:44.2271349Z configuration_minicpm.py: 9.23kB [00:00, 60.2MB/s] 2025-12-04T10:23:44.2362056Z A new version of the following files was downloaded from https://huggingface.co/openbmb/MiniCPM3-4B: 2025-12-04T10:23:44.2362587Z - configuration_minicpm.py 2025-12-04T10:23:44.2363222Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-12-04T10:23:44.2709181Z INFO 12-04 10:23:44 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-12-04T10:23:50.6789074Z INFO 12-04 10:23:50 [model.py:653] Resolved architecture: MiniCPM3ForCausalLM 2025-12-04T10:23:50.6789604Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:23:50.7045478Z INFO 12-04 10:23:50 [model.py:1714] Using max model len 32768 2025-12-04T10:23:50.7046954Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-12-04T10:23:50.7813392Z INFO 12-04 10:23:50 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-12-04T10:23:50.8477211Z 2025-12-04T10:23:50.8479180Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:23:50.8479701Z tokenizer_config.json: 10.4kB [00:00, 63.2MB/s] 2025-12-04T10:23:50.9745555Z 2025-12-04T10:23:51.3130153Z tokenizer.model: 0% 0.00/1.18M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:23:53.0311860Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:53.0339952Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:53.0346851Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:53.0354890Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:53.0362774Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:53.0370781Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:53.0374163Z (EngineCore_DP0 pid=30292) INFO 12-04 10:23:53 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:23:53.3235252Z (EngineCore_DP0 pid=30292) INFO 12-04 10:23:53 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:23:53.3954511Z (EngineCore_DP0 pid=30292) INFO 12-04 10:23:53 [gpu_model_runner.py:2840] Starting to load model openbmb/MiniCPM3-4B... 2025-12-04T10:23:53.6760109Z (EngineCore_DP0 pid=30292) INFO 12-04 10:23:53 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:23:53.7431571Z (EngineCore_DP0 pid=30292) INFO 12-04 10:23:53 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:23:54.3554337Z (EngineCore_DP0 pid=30292) INFO 12-04 10:23:54 [gpu_model_runner.py:2902] Model loading took 0.4724 GiB and 0.127249 seconds 2025-12-04T10:23:54.3559185Z (EngineCore_DP0 pid=30292) INFO 12-04 10:23:54 [kv_cache_utils.py:1199] GPU KV cache size: 699,040 tokens 2025-12-04T10:23:54.3559903Z (EngineCore_DP0 pid=30292) INFO 12-04 10:23:54 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 21.33x 2025-12-04T10:23:54.9726076Z (EngineCore_DP0 pid=30292) INFO 12-04 10:23:54 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:23:54.9749709Z INFO 12-04 10:23:54 [llm.py:337] Supported tasks: ['generate'] 2025-12-04T10:23:55.4338207Z PASSED 2025-12-04T10:23:55.4487808Z models/test_initialization.py::test_can_initialize_large_subset[RobertaModel] Fork a new process to run a test 30362 2025-12-04T10:23:55.4497400Z Fork a new process to run a test 0 2025-12-04T10:23:55.4769838Z INFO 12-04 10:23:55 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='RobertaModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'sentence-transformers/stsb-roberta-base-v2'} 2025-12-04T10:23:55.6168474Z 2025-12-04T10:23:55.6169463Z config.json: 0% 0.00/675 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 72, 'local_cache_dir': None} 2025-12-04T10:23:58.1973945Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:58.2004471Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:58.2011327Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:58.2018948Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:58.2026875Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:58.2034065Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:23:58.2036960Z (EngineCore_DP0 pid=30370) INFO 12-04 10:23:58 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:23:58.4915207Z (EngineCore_DP0 pid=30370) INFO 12-04 10:23:58 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:23:58.5051317Z (EngineCore_DP0 pid=30370) INFO 12-04 10:23:58 [gpu_model_runner.py:2840] Starting to load model sentence-transformers/stsb-roberta-base-v2... 2025-12-04T10:23:58.7762842Z (EngineCore_DP0 pid=30370) INFO 12-04 10:23:58 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:23:58.7818714Z (EngineCore_DP0 pid=30370) INFO 12-04 10:23:58 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:23:59.3954154Z (EngineCore_DP0 pid=30370) INFO 12-04 10:23:59 [gpu_model_runner.py:2902] Model loading took 0.0862 GiB and 0.067128 seconds 2025-12-04T10:23:59.5709457Z (EngineCore_DP0 pid=30370) INFO 12-04 10:23:59 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T10:23:59.5737353Z (EngineCore_DP0 pid=30370) INFO 12-04 10:23:59 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T10:23:59.8562255Z (EngineCore_DP0 pid=30370) INFO 12-04 10:23:59 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:23:59.8576301Z INFO 12-04 10:23:59 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-12-04T10:24:00.2875302Z PASSED 2025-12-04T10:24:00.3024295Z models/test_initialization.py::test_can_initialize_large_subset[BertForTokenClassification] Fork a new process to run a test 30440 2025-12-04T10:24:00.3035107Z Fork a new process to run a test 0 2025-12-04T10:24:00.3313868Z INFO 12-04 10:24:00 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='BertForTokenClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'boltuix/NeuroBERT-NER'} 2025-12-04T10:24:00.4511760Z 2025-12-04T10:24:00.4514239Z config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:24:00.4514868Z config.json: 1.84kB [00:00, 10.2MB/s] 2025-12-04T10:24:06.8688280Z INFO 12-04 10:24:06 [model.py:915] Resolved `--runner auto` to `--runner pooling`. Pass the value explicitly to silence this message. 2025-12-04T10:24:06.8689102Z INFO 12-04 10:24:06 [model.py:653] Resolved architecture: BertForTokenClassification 2025-12-04T10:24:06.8689825Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-12-04T10:24:06.8942106Z INFO 12-04 10:24:06 [model.py:1939] Downcasting torch.float32 to torch.float16. 2025-12-04T10:24:06.9260182Z 2025-12-04T10:24:06.9261745Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-12-04T10:24:06.9262128Z tokenizer_config.json: 1.30kB [00:00, 11.4MB/s] 2025-12-04T10:24:06.9269783Z INFO 12-04 10:24:06 [model.py:1714] Using max model len 512 2025-12-04T10:24:06.9534320Z INFO 12-04 10:24:06 [arg_utils.py:1725] (Disabling) chunked prefill by default 2025-12-04T10:24:06.9534826Z INFO 12-04 10:24:06 [arg_utils.py:1728] (Disabling) prefix caching by default 2025-12-04T10:24:07.0947500Z INFO 12-04 10:24:07 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T10:24:07.1969748Z 2025-12-04T10:24:07.1996764Z vocab.txt: 0.00B [00:00, ?B/s] 2025-12-04T10:24:07.1997082Z vocab.txt: 232kB [00:00, 85.9MB/s] 2025-12-04T10:24:07.2378517Z 2025-12-04T10:24:07.2412276Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-12-04T10:24:07.2412763Z tokenizer.json: 711kB [00:00, 208MB/s] 2025-12-04T10:24:07.3107430Z 2025-12-04T10:24:07.3108615Z special_tokens_map.json: 0% 0.00/695 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:24:08.7965243Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:08.7996038Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:08.8003289Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:08.8010553Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:08.8018796Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:08.8026207Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:08.8028924Z (EngineCore_DP0 pid=30499) INFO 12-04 10:24:08 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:24:09.0877050Z (EngineCore_DP0 pid=30499) INFO 12-04 10:24:09 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:24:09.0983394Z (EngineCore_DP0 pid=30499) INFO 12-04 10:24:09 [gpu_model_runner.py:2840] Starting to load model boltuix/NeuroBERT-NER... 2025-12-04T10:24:09.3686454Z (EngineCore_DP0 pid=30499) INFO 12-04 10:24:09 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:24:09.3732354Z (EngineCore_DP0 pid=30499) INFO 12-04 10:24:09 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:24:09.9825524Z (EngineCore_DP0 pid=30499) INFO 12-04 10:24:09 [gpu_model_runner.py:2902] Model loading took 0.0163 GiB and 0.064962 seconds 2025-12-04T10:24:10.0977001Z (EngineCore_DP0 pid=30499) INFO 12-04 10:24:10 [core.py:142] Disabling chunked prefill for model without KVCache 2025-12-04T10:24:10.1003783Z (EngineCore_DP0 pid=30499) INFO 12-04 10:24:10 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-12-04T10:24:10.3845750Z (EngineCore_DP0 pid=30499) INFO 12-04 10:24:10 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:24:10.3861589Z INFO 12-04 10:24:10 [llm.py:337] Supported tasks: ['encode'] 2025-12-04T10:24:10.8102979Z PASSED 2025-12-04T10:24:10.8251073Z models/test_initialization.py::test_can_initialize_large_subset[Ovis] Fork a new process to run a test 30569 2025-12-04T10:24:10.8260832Z Fork a new process to run a test 0 2025-12-04T10:24:10.8264671Z `transformers==4.56.2` installed, but `transformers<=4.53` is required to run this model. Reason: HF model is not compatible 2025-12-04T10:24:11.1118682Z PASSED 2025-12-04T10:24:11.1268060Z models/test_initialization.py::test_implicit_converted_models[GemmaForSequenceClassification] Fork a new process to run a test 30570 2025-12-04T10:24:11.1278427Z Fork a new process to run a test 0 2025-12-04T10:24:11.1554729Z INFO 12-04 10:24:11 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GemmaForSequenceClassification', exist_overrides={'architectures': ['GemmaForSequenceClassification'], 'classifier_from_token': ['Yes'], 'method': 'no_post_processing'}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'BAAI/bge-reranker-v2-gemma'} 2025-12-04T10:24:11.2621229Z 2025-12-04T10:24:11.2622064Z config.json: 0% 0.00/659 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:24:14.5690507Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:14.5718045Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:14.5724702Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:14.5732137Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:14.5739901Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:14.5747242Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:14.5749417Z (EngineCore_DP0 pid=30599) INFO 12-04 10:24:14 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:24:14.8602656Z (EngineCore_DP0 pid=30599) INFO 12-04 10:24:14 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:24:14.9000857Z (EngineCore_DP0 pid=30599) INFO 12-04 10:24:14 [gpu_model_runner.py:2840] Starting to load model BAAI/bge-reranker-v2-gemma... 2025-12-04T10:24:15.1712606Z (EngineCore_DP0 pid=30599) INFO 12-04 10:24:15 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:24:15.2398017Z (EngineCore_DP0 pid=30599) INFO 12-04 10:24:15 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:24:15.2408047Z (EngineCore_DP0 pid=30599) WARNING 12-04 10:24:15 [gemma.py:67] Gemma's activation function was incorrectly set to exact GeLU in the config JSON file when it was initially released. Changing the activation function to approximate GeLU (`gelu_pytorch_tanh`). If you want to use the legacy `gelu`, edit the config JSON to set `hidden_activation=gelu` instead of `hidden_act`. See https://github.com/huggingface/transformers/pull/29402 for more details. 2025-12-04T10:24:15.8393696Z (EngineCore_DP0 pid=30599) INFO 12-04 10:24:15 [gpu_model_runner.py:2902] Model loading took 1.1856 GiB and 0.129531 seconds 2025-12-04T10:24:15.8397377Z (EngineCore_DP0 pid=30599) INFO 12-04 10:24:15 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T10:24:15.8398109Z (EngineCore_DP0 pid=30599) INFO 12-04 10:24:15 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 1280.00x 2025-12-04T10:24:17.9499677Z (EngineCore_DP0 pid=30599) INFO 12-04 10:24:17 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:24:17.9514196Z INFO 12-04 10:24:17 [llm.py:337] Supported tasks: ['classify', 'score'] 2025-12-04T10:24:18.4341605Z PASSED 2025-12-04T10:24:18.4490883Z models/test_initialization.py::test_implicit_converted_models[LlamaForSequenceClassification] Fork a new process to run a test 30669 2025-12-04T10:24:18.4501784Z Fork a new process to run a test 0 2025-12-04T10:24:18.4776647Z INFO 12-04 10:24:18 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='LlamaForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Skywork/Skywork-Reward-V2-Llama-3.2-1B'} 2025-12-04T10:24:21.0821159Z 2025-12-04T10:24:21.0822433Z config.json: 0% 0.00/991 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:24:24.1269856Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:24.1297095Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:24.1304579Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:24.1312540Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:24.1320505Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:24.1328833Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:24.1331868Z (EngineCore_DP0 pid=30697) INFO 12-04 10:24:24 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:24:24.4244297Z (EngineCore_DP0 pid=30697) INFO 12-04 10:24:24 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:24:24.4845639Z (EngineCore_DP0 pid=30697) INFO 12-04 10:24:24 [gpu_model_runner.py:2840] Starting to load model Skywork/Skywork-Reward-V2-Llama-3.2-1B... 2025-12-04T10:24:24.7669179Z (EngineCore_DP0 pid=30697) INFO 12-04 10:24:24 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:24:24.8728599Z (EngineCore_DP0 pid=30697) INFO 12-04 10:24:24 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:24:25.4868343Z (EngineCore_DP0 pid=30697) INFO 12-04 10:24:25 [gpu_model_runner.py:2902] Model loading took 0.6192 GiB and 0.167051 seconds 2025-12-04T10:24:25.4873035Z (EngineCore_DP0 pid=30697) INFO 12-04 10:24:25 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-12-04T10:24:25.4874117Z (EngineCore_DP0 pid=30697) INFO 12-04 10:24:25 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 40.00x 2025-12-04T10:24:26.8065753Z (EngineCore_DP0 pid=30697) INFO 12-04 10:24:26 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:24:26.8080682Z INFO 12-04 10:24:26 [llm.py:337] Supported tasks: ['classify', 'score'] 2025-12-04T10:24:27.3035382Z PASSED 2025-12-04T10:24:27.3184426Z models/test_initialization.py::test_implicit_converted_models[Qwen2ForSequenceClassification] Fork a new process to run a test 30767 2025-12-04T10:24:27.3194502Z Fork a new process to run a test 0 2025-12-04T10:24:27.3467663Z INFO 12-04 10:24:27 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2ForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'jason9693/Qwen2.5-1.5B-apeach'} 2025-12-04T10:24:27.4688676Z 2025-12-04T10:24:27.4689652Z config.json: 0% 0.00/950 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:24:30.4633173Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:30.4660112Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:30.4667282Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:30.4675340Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:30.4683160Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:30.4691092Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:30.4694428Z (EngineCore_DP0 pid=30797) INFO 12-04 10:24:30 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:24:30.7576778Z (EngineCore_DP0 pid=30797) INFO 12-04 10:24:30 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:24:30.8175854Z (EngineCore_DP0 pid=30797) INFO 12-04 10:24:30 [gpu_model_runner.py:2840] Starting to load model jason9693/Qwen2.5-1.5B-apeach... 2025-12-04T10:24:31.1005405Z (EngineCore_DP0 pid=30797) INFO 12-04 10:24:31 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:24:31.1704146Z (EngineCore_DP0 pid=30797) INFO 12-04 10:24:31 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:24:31.7831984Z (EngineCore_DP0 pid=30797) INFO 12-04 10:24:31 [gpu_model_runner.py:2902] Model loading took 0.5540 GiB and 0.130207 seconds 2025-12-04T10:24:31.7836414Z (EngineCore_DP0 pid=30797) INFO 12-04 10:24:31 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-12-04T10:24:31.7837380Z (EngineCore_DP0 pid=30797) INFO 12-04 10:24:31 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 80.00x 2025-12-04T10:24:33.6171315Z (EngineCore_DP0 pid=30797) INFO 12-04 10:24:33 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:24:33.6186115Z INFO 12-04 10:24:33 [llm.py:337] Supported tasks: ['classify'] 2025-12-04T10:24:34.1026672Z PASSED 2025-12-04T10:24:34.1175249Z models/test_initialization.py::test_implicit_converted_models[Qwen3ForSequenceClassification] Fork a new process to run a test 30867 2025-12-04T10:24:34.1186622Z Fork a new process to run a test 0 2025-12-04T10:24:34.1462962Z INFO 12-04 10:24:34 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen3ForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'tomaarsen/Qwen3-Reranker-0.6B-seq-cls'} 2025-12-04T10:24:34.2323833Z 2025-12-04T10:24:34.2324804Z config.json: 0% 0.00/844 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-12-04T10:24:36.8731414Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:36.8758651Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:36.8766007Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:36.8774026Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:36.8781837Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:36.8789439Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-12-04T10:24:36.8792346Z (EngineCore_DP0 pid=30895) INFO 12-04 10:24:36 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-12-04T10:24:37.1567086Z (EngineCore_DP0 pid=30895) INFO 12-04 10:24:37 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-12-04T10:24:37.1926943Z (EngineCore_DP0 pid=30895) INFO 12-04 10:24:37 [gpu_model_runner.py:2840] Starting to load model tomaarsen/Qwen3-Reranker-0.6B-seq-cls... 2025-12-04T10:24:37.4618077Z (EngineCore_DP0 pid=30895) INFO 12-04 10:24:37 [gpu_model_runner.py:2870] Loading model from scratch... 2025-12-04T10:24:37.5354721Z (EngineCore_DP0 pid=30895) INFO 12-04 10:24:37 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-12-04T10:24:38.1401108Z (EngineCore_DP0 pid=30895) INFO 12-04 10:24:38 [gpu_model_runner.py:2902] Model loading took 0.3284 GiB and 0.134689 seconds 2025-12-04T10:24:38.1404741Z (EngineCore_DP0 pid=30895) INFO 12-04 10:24:38 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-12-04T10:24:38.1405480Z (EngineCore_DP0 pid=30895) INFO 12-04 10:24:38 [kv_cache_utils.py:1204] Maximum concurrency for 40,960 tokens per request: 64.00x 2025-12-04T10:24:39.0809274Z (EngineCore_DP0 pid=30895) INFO 12-04 10:24:39 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-12-04T10:24:39.0826851Z INFO 12-04 10:24:39 [llm.py:337] Supported tasks: ['classify', 'score'] 2025-12-04T10:24:39.5537514Z PASSED 2025-12-04T10:24:39.5537700Z 2025-12-04T10:24:39.5537842Z =================================== FAILURES =================================== 2025-12-04T10:24:39.5538404Z _______________ test_can_initialize_small_subset[UltravoxModel] ________________ 2025-12-04T10:24:39.5538742Z 2025-12-04T10:24:39.5538836Z model_arch = 'UltravoxModel' 2025-12-04T10:24:39.5539335Z monkeypatch = <_pytest.monkeypatch.MonkeyPatch object at 0x7fea58a7f380> 2025-12-04T10:24:39.5539909Z 2025-12-04T10:24:39.5925279Z @pytest.mark.parametrize("model_arch", MINIMAL_MODEL_ARCH_LIST) 2025-12-04T10:24:39.5926508Z def test_can_initialize_small_subset(model_arch: str, monkeypatch: pytest.MonkeyPatch): 2025-12-04T10:24:39.5927380Z  """Test initializing small subset of supported models""" 2025-12-04T10:24:39.5928347Z if model_arch == "Lfm2ForCausalLM": 2025-12-04T10:24:39.5929260Z pytest.skip("Skipping until test supports V1-only models") 2025-12-04T10:24:39.5930021Z > can_initialize(model_arch, monkeypatch, HF_EXAMPLE_MODELS) 2025-12-04T10:24:39.5930461Z 2025-12-04T10:24:39.5930650Z models/test_initialization.py:137: 2025-12-04T10:24:39.5931141Z _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 2025-12-04T10:24:39.5931523Z utils.py:938: in wrapper 2025-12-04T10:24:39.5931901Z raise original_exception 2025-12-04T10:24:39.5932487Z utils.py:879: in wrapper 2025-12-04T10:24:39.5933009Z func(*args, **kwargs) 2025-12-04T10:24:39.5933451Z models/test_initialization.py:105: in can_initialize 2025-12-04T10:24:39.5933831Z LLM( 2025-12-04T10:24:39.5934569Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/entrypoints/llm.py:328: in __init__ 2025-12-04T10:24:39.5935416Z self.llm_engine = LLMEngine.from_engine_args( 2025-12-04T10:24:39.5936184Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/v1/engine/llm_engine.py:186: in from_engine_args 2025-12-04T10:24:39.5936767Z return cls( 2025-12-04T10:24:39.5937402Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/v1/engine/llm_engine.py:121: in __init__ 2025-12-04T10:24:39.5938005Z self.engine_core = EngineCoreClient.make_client( 2025-12-04T10:24:39.5938830Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/v1/engine/core_client.py:93: in make_client 2025-12-04T10:24:39.5939843Z return SyncMPClient(vllm_config, executor_class, log_stats) 2025-12-04T10:24:39.5940472Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/v1/engine/core_client.py:641: in __init__ 2025-12-04T10:24:39.5940982Z super().__init__( 2025-12-04T10:24:39.5941757Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/v1/engine/core_client.py:470: in __init__ 2025-12-04T10:24:39.5942458Z with launch_core_engines(vllm_config, executor_class, log_stats) as ( 2025-12-04T10:24:39.5943012Z /opt/conda/envs/py_3.12/lib/python3.12/contextlib.py:144: in __exit__ 2025-12-04T10:24:39.5943423Z next(self.gen) 2025-12-04T10:24:39.5943963Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/v1/engine/utils.py:815: in launch_core_engines 2025-12-04T10:24:39.5944741Z wait_for_engine_startup( 2025-12-04T10:24:39.5945002Z _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 2025-12-04T10:24:39.5945174Z 2025-12-04T10:24:39.5945359Z > raise RuntimeError( 2025-12-04T10:24:39.5945826Z "Engine core initialization failed. " 2025-12-04T10:24:39.5946454Z "See root cause above. " 2025-12-04T10:24:39.5947029Z f"Failed core proc(s): {finished}" 2025-12-04T10:24:39.5947436Z ) 2025-12-04T10:24:39.5947859Z E RuntimeError: Engine core initialization failed. See root cause above. Failed core proc(s): {} 2025-12-04T10:24:39.5948194Z 2025-12-04T10:24:39.5948518Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/v1/engine/utils.py:872: RuntimeError 2025-12-04T10:24:39.5949034Z =============================== warnings summary =============================== 2025-12-04T10:24:39.5962648Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/jit/_script.py:362: 14 warnings 2025-12-04T10:24:39.5963596Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/jit/_script.py:362: DeprecationWarning: `torch.jit.script_method` is deprecated. Please switch to `torch.compile` or `torch.export`. 2025-12-04T10:24:39.5964251Z warnings.warn( 2025-12-04T10:24:39.5964353Z 2025-12-04T10:24:39.5964444Z :488 2025-12-04T10:24:39.5964987Z :488: DeprecationWarning: builtin type SwigPyPacked has no __module__ attribute 2025-12-04T10:24:39.5965496Z 2025-12-04T10:24:39.5965576Z :488 2025-12-04T10:24:39.5966000Z :488: DeprecationWarning: builtin type SwigPyObject has no __module__ attribute 2025-12-04T10:24:39.5966356Z 2025-12-04T10:24:39.5967100Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305 2025-12-04T10:24:39.5968415Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305: DeprecationWarning: jsonschema.exceptions.RefResolutionError is deprecated as of version 4.18.0. If you wish to catch potential reference resolution errors, directly catch referencing.exceptions.Unresolvable. 2025-12-04T10:24:39.5969503Z ref_error: type[Exception] = jsonschema.RefResolutionError, 2025-12-04T10:24:39.5969709Z 2025-12-04T10:24:39.5969821Z tests/models/test_initialization.py: 224 warnings 2025-12-04T10:24:39.5970503Z /var/lib/jenkins/workspace/vllm/tests/utils.py:872: DeprecationWarning: This process (pid=3920) is multi-threaded, use of fork() may lead to deadlocks in the child. 2025-12-04T10:24:39.5971363Z pid = os.fork() 2025-12-04T10:24:39.5971472Z 2025-12-04T10:24:39.5971661Z -- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html 2025-12-04T10:24:39.5972872Z =========================== short test summary info ============================ 2025-12-04T10:24:39.5974016Z FAILED models/test_initialization.py::test_can_initialize_small_subset[UltravoxModel] - RuntimeError: Engine core initialization failed. See root cause above. Failed core proc(s): {} 2025-12-04T10:24:39.5974983Z ===== 1 failed, 223 passed, 1 skipped, 241 warnings in 2685.70s (0:44:45) ====== 2025-12-04T10:24:39.8452672Z sys:1: DeprecationWarning: builtin type swigvarlink has no __module__ attribute 2025-12-04T10:24:41.0761262Z 2025-12-04 10:24:41,075 [INFO] cli.lib.core.vllm.lib: Finish running step: pytest -v -s models/test_initialization.py 2025-12-04T10:24:41.1414536Z 2025-12-04 10:24:41,085 [ERROR] cli.lib.core.vllm.lib: Failed tests: ['pytest -v -s models/test_initialization.py'] 2025-12-04T10:24:41.1660751Z 2025-12-04 10:24:41,165 [INFO] cli.lib.core.vllm.vllm_test: Double check installed packages 2025-12-04T10:24:41.3924387Z 2025-12-04 10:24:41,391 [INFO] cli.lib.common.pip_helper: torch already exist with version: 2.10.0a0+gitffd9b0f 2025-12-04T10:24:41.3952414Z 2025-12-04 10:24:41,395 [INFO] cli.lib.common.pip_helper: xformers already exist with version: 0.0.34+3f91ad6f.d20251204 2025-12-04T10:24:41.4135490Z 2025-12-04 10:24:41,413 [INFO] cli.lib.common.pip_helper: torchvision already exist with version: 0.25.0a0+617079d 2025-12-04T10:24:41.4214910Z 2025-12-04 10:24:41,421 [INFO] cli.lib.common.pip_helper: torchaudio already exist with version: 2.10.0a0+e90a398 2025-12-04T10:24:41.4316724Z 2025-12-04 10:24:41,431 [INFO] cli.lib.common.pip_helper: vllm already exist with version: 0.11.0rc2.dev389+ge51928192.d20251204 2025-12-04T10:24:41.4318106Z 2025-12-04 10:24:41,431 [INFO] cli.lib.core.vllm.vllm_test: Done. checked installed packages 2025-12-04T10:24:41.4342724Z Traceback (most recent call last): 2025-12-04T10:24:41.4343414Z File "", line 198, in _run_module_as_main 2025-12-04T10:24:41.4344077Z File "", line 88, in _run_code 2025-12-04T10:24:41.4349341Z File "/var/lib/jenkins/workspace/.ci/lumen_cli/cli/run.py", line 40, in 2025-12-04T10:24:41.4397193Z main() 2025-12-04T10:24:41.4408957Z File "/var/lib/jenkins/workspace/.ci/lumen_cli/cli/run.py", line 34, in main 2025-12-04T10:24:41.4410262Z args.func(args) 2025-12-04T10:24:41.4427054Z File "/var/lib/jenkins/workspace/.ci/lumen_cli/cli/lib/common/cli_helper.py", line 67, in 2025-12-04T10:24:41.4432849Z func=lambda args, cls=spec["runner"]: cls(args).run(), 2025-12-04T10:24:41.4439547Z ^^^^^^^^^^^^^^^ 2025-12-04T10:24:41.4446080Z File "/var/lib/jenkins/workspace/.ci/lumen_cli/cli/lib/core/vllm/vllm_test.py", line 124, in run 2025-12-04T10:24:41.4453054Z run_test_plan( 2025-12-04T10:24:41.4453562Z File "/var/lib/jenkins/workspace/.ci/lumen_cli/cli/lib/core/vllm/lib.py", line 262, in run_test_plan 2025-12-04T10:24:41.4460891Z raise RuntimeError(f"{len(failures)} pytest runs failed: {failures}") 2025-12-04T10:24:41.4489360Z RuntimeError: 1 pytest runs failed: ['pytest -v -s models/test_initialization.py'] 2025-12-04T10:24:42.6369618Z + sccache_epilogue 2025-12-04T10:24:42.6391330Z + echo '::group::Sccache Compilation Log' 2025-12-04T10:24:42.6392897Z ##[group]Sccache Compilation Log 2025-12-04T10:24:42.6393518Z + echo '=================== sccache compilation log ===================' 2025-12-04T10:24:42.6394217Z =================== sccache compilation log =================== 2025-12-04T10:24:42.6399626Z + python /var/lib/jenkins/workspace/.ci/pytorch/print_sccache_log.py /var/lib/jenkins/sccache_error.log 2025-12-04T10:24:42.6696086Z + echo '=========== If your build fails, please take a look at the log above for possible reasons ===========' 2025-12-04T10:24:42.6696849Z =========== If your build fails, please take a look at the log above for possible reasons =========== 2025-12-04T10:24:42.6697306Z + sccache --show-stats 2025-12-04T10:24:42.6730446Z Compile requests 19 2025-12-04T10:24:42.6730860Z Compile requests executed 4 2025-12-04T10:24:42.6731150Z Cache hits 4 2025-12-04T10:24:42.6731421Z Cache hits (C/C++) 4 2025-12-04T10:24:42.6731681Z Cache misses 0 2025-12-04T10:24:42.6731950Z Cache hits rate 100.00 % 2025-12-04T10:24:42.6732220Z Cache hits rate (C/C++) 100.00 % 2025-12-04T10:24:42.6732487Z Cache timeouts 0 2025-12-04T10:24:42.6732748Z Cache read errors 0 2025-12-04T10:24:42.6733006Z Forced recaches 0 2025-12-04T10:24:42.6733268Z Cache write errors 0 2025-12-04T10:24:42.6733524Z Cache errors 0 2025-12-04T10:24:42.6733789Z Compilations 0 2025-12-04T10:24:42.6734066Z Compilation failures 0 2025-12-04T10:24:42.6734340Z Non-cacheable compilations 0 2025-12-04T10:24:42.6734649Z Non-cacheable calls 0 2025-12-04T10:24:42.6735134Z Non-compilation calls 15 2025-12-04T10:24:42.6735428Z Unsupported compiler calls 0 2025-12-04T10:24:42.6735699Z Average cache write 0.000 s 2025-12-04T10:24:42.6736140Z Average compiler 0.000 s 2025-12-04T10:24:42.6736430Z Average cache read hit 0.079 s 2025-12-04T10:24:42.6736706Z Failed distributed compilations 0 2025-12-04T10:24:42.6737115Z Cache location s3, name: ossci-compiler-cache-circleci-v2, prefix: / 2025-12-04T10:24:42.6737538Z Version (client) 0.10.0 2025-12-04T10:24:42.6737819Z + sccache --stop-server 2025-12-04T10:24:42.6758931Z Stopping sccache server... 2025-12-04T10:24:42.6762069Z Compile requests 19 2025-12-04T10:24:42.6762544Z Compile requests executed 4 2025-12-04T10:24:42.6762826Z Cache hits 4 2025-12-04T10:24:42.6763093Z Cache hits (C/C++) 4 2025-12-04T10:24:42.6763352Z Cache misses 0 2025-12-04T10:24:42.6763611Z Cache hits rate 100.00 % 2025-12-04T10:24:42.6763875Z Cache hits rate (C/C++) 100.00 % 2025-12-04T10:24:42.6764155Z Cache timeouts 0 2025-12-04T10:24:42.6764424Z Cache read errors 0 2025-12-04T10:24:42.6764685Z Forced recaches 0 2025-12-04T10:24:42.6764949Z Cache write errors 0 2025-12-04T10:24:42.6765205Z Cache errors 0 2025-12-04T10:24:42.6765454Z Compilations 0 2025-12-04T10:24:42.6766095Z Compilation failures 0 2025-12-04T10:24:42.6766549Z Non-cacheable compilations 0 2025-12-04T10:24:42.6767054Z Non-cacheable calls 0 2025-12-04T10:24:42.6767358Z Non-compilation calls 15 2025-12-04T10:24:42.6767631Z Unsupported compiler calls 0 2025-12-04T10:24:42.6767903Z Average cache write 0.000 s 2025-12-04T10:24:42.6768180Z Average compiler 0.000 s 2025-12-04T10:24:42.6768468Z Average cache read hit 0.079 s 2025-12-04T10:24:42.6768737Z Failed distributed compilations 0 2025-12-04T10:24:42.6769050Z Cache location s3, name: ossci-compiler-cache-circleci-v2, prefix: / 2025-12-04T10:24:42.6769372Z Version (client) 0.10.0 2025-12-04T10:24:42.6769590Z + echo ::endgroup:: 2025-12-04T10:24:42.6769933Z ##[endgroup] 2025-12-04T10:24:42.6770084Z + cleanup_workspace 2025-12-04T10:24:42.6770509Z + echo 'sudo may print the following warning message that can be ignored. The chown command will still run.' 2025-12-04T10:24:42.6771359Z sudo may print the following warning message that can be ignored. The chown command will still run. 2025-12-04T10:24:42.6771825Z + echo ' sudo: setrlimit(RLIMIT_STACK): Operation not permitted' 2025-12-04T10:24:42.6772145Z sudo: setrlimit(RLIMIT_STACK): Operation not permitted 2025-12-04T10:24:42.6775495Z + echo 'For more details refer to https://github.com/sudo-project/sudo/issues/42' 2025-12-04T10:24:42.6776007Z For more details refer to https://github.com/sudo-project/sudo/issues/42 2025-12-04T10:24:42.6776349Z + sudo chown -R 1000 /var/lib/jenkins/workspace 2025-12-04T10:24:45.1575097Z ##[error]Process completed with exit code 1. 2025-12-04T10:24:45.1655110Z Prepare all required actions 2025-12-04T10:24:45.1655906Z Getting action download info 2025-12-04T10:24:45.3194691Z ##[group]Run ./.github/actions/pytest-cache-upload 2025-12-04T10:24:45.3194942Z with: 2025-12-04T10:24:45.3195106Z cache_dir: .pytest_cache 2025-12-04T10:24:45.3195298Z shard: 1 2025-12-04T10:24:45.3195469Z sha: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T10:24:45.3195712Z test_config: vllm_basic_models_test 2025-12-04T10:24:45.3195990Z job_identifier: vllm-test_linux-jammy-cuda12.8-py3.12-gcc11 2025-12-04T10:24:45.3196260Z env: 2025-12-04T10:24:45.3196405Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:24:45.3196591Z HAS_NVIDIA_GPU: true 2025-12-04T10:24:45.3196817Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:24:45.3197196Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:24:45.3197675Z ##[endgroup] 2025-12-04T10:24:45.3225150Z ##[group]Run nick-fields/retry@v3.0.0 2025-12-04T10:24:45.3225398Z with: 2025-12-04T10:24:45.3225546Z shell: bash 2025-12-04T10:24:45.3225702Z timeout_minutes: 5 2025-12-04T10:24:45.3225870Z max_attempts: 5 2025-12-04T10:24:45.3226031Z retry_wait_seconds: 30 2025-12-04T10:24:45.3226267Z command: set -eu python3 -m pip install boto3==1.35.42 2025-12-04T10:24:45.3226542Z polling_interval_seconds: 1 2025-12-04T10:24:45.3226741Z warning_on_retry: true 2025-12-04T10:24:45.3226919Z continue_on_error: false 2025-12-04T10:24:45.3227093Z env: 2025-12-04T10:24:45.3227237Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:24:45.3227413Z HAS_NVIDIA_GPU: true 2025-12-04T10:24:45.3227628Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:24:45.3228007Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:24:45.3228356Z ##[endgroup] 2025-12-04T10:24:45.9507937Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T10:24:47.0174567Z Collecting boto3==1.35.42 2025-12-04T10:24:47.0336750Z Downloading boto3-1.35.42-py3-none-any.whl (139 kB) 2025-12-04T10:24:47.1041238Z Collecting s3transfer<0.11.0,>=0.10.0 2025-12-04T10:24:47.1074045Z Downloading s3transfer-0.10.4-py3-none-any.whl (83 kB) 2025-12-04T10:24:47.1145480Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.42) (0.10.0) 2025-12-04T10:24:48.1991976Z Collecting botocore<1.36.0,>=1.35.42 2025-12-04T10:24:48.2034669Z Downloading botocore-1.35.99-py3-none-any.whl (13.3 MB) 2025-12-04T10:24:48.3362532Z Requirement already satisfied: python-dateutil<3.0.0,>=2.1 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.42->boto3==1.35.42) (2.8.1) 2025-12-04T10:24:48.3369462Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.42->boto3==1.35.42) (1.25.10) 2025-12-04T10:24:48.5172443Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil<3.0.0,>=2.1->botocore<1.36.0,>=1.35.42->boto3==1.35.42) (1.15.0) 2025-12-04T10:24:48.5954400Z Installing collected packages: botocore, s3transfer, boto3 2025-12-04T10:24:49.1726495Z Successfully installed boto3-1.35.42 botocore-1.35.99 s3transfer-0.10.4 2025-12-04T10:24:49.3939429Z Command completed after 1 attempt(s). 2025-12-04T10:24:49.4001955Z ##[group]Run python3 .github/scripts/pytest_cache.py \ 2025-12-04T10:24:49.4002298Z python3 .github/scripts/pytest_cache.py \ 2025-12-04T10:24:49.4002549Z  --upload \ 2025-12-04T10:24:49.4002770Z  --cache_dir "$GITHUB_WORKSPACE/$CACHE_DIR" \ 2025-12-04T10:24:49.4003056Z  --pr_identifier "$GITHUB_REF" \ 2025-12-04T10:24:49.4003307Z  --job_identifier "$JOB_IDENTIFIER" \ 2025-12-04T10:24:49.4003541Z  --sha "$SHA" \ 2025-12-04T10:24:49.4003741Z  --test_config "$TEST_CONFIG" \ 2025-12-04T10:24:49.4003969Z  --shard "$SHARD" \ 2025-12-04T10:24:49.4004386Z  --repo "$REPO" \ 2025-12-04T10:24:49.4004622Z  --temp_dir "$RUNNER_TEMP" \ 2025-12-04T10:24:49.4017231Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:24:49.4017502Z env: 2025-12-04T10:24:49.4017658Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:24:49.4017844Z HAS_NVIDIA_GPU: true 2025-12-04T10:24:49.4018069Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:24:49.4018461Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:24:49.4018817Z CACHE_DIR: .pytest_cache 2025-12-04T10:24:49.4019069Z JOB_IDENTIFIER: vllm-test_linux-jammy-cuda12.8-py3.12-gcc11 2025-12-04T10:24:49.4019533Z SHA: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T10:24:49.4019786Z TEST_CONFIG: vllm_basic_models_test 2025-12-04T10:24:49.4019995Z SHARD: 1 2025-12-04T10:24:49.4020143Z REPO: pytorch/pytorch 2025-12-04T10:24:49.4020443Z ##[endgroup] 2025-12-04T10:24:49.5385067Z PR identifier for `refs/heads/main` is `96e092540d6b3c4076e3d2bc6f1f9013` 2025-12-04T10:24:49.5387457Z Uploading cache with args Namespace(upload=True, download=False, cache_dir='/home/ec2-user/actions-runner/_work/pytorch/pytorch/.pytest_cache', pr_identifier='refs/heads/main', job_identifier='vllm-test_linux-jammy-cuda12.8-py3.12-gcc11', sha='ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32', test_config='vllm_basic_models_test', shard='1', repo='pytorch/pytorch', temp_dir='/home/ec2-user/actions-runner/_work/_temp', bucket=None) 2025-12-04T10:24:49.5390032Z The pytest cache dir `/home/ec2-user/actions-runner/_work/pytorch/pytorch/.pytest_cache` does not exist. Skipping upload 2025-12-04T10:24:49.5575924Z ##[group]Run cat test/**/*_toprint.log || true 2025-12-04T10:24:49.5576261Z cat test/**/*_toprint.log || true 2025-12-04T10:24:49.5584520Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:24:49.5584793Z env: 2025-12-04T10:24:49.5584960Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:24:49.5585151Z HAS_NVIDIA_GPU: true 2025-12-04T10:24:49.5585384Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:24:49.5585782Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:24:49.5586130Z ##[endgroup] 2025-12-04T10:24:49.5681055Z cat: 'test/**/*_toprint.log': No such file or directory 2025-12-04T10:24:49.5712575Z ##[group]Run kill "$MONITOR_SCRIPT_PID" 2025-12-04T10:24:49.5712879Z kill "$MONITOR_SCRIPT_PID" 2025-12-04T10:24:49.5719963Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:24:49.5720241Z env: 2025-12-04T10:24:49.5720398Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:24:49.5720582Z HAS_NVIDIA_GPU: true 2025-12-04T10:24:49.5720807Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:24:49.5721197Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:24:49.5721559Z MONITOR_SCRIPT_PID: 61327 2025-12-04T10:24:49.5721749Z ##[endgroup] 2025-12-04T10:24:49.5748776Z /home/ec2-user/actions-runner/_work/_temp/111093d9-b53a-4834-a56d-eb3ce63f5c63.sh: line 1: kill: (61327) - No such process 2025-12-04T10:24:49.5751597Z ##[error]Process completed with exit code 1. 2025-12-04T10:24:49.5843604Z Prepare all required actions 2025-12-04T10:24:49.5843964Z Getting action download info 2025-12-04T10:24:49.7417350Z Download action repository 'seemethere/upload-artifact-s3@v5' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-12-04T10:24:49.9877234Z Download action repository 'actions/upload-artifact@v4' (SHA:ea165f8d65b6e75b540449e92b4886f43607fa02) 2025-12-04T10:24:50.5341346Z ##[group]Run ./.github/actions/upload-test-artifacts 2025-12-04T10:24:50.5341734Z with: 2025-12-04T10:24:50.5342242Z file-suffix: test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_57118543727 2025-12-04T10:24:50.5342862Z s3-bucket: gha-artifacts 2025-12-04T10:24:50.5343141Z env: 2025-12-04T10:24:50.5343378Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:24:50.5343661Z HAS_NVIDIA_GPU: true 2025-12-04T10:24:50.5343995Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:24:50.5344647Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:24:50.5345192Z ##[endgroup] 2025-12-04T10:24:50.5372851Z ##[group]Run # Remove any previous test jsons if they exist 2025-12-04T10:24:50.5373357Z # Remove any previous test jsons if they exist 2025-12-04T10:24:50.5373762Z rm -f test-jsons-*.zip 2025-12-04T10:24:50.5374232Z zip -r "test-jsons-${FILE_SUFFIX}.zip" test/test-reports -i '*.json' 2025-12-04T10:24:50.5384208Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:24:50.5384616Z env: 2025-12-04T10:24:50.5384844Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:24:50.5385128Z HAS_NVIDIA_GPU: true 2025-12-04T10:24:50.5385458Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:24:50.5386211Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:24:50.5387036Z FILE_SUFFIX: test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_57118543727 2025-12-04T10:24:50.5387635Z ##[endgroup] 2025-12-04T10:24:50.5536902Z zip warning: name not matched: test/test-reports 2025-12-04T10:24:50.5537933Z zip warning: zip file empty 2025-12-04T10:24:50.5569039Z ##[group]Run # Remove any previous test reports if they exist 2025-12-04T10:24:50.5569607Z # Remove any previous test reports if they exist 2025-12-04T10:24:50.5570064Z rm -f test-reports-*.zip 2025-12-04T10:24:50.5570654Z zip -r "test-reports-${FILE_SUFFIX}.zip" test/test-reports -i '*.xml' -i '*.csv' 2025-12-04T10:24:50.5581493Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:24:50.5581960Z env: 2025-12-04T10:24:50.5582211Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:24:50.5582537Z HAS_NVIDIA_GPU: true 2025-12-04T10:24:50.5582935Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:24:50.5583625Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:24:50.5584567Z FILE_SUFFIX: test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_57118543727 2025-12-04T10:24:50.5585243Z ##[endgroup] 2025-12-04T10:24:50.5638604Z zip warning: name not matched: test/test-reports 2025-12-04T10:24:50.5639372Z zip warning: zip file empty 2025-12-04T10:24:50.5662453Z ##[group]Run # Remove any previous usage logs if they exist 2025-12-04T10:24:50.5662789Z # Remove any previous usage logs if they exist 2025-12-04T10:24:50.5663081Z rm -f logs-*.zip 2025-12-04T10:24:50.5663502Z zip "logs-${FILE_SUFFIX}.zip" 'usage_log.txt' || true 2025-12-04T10:24:50.5663863Z zip -r "logs-${FILE_SUFFIX}.zip" test/test-reports -i '*.log' || true 2025-12-04T10:24:50.5671511Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:24:50.5671793Z env: 2025-12-04T10:24:50.5671944Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:24:50.5672136Z HAS_NVIDIA_GPU: true 2025-12-04T10:24:50.5672353Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:24:50.5672934Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:24:50.5673467Z FILE_SUFFIX: test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_57118543727 2025-12-04T10:24:50.5673850Z ##[endgroup] 2025-12-04T10:24:50.5750416Z adding: usage_log.txt (deflated 58%) 2025-12-04T10:24:50.5761809Z zip warning: name not matched: test/test-reports 2025-12-04T10:24:50.5762163Z 2025-12-04T10:24:50.5762559Z zip error: Nothing to do! (logs-test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_57118543727.zip) 2025-12-04T10:24:50.5789260Z ##[group]Run # Remove any previous debugging artifacts if they exist 2025-12-04T10:24:50.5789818Z # Remove any previous debugging artifacts if they exist 2025-12-04T10:24:50.5790271Z rm -f debug-*.zip 2025-12-04T10:24:50.5790577Z if [ -d 'test/debug' ]; then 2025-12-04T10:24:50.5790966Z  zip -r "debug-${FILE_SUFFIX}.zip" test/debug 2025-12-04T10:24:50.5791335Z fi 2025-12-04T10:24:50.5800210Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:24:50.5800630Z env: 2025-12-04T10:24:50.5800864Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:24:50.5801160Z HAS_NVIDIA_GPU: true 2025-12-04T10:24:50.5801500Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:24:50.5802119Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:24:50.5802949Z FILE_SUFFIX: test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_57118543727 2025-12-04T10:24:50.5803550Z ##[endgroup] 2025-12-04T10:24:50.5907869Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T10:24:50.5908103Z with: 2025-12-04T10:24:50.5908257Z s3-bucket: gha-artifacts 2025-12-04T10:24:50.5908600Z s3-prefix: pytorch/pytorch/19922031852/1/artifact 2025-12-04T10:24:50.5908850Z retention-days: 14 2025-12-04T10:24:50.5909026Z if-no-files-found: warn 2025-12-04T10:24:50.5909208Z path: test-jsons-*.zip 2025-12-04T10:24:50.5909395Z name: artifact 2025-12-04T10:24:50.5909556Z region: us-east-1 2025-12-04T10:24:50.5909707Z env: 2025-12-04T10:24:50.5909856Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:24:50.5910041Z HAS_NVIDIA_GPU: true 2025-12-04T10:24:50.5910251Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:24:50.5910648Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:24:50.5910997Z ##[endgroup] 2025-12-04T10:24:50.8818729Z NOTE: s3-prefix specified, ignoring name parameter 2025-12-04T10:24:50.8819263Z With the provided path, there will be 1 file uploaded 2025-12-04T10:24:50.8819684Z Uploading to s3 prefix: pytorch/pytorch/19922031852/1/artifact 2025-12-04T10:24:50.8887773Z Starting upload of test-jsons-test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_57118543727.zip 2025-12-04T10:24:50.9983665Z Finished upload of test-jsons-test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_57118543727.zip 2025-12-04T10:24:51.0211248Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T10:24:51.0211488Z with: 2025-12-04T10:24:51.0211649Z s3-bucket: gha-artifacts 2025-12-04T10:24:51.0212022Z s3-prefix: pytorch/pytorch/19922031852/1/artifact 2025-12-04T10:24:51.0212266Z retention-days: 14 2025-12-04T10:24:51.0212448Z if-no-files-found: error 2025-12-04T10:24:51.0212643Z path: test-reports-*.zip 2025-12-04T10:24:51.0212818Z name: artifact 2025-12-04T10:24:51.0212978Z region: us-east-1 2025-12-04T10:24:51.0213139Z env: 2025-12-04T10:24:51.0213296Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:24:51.0213487Z HAS_NVIDIA_GPU: true 2025-12-04T10:24:51.0213721Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:24:51.0214121Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:24:51.0214481Z ##[endgroup] 2025-12-04T10:24:51.3439794Z NOTE: s3-prefix specified, ignoring name parameter 2025-12-04T10:24:51.3440478Z With the provided path, there will be 1 file uploaded 2025-12-04T10:24:51.3441497Z Uploading to s3 prefix: pytorch/pytorch/19922031852/1/artifact 2025-12-04T10:24:51.3508596Z Starting upload of test-reports-test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_57118543727.zip 2025-12-04T10:24:51.4485585Z Finished upload of test-reports-test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_57118543727.zip 2025-12-04T10:24:51.4719523Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T10:24:51.4719772Z with: 2025-12-04T10:24:51.4719936Z s3-bucket: gha-artifacts 2025-12-04T10:24:51.4720169Z s3-prefix: pytorch/pytorch/19922031852/1/artifact 2025-12-04T10:24:51.4720419Z retention-days: 14 2025-12-04T10:24:51.4720597Z if-no-files-found: ignore 2025-12-04T10:24:51.4720797Z path: logs-*.zip 2025-12-04T10:24:51.4720959Z name: artifact 2025-12-04T10:24:51.4721119Z region: us-east-1 2025-12-04T10:24:51.4721278Z env: 2025-12-04T10:24:51.4721442Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:24:51.4721648Z HAS_NVIDIA_GPU: true 2025-12-04T10:24:51.4721868Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:24:51.4722270Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:24:51.4722623Z ##[endgroup] 2025-12-04T10:24:51.7624537Z NOTE: s3-prefix specified, ignoring name parameter 2025-12-04T10:24:51.7625105Z With the provided path, there will be 1 file uploaded 2025-12-04T10:24:51.7625562Z Uploading to s3 prefix: pytorch/pytorch/19922031852/1/artifact 2025-12-04T10:24:51.7693122Z Starting upload of logs-test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_57118543727.zip 2025-12-04T10:24:51.8753012Z Finished upload of logs-test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_57118543727.zip 2025-12-04T10:24:51.8992613Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T10:24:51.8992853Z with: 2025-12-04T10:24:51.8993012Z s3-bucket: gha-artifacts 2025-12-04T10:24:51.8993259Z s3-prefix: pytorch/pytorch/19922031852/1/artifact 2025-12-04T10:24:51.8993501Z retention-days: 14 2025-12-04T10:24:51.8993695Z if-no-files-found: ignore 2025-12-04T10:24:51.8993897Z path: debug-*.zip 2025-12-04T10:24:51.8994054Z name: artifact 2025-12-04T10:24:51.8994214Z region: us-east-1 2025-12-04T10:24:51.8994371Z env: 2025-12-04T10:24:51.8994515Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:24:51.8994701Z HAS_NVIDIA_GPU: true 2025-12-04T10:24:51.8994939Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:24:51.8995338Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:24:51.8995688Z ##[endgroup] 2025-12-04T10:24:52.1844954Z No files were found with the provided path: debug-*.zip. No artifacts will be uploaded. 2025-12-04T10:24:52.2089483Z ##[group]Run # shellcheck disable=SC2156 2025-12-04T10:24:52.2089806Z # shellcheck disable=SC2156 2025-12-04T10:24:52.2090255Z find . -iname "core.[1-9]*" -exec docker exec "${DOCKER_CONTAINER_ID}" sh -c "gdb python {} -ex 'bt' -ex 'q'" \; 2025-12-04T10:24:52.2098904Z shell: /usr/bin/bash -e {0} 2025-12-04T10:24:52.2099352Z env: 2025-12-04T10:24:52.2099511Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:24:52.2099697Z HAS_NVIDIA_GPU: true 2025-12-04T10:24:52.2099923Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:24:52.2100325Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:24:52.2100680Z ##[endgroup] 2025-12-04T10:24:52.6114778Z ##[group]Run seemethere/upload-artifact-s3@baba72d0712b404f646cebe0730933554ebce96a 2025-12-04T10:24:52.6115430Z with: 2025-12-04T10:24:52.6115945Z name: coredumps-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu 2025-12-04T10:24:52.6116609Z retention-days: 14 2025-12-04T10:24:52.6116907Z if-no-files-found: ignore 2025-12-04T10:24:52.6117235Z path: ./**/core.[1-9]* 2025-12-04T10:24:52.6117569Z s3-bucket: gha-artifacts 2025-12-04T10:24:52.6117894Z region: us-east-1 2025-12-04T10:24:52.6118168Z env: 2025-12-04T10:24:52.6118420Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:24:52.6118724Z HAS_NVIDIA_GPU: true 2025-12-04T10:24:52.6119121Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:24:52.6119830Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:24:52.6120462Z ##[endgroup] 2025-12-04T10:25:02.8461252Z No files were found with the provided path: ./**/core.[1-9]*. No artifacts will be uploaded. 2025-12-04T10:25:02.8809895Z Prepare all required actions 2025-12-04T10:25:02.8810228Z Getting action download info 2025-12-04T10:25:03.0359660Z Download action repository 'actions/setup-python@v6' (SHA:83679a892e2d95755f2dac6acb0bfd1e9ac5d548) 2025-12-04T10:25:03.4315401Z ##[group]Run ./.github/actions/upload-utilization-stats 2025-12-04T10:25:03.4315666Z with: 2025-12-04T10:25:03.4315822Z job_id: 57118543727 2025-12-04T10:25:03.4316190Z job_name: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu, unstable) 2025-12-04T10:25:03.4316598Z workflow_name: vllm-test 2025-12-04T10:25:03.4316790Z workflow_run_id: 19922031852 2025-12-04T10:25:03.4316987Z workflow_attempt: 1 2025-12-04T10:25:03.4317152Z env: 2025-12-04T10:25:03.4317295Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:25:03.4317475Z HAS_NVIDIA_GPU: true 2025-12-04T10:25:03.4317695Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:25:03.4318119Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:25:03.4318465Z ##[endgroup] 2025-12-04T10:25:03.4349344Z ##[group]Run actions/setup-python@v6 2025-12-04T10:25:03.4349580Z with: 2025-12-04T10:25:03.4349735Z python-version: 3.10 2025-12-04T10:25:03.4350045Z check-latest: false 2025-12-04T10:25:03.4350322Z token: *** 2025-12-04T10:25:03.4350493Z update-environment: true 2025-12-04T10:25:03.4350688Z allow-prereleases: false 2025-12-04T10:25:03.4350875Z freethreaded: false 2025-12-04T10:25:03.4351042Z env: 2025-12-04T10:25:03.4351185Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:25:03.4351368Z HAS_NVIDIA_GPU: true 2025-12-04T10:25:03.4351593Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:25:03.4351979Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:25:03.4352327Z ##[endgroup] 2025-12-04T10:25:03.9292286Z ##[group]Installed versions 2025-12-04T10:25:03.9301045Z Version 3.10 was not found in the local cache 2025-12-04T10:25:03.9572600Z (node:95523) [DEP0040] DeprecationWarning: The `punycode` module is deprecated. Please use a userland alternative instead. 2025-12-04T10:25:03.9573636Z (Use `node --trace-deprecation ...` to show where the warning was created) 2025-12-04T10:25:04.2686729Z ##[error]The version '3.10' with architecture 'x64' was not found for this operating system. The list of all available versions can be found here: https://raw.githubusercontent.com/actions/python-versions/main/versions-manifest.json 2025-12-04T10:25:04.2892156Z ##[group]Run pytorch/test-infra/.github/actions/teardown-linux@main 2025-12-04T10:25:04.2892628Z with: 2025-12-04T10:25:04.2892771Z env: 2025-12-04T10:25:04.2892924Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:25:04.2893118Z HAS_NVIDIA_GPU: true 2025-12-04T10:25:04.2893346Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:25:04.2893753Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:25:04.2894094Z ##[endgroup] 2025-12-04T10:25:04.2907928Z ##[group]Run set -eou pipefail 2025-12-04T10:25:04.2908170Z set -eou pipefail 2025-12-04T10:25:04.2908362Z  2025-12-04T10:25:04.2908622Z echo "Holding runner for 2 hours until all ssh sessions have logged out" 2025-12-04T10:25:04.2908956Z for _ in $(seq 1440); do 2025-12-04T10:25:04.2909185Z  # Break if no ssh session exists anymore 2025-12-04T10:25:04.2909426Z  if [ "$(who)" = "" ]; then 2025-12-04T10:25:04.2909658Z  break 2025-12-04T10:25:04.2909829Z  fi 2025-12-04T10:25:04.2909988Z  echo "." 2025-12-04T10:25:04.2910164Z  sleep 5 2025-12-04T10:25:04.2910319Z done 2025-12-04T10:25:04.2918018Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:25:04.2918281Z env: 2025-12-04T10:25:04.2918431Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:25:04.2918617Z HAS_NVIDIA_GPU: true 2025-12-04T10:25:04.2918838Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:25:04.2919240Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:25:04.2919595Z ##[endgroup] 2025-12-04T10:25:04.2947485Z Holding runner for 2 hours until all ssh sessions have logged out 2025-12-04T10:25:04.3056039Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T10:25:04.3056438Z # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T10:25:04.3056745Z # shellcheck disable=SC2046 2025-12-04T10:25:04.3056990Z docker stop $(docker ps -q) || true 2025-12-04T10:25:04.3057251Z # Prune all of the docker images 2025-12-04T10:25:04.3057474Z docker system prune -af 2025-12-04T10:25:04.3064601Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:25:04.3064874Z env: 2025-12-04T10:25:04.3065021Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:25:04.3065205Z HAS_NVIDIA_GPU: true 2025-12-04T10:25:04.3065426Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:25:04.3065819Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:25:04.3066170Z ##[endgroup] 2025-12-04T10:25:16.3018660Z dc9d197ee609 2025-12-04T10:25:25.8261292Z Deleted Containers: 2025-12-04T10:25:25.8261715Z dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:25:25.8262136Z 2025-12-04T10:25:35.2201022Z Deleted Images: 2025-12-04T10:25:35.2201961Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T10:25:35.2203134Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image@sha256:26aca0736a4ce89ce91ccedb1f5d90ce14b0dd6252f01598064e66377063a214 2025-12-04T10:25:35.2203952Z deleted: sha256:0ccbf4fab5f55e557ff63d66607494707567ca363d0460db6f3a59a889b43505 2025-12-04T10:25:35.2204523Z deleted: sha256:d6d89e483caae7b93ff6002401579373a1c56ccbb004e00bd52a0d61b5502917 2025-12-04T10:25:35.2205060Z deleted: sha256:6046dce7adef7507f779762847a250b007cfbf29a76214261c11c11fa15b5582 2025-12-04T10:25:35.2205597Z deleted: sha256:82b58657615bd7679d9d03b5762819b1aaed3b334fae20f3a89c1f2b4e29b57e 2025-12-04T10:25:35.2206075Z deleted: sha256:715e1d20a0f3534120ff63837ef5d98f070503d3d7cca605d86e4dedff1483f2 2025-12-04T10:25:35.2206500Z deleted: sha256:50317a1ace6bcc016871f9a46ad2bcef533c73854ecb3859e57a1330a4816355 2025-12-04T10:25:35.2207251Z deleted: sha256:07440812a080f81619e329e4a5e57b4763d527b0f1cbfbde10e12d1fa3e8665c 2025-12-04T10:25:35.2207721Z deleted: sha256:dae702cd12fb11545262c01f4f1d2accce6c45325ef2d77336a2e9546da8470c 2025-12-04T10:25:35.2208270Z deleted: sha256:9ac3e4517d9f30edce3e2bb08401e068904509ee74fc46a16a8149f9f9f7bacd 2025-12-04T10:25:35.2208709Z deleted: sha256:ac135c8b6bc69f23b5520369c6a1b31f4bef2b53f97002c444628141de9d8654 2025-12-04T10:25:35.2209144Z deleted: sha256:abc1bd3d1bf2f3ef31dbb6a1df71a5cce99ab19842e107778c813c7a4679dd20 2025-12-04T10:25:35.2209577Z deleted: sha256:3de6e9ed31faca4e43f13801f651c407757095ca56b92d55613519cc331bd71e 2025-12-04T10:25:35.2210004Z deleted: sha256:5b8839850e12dd4597c9e9c795b948a6866177ce5e1ca8c21ed83efa5f113f01 2025-12-04T10:25:35.2210432Z deleted: sha256:229d26bb57f50f8053e7afef135e466d8f9aa5e34c147a228cff74237d5520c2 2025-12-04T10:25:35.2210863Z deleted: sha256:31ffbd21c0ab8f091ff660a91e23531a66816ff8e5132a7bb8c812f02bc5307b 2025-12-04T10:25:35.2211293Z deleted: sha256:4f676ab679d9728439cc1d56082776eb77d974472de5a7adf4aa6758e331d4e0 2025-12-04T10:25:35.2211719Z deleted: sha256:eaef4de65cf0416ad4d19f60a1219def9a7530562bcfd1685c27dbde0aeaa314 2025-12-04T10:25:35.2212159Z deleted: sha256:fefc1b0f00f7c4d06ed919834ed541d68953820b7dfe8e42bca221d8363a53a1 2025-12-04T10:25:35.2212587Z deleted: sha256:6af11d6c331c5152b1f021309e86e6f2361642184222d59cca5fa8278252e998 2025-12-04T10:25:35.2213007Z deleted: sha256:92b0b3452c273450bef1462ee49f4042097211ff4929f7edacdb358b52181c8d 2025-12-04T10:25:35.2213425Z deleted: sha256:bf7154aa212f05d0945b39c7f776973ec958230fc5de114bf0f508dfde104362 2025-12-04T10:25:35.2213860Z deleted: sha256:85cf7cb44822fd42d15c866bfaeacbb85edce53c6fdd66a132b29d9f2d0cc101 2025-12-04T10:25:35.2214312Z deleted: sha256:1b568bc90bd1db1955a72a5ae4251ef7867acaa0837df1f4c03c72316ff2a5a7 2025-12-04T10:25:35.2214742Z deleted: sha256:1b979582b9703d9b1ae45b022715e4cfd83a3b2b2983201be01d04a45c91b827 2025-12-04T10:25:35.2215179Z deleted: sha256:473bdb3efc53041dedd3f3b31bf360e4e98af379240ee854b52e73042988bec5 2025-12-04T10:25:35.2215628Z deleted: sha256:adc41763227119ff0efc0c9932890ae7a5a85efbb3aa021320a7eb32855a6527 2025-12-04T10:25:35.2216072Z deleted: sha256:f1c09347aedf1e3e94f51980d8031b0e79f539b3815a4aeb2f247c6765955a98 2025-12-04T10:25:35.2216487Z deleted: sha256:6494f600573048d464a065deab11992f4ee00c1db1ccdf3f54b8833437bd162c 2025-12-04T10:25:35.2216921Z deleted: sha256:8d94e9665e84daedc6dab86e3db3a14e2646613fbe62124afea1627d4852a680 2025-12-04T10:25:35.2217356Z deleted: sha256:adb7c970ad4132884c6fdf2cb802733fee2d6b70e3ab6fe1e3f016643f691c76 2025-12-04T10:25:35.2217827Z deleted: sha256:5e7c872ebf4e697a6d3c6f45b87d9ea156f6c47291be3db5b5d51dd6b5c178bc 2025-12-04T10:25:35.2218254Z deleted: sha256:9639c558f4c7da0e52a767ecc0d82b8c68060bc54d03e9550039043730d9bf7b 2025-12-04T10:25:35.2218777Z deleted: sha256:0198a32bb3f4052e927c402c7a51dedeaa449d0940616d7fed422a5a32f8aea3 2025-12-04T10:25:35.2219338Z deleted: sha256:2a4a8be64e29c7a6037003cb94bdec1b2c438b783c8c2614ff2ad095c930eb71 2025-12-04T10:25:35.2219778Z deleted: sha256:42a7b298ec1108139de9a7c40cf959b9753397afb7d8657787da03c1e3bdc9f3 2025-12-04T10:25:35.2220206Z deleted: sha256:24333eea3d7417550b632d8204b62fd3703d34eec9aacd7d092f9847ca4be02d 2025-12-04T10:25:35.2230438Z deleted: sha256:aa0d5bf0be63c55d5d0af39dd279a87dcfe9804b28609328250d8fcef65c6d0f 2025-12-04T10:25:35.2230926Z deleted: sha256:2f92f199a43a7c3f5d5eb4ce8e1c60739bccb73c6228c4f6da665d287c73747c 2025-12-04T10:25:35.2231377Z deleted: sha256:a782a6e8c5eeb31c8a0c8b191eaf841cf7b4e60118e35331af9575a603b7437d 2025-12-04T10:25:35.2231830Z deleted: sha256:17d12b900a816c35a6653cde50088ca434b7b55fb4b8aee238112921b8ff76e2 2025-12-04T10:25:35.2232286Z deleted: sha256:7ca2614c76f7d8c4b7a8757c0691f6525b9ae48999f5f657a20cdc45f4600501 2025-12-04T10:25:35.2232741Z deleted: sha256:d3192a402a87e956603694228a377efc787caf6d6d3629eb656ab22fc0661fd4 2025-12-04T10:25:35.2233170Z deleted: sha256:55cf1bc8ad4d93d4216d89afd208938bfe95736df6bd22fe38ce65d88aa85ae7 2025-12-04T10:25:35.2233766Z deleted: sha256:48ff0b3405a517a3fa2e8b1b010a06be39059970f23dec077b90566ef2c8be7f 2025-12-04T10:25:35.2234222Z deleted: sha256:1bbadee83428d7823f312af1602d95f4c3877592a8fbac20a89eee14054f5317 2025-12-04T10:25:35.2234702Z deleted: sha256:2d36d8a0aa92501354e41dea5c096f44602b99d187221219df375fb6da8ce399 2025-12-04T10:25:35.2235136Z deleted: sha256:5f30cc351960c0de3011f71ecdabcdac4084ae0baa69d48a670804a8d5c5940b 2025-12-04T10:25:35.2235565Z deleted: sha256:a8587f4b7f2c65f78601848588796ea0c0b7c6a1cac811102064236be5559ac6 2025-12-04T10:25:35.2235998Z deleted: sha256:41effb6079fcd08f13a6ae70359e35e7cf1191bcdc006557f3013646c3b2fdb6 2025-12-04T10:25:35.2236445Z deleted: sha256:0eb6eb00fcf75aa4b07cd49152fa88531cc8d9a0aac86f50ec3bee01b78e1230 2025-12-04T10:25:35.2236894Z deleted: sha256:96e8723cadec37845759e2540ed12135470addffaa1e7c6a7d2583499ab0354f 2025-12-04T10:25:35.2237320Z deleted: sha256:17145782650a813072b769c21d242c084273ef7f363d6e99927d8b426ae2e323 2025-12-04T10:25:35.2237728Z deleted: sha256:e4051a43633632655846c780ad553003608a922e40a79c3e71eb2456993afe30 2025-12-04T10:25:35.2238159Z deleted: sha256:e409e265b0b0a13bd0e7ba7078511b61afc5c8f4353e73c1b2acb3f6f807122a 2025-12-04T10:25:35.2238600Z deleted: sha256:9d2469d6fbb9da43819cdb1fabcb1b222421ef4f0ebbf04436c631ee50502666 2025-12-04T10:25:35.2239033Z deleted: sha256:ce32af9f70868415c59b2b094b90f0e77d9f0dd9ba1bc91282c3ee0334adc8a1 2025-12-04T10:25:35.2239455Z deleted: sha256:f61b3e95401d10bec28028d964e8f17c15d580af9bfe7e395b769b2ce53f6a21 2025-12-04T10:25:35.2239879Z deleted: sha256:6591a75ff8c2005f7e0b0cbb571908a5d97b4d016d3bc90b387710843b769274 2025-12-04T10:25:35.2240308Z deleted: sha256:e369a5e07bfbe78b4526b3e13a5e66899a9e43fa26c98b2f68fb903d73fe2559 2025-12-04T10:25:35.2240743Z deleted: sha256:b680eca9dfbf17d65d597a39fe2538288a1eaeca2f08934baf859e7c7825d752 2025-12-04T10:25:35.2241197Z deleted: sha256:a0570baf0430eff1dfe0726be745dbeae81c7e5d306ff5e33d1f27b19c084401 2025-12-04T10:25:35.2241644Z deleted: sha256:9398cda2228787e403e29955938ff118c3f739fce7873bcb3ac9f271b120bf8c 2025-12-04T10:25:35.2242070Z deleted: sha256:d644da137639d8587342055e9126e84ddb9b02d4b81d4643841029a0c8cc8a24 2025-12-04T10:25:35.2242498Z deleted: sha256:cc53b7d607b3358cdcd2d8a16abb7afeef7fd7a5c18594ee6e5e1754615e1468 2025-12-04T10:25:35.2242940Z deleted: sha256:b8241fa8b18e8ed419b7b5e514c071d56e63d387d45834c703fecc017b2de3f9 2025-12-04T10:25:35.2243390Z deleted: sha256:d9248bd92b0d14e59fada1c3d06b2decae56c719a65f0a93bb3858f8668fc0a4 2025-12-04T10:25:35.2243823Z deleted: sha256:7b5268889149786cd207a90fe9a34f01085219e0046ac3e6515cf067b9994eb1 2025-12-04T10:25:35.2244245Z deleted: sha256:8c7ee1b853c0ee7cb8a7623a1cccc5db1392e99ca605d511e5f66db6fbe26aba 2025-12-04T10:25:35.2244683Z deleted: sha256:3ba56019c8a2981290b7dd24caca7f31103ff85b51a8398ab277edd17edde3d8 2025-12-04T10:25:35.2245167Z deleted: sha256:5e53c57b4a34c308e85b66c0dbf36890288a4921fa2419b93ea5db95ec6269bf 2025-12-04T10:25:35.2245601Z deleted: sha256:c1c39a1ef57a5de8a3e36f4d166d60a9ee0a45fc4e9a93382459dd93cad88419 2025-12-04T10:25:35.2246025Z deleted: sha256:35e32f19686b2917502fe72e9533b7a038615d78a6b9881584649073accadc75 2025-12-04T10:25:35.2246442Z deleted: sha256:75684b69423f25ec20d72bdedc8e2e26f9fe68706ea825496f56b0a8e701fe4f 2025-12-04T10:25:35.2247048Z deleted: sha256:2f7f7c60f0e2606288945684e4f98431b9009c827adf6047f1c422b759ee74cf 2025-12-04T10:25:35.2247798Z deleted: sha256:c5e7eb42593fd6b737557df0eb7ba544929bee03db5255c97cd63a89755413eb 2025-12-04T10:25:35.2248519Z deleted: sha256:410586d156a1c662668a6ce4c4f05eeed4fdffe1314f00b1274a23190754e740 2025-12-04T10:25:35.2249004Z deleted: sha256:5992ca9d69ab1b04d3814299d522118bd1e7bbe34c7e57d5e98ccef74c4034c1 2025-12-04T10:25:35.2249427Z deleted: sha256:28e562eb98b063dec35d0c6b00c0880b944dc153a3172d1dd55e055b9de230be 2025-12-04T10:25:35.2249869Z deleted: sha256:b63c3db121a379cb83c27b98d44c3337b358dce01ac9df1aae1b0aca6e6ebf99 2025-12-04T10:25:35.2250298Z deleted: sha256:901e91430cda8325696785092962b7212f98d2c41487d7c0a0b2993acd06e49c 2025-12-04T10:25:35.2250856Z deleted: sha256:429a744a1c9c356d9d7ff639dfc8ae2be29b1f47ade5027d641e980c72ba3056 2025-12-04T10:25:35.2251297Z deleted: sha256:306e23501fe1a51ea1702bf09ebd81c901e18a13f0cb0b3292b32f718e2b300e 2025-12-04T10:25:35.2251772Z deleted: sha256:7c3a28a64825dfed0b482cf73a833043dd538d6496eb04fb1e5ee42a2db6489d 2025-12-04T10:25:35.2252205Z deleted: sha256:0852f61af41399ab6a9cf19c12c789ce58a4d05907da048d26af71afa10f8919 2025-12-04T10:25:35.2252635Z deleted: sha256:73974f74b436f39a2fdb6461b1e3f7c3e41c73325776fa71d16b942a5b4a365b 2025-12-04T10:25:35.2253000Z untagged: public.ecr.aws/docker/library/python:3.13 2025-12-04T10:25:35.2253501Z untagged: public.ecr.aws/docker/library/python@sha256:3f986299a7b8b44b0d8cf9bda2b22361ce5c3058ef5d7cb17fb7452506680ab0 2025-12-04T10:25:35.2254099Z deleted: sha256:44438aecfedf7b6086fce506dae0db5ba7fc0027f9b743f1a75a6b5cbc7de70a 2025-12-04T10:25:35.2254553Z deleted: sha256:6f09a1f5d8a107c2532fbd116e75116cb75fa77b1a7d72d3bdf1ac12de152acd 2025-12-04T10:25:35.2255006Z deleted: sha256:fe5f3ac0be086125eb1e3cd10cc33e8e426f4e079381f7ce5a987b626e99fa67 2025-12-04T10:25:35.2255444Z deleted: sha256:79dd2061a22cf919cfc4f1f02704bfda09afadb017265e670ee54441d296c06c 2025-12-04T10:25:35.2255889Z deleted: sha256:9447ad402aafdbee17e999b0ec84ad89c2646dbebf054d469d4f8bee77f66212 2025-12-04T10:25:35.2256326Z deleted: sha256:7a4909f3c1975be52292f53107495ee1b41c17494918767ccedf1cf1688ae318 2025-12-04T10:25:35.2256756Z deleted: sha256:3474923d97f1f498237650a7d51bd4aea37d5e6b9d8a778777920584af5dd560 2025-12-04T10:25:35.2257190Z deleted: sha256:683afd1773444401a9cbd24842ee5d9154a11abb4fab63ddea5c03df788597ee 2025-12-04T10:25:35.2257444Z 2025-12-04T10:25:35.2257543Z Total reclaimed space: 66.18GB 2025-12-04T10:25:35.2305965Z ##[group]Run set +e 2025-12-04T10:25:35.2306247Z set +e 2025-12-04T10:25:35.2306425Z set -x 2025-12-04T10:25:35.2306579Z  2025-12-04T10:25:35.2306738Z nvidia-smi 2025-12-04T10:25:35.2307060Z # NB: Surprisingly, nvidia-smi command returns successfully with return code 0 even in 2025-12-04T10:25:35.2307536Z # the case where the driver has already crashed as it still can get the driver version 2025-12-04T10:25:35.2308002Z # and some basic information like the bus ID. However, the rest of the information 2025-12-04T10:25:35.2308365Z # would be missing (ERR!), for example: 2025-12-04T10:25:35.2308592Z # 2025-12-04T10:25:35.2308808Z # +-----------------------------------------------------------------------------+ 2025-12-04T10:25:35.2309194Z # | NVIDIA-SMI 525.89.02 Driver Version: 525.89.02 CUDA Version: 12.0 | 2025-12-04T10:25:35.2309572Z # |-------------------------------+----------------------+----------------------+ 2025-12-04T10:25:35.2310053Z # | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T10:25:35.2310454Z # | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | 2025-12-04T10:25:35.2310788Z # | | | MIG M. | 2025-12-04T10:25:35.2311039Z # |===============================+======================+======================| 2025-12-04T10:25:35.2311319Z # | 0 ERR! Off | 00000000:00:1E.0 Off | ERR! | 2025-12-04T10:25:35.2311642Z # |ERR! ERR! ERR! ERR! / ERR! | 4184MiB / 23028MiB | ERR! Default | 2025-12-04T10:25:35.2311937Z # | | | ERR! | 2025-12-04T10:25:35.2312226Z # +-------------------------------+----------------------+----------------------+ 2025-12-04T10:25:35.2312480Z # 2025-12-04T10:25:35.2312686Z # +-----------------------------------------------------------------------------+ 2025-12-04T10:25:35.2312997Z # | Processes: | 2025-12-04T10:25:35.2313327Z # | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T10:25:35.2313624Z # | ID ID Usage | 2025-12-04T10:25:35.2313959Z # |=============================================================================| 2025-12-04T10:25:35.2314238Z # +-----------------------------------------------------------------------------+ 2025-12-04T10:25:35.2314492Z # 2025-12-04T10:25:35.2314749Z # This should be reported as a failure instead as it will guarantee to fail when 2025-12-04T10:25:35.2315101Z # Docker tries to run with --gpus all 2025-12-04T10:25:35.2315319Z # 2025-12-04T10:25:35.2315577Z # So, the correct check here is to query one of the missing piece of info like 2025-12-04T10:25:35.2315939Z # GPU name, so that the command can fail accordingly 2025-12-04T10:25:35.2316277Z nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 2025-12-04T10:25:35.2316573Z NVIDIA_SMI_STATUS=$? 2025-12-04T10:25:35.2316750Z  2025-12-04T10:25:35.2317052Z # These are acceptable return code from nvidia-smi as copied from setup-nvidia GitHub action 2025-12-04T10:25:35.2317504Z if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then 2025-12-04T10:25:35.2317906Z  echo "NVIDIA driver installation has failed, shutting down the runner..." 2025-12-04T10:25:35.2318245Z  .github/scripts/stop_runner_service.sh 2025-12-04T10:25:35.2318469Z fi 2025-12-04T10:25:35.2318612Z  2025-12-04T10:25:35.2318998Z # For runner with multiple GPUs, we also want to confirm that the number of GPUs are the 2025-12-04T10:25:35.2319444Z # power of 2, i.e. 1, 2, 4, or 8. This is to avoid flaky test issue when one GPU fails 2025-12-04T10:25:35.2319820Z # https://github.com/pytorch/test-infra/issues/4000 2025-12-04T10:25:35.2320125Z GPU_COUNT=$(nvidia-smi --list-gpus | wc -l) 2025-12-04T10:25:35.2320368Z NVIDIA_SMI_STATUS=$? 2025-12-04T10:25:35.2320551Z  2025-12-04T10:25:35.2320846Z # These are acceptable return code from nvidia-smi as copied from setup-nvidia GitHub action 2025-12-04T10:25:35.2321296Z if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then 2025-12-04T10:25:35.2321702Z  echo "NVIDIA driver installation has failed, shutting down the runner..." 2025-12-04T10:25:35.2322051Z  .github/scripts/stop_runner_service.sh 2025-12-04T10:25:35.2322273Z fi 2025-12-04T10:25:35.2322412Z  2025-12-04T10:25:35.2322582Z # Check the GPU count to be a power of 2 2025-12-04T10:25:35.2323014Z if [ "$GPU_COUNT" -le 8 ] && [ "$GPU_COUNT" -ne 1 ] && [ "$GPU_COUNT" -ne 2 ] && [ "$GPU_COUNT" -ne 4 ] && [ "$GPU_COUNT" -ne 8 ]; then 2025-12-04T10:25:35.2323533Z  echo "NVIDIA driver detects $GPU_COUNT GPUs. The runner has a broken GPU, shutting it down..." 2025-12-04T10:25:35.2323917Z  .github/scripts/stop_runner_service.sh 2025-12-04T10:25:35.2324139Z fi 2025-12-04T10:25:35.2336471Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:25:35.2336745Z env: 2025-12-04T10:25:35.2336906Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:25:35.2337097Z HAS_NVIDIA_GPU: true 2025-12-04T10:25:35.2337317Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:25:35.2337717Z DOCKER_CONTAINER_ID: dc9d197ee60933b79606af94de83a2f7e3459d87dd2d6ecf916822144ab1e95f 2025-12-04T10:25:35.2338071Z ##[endgroup] 2025-12-04T10:25:35.2369604Z + nvidia-smi 2025-12-04T10:25:35.2650095Z Thu Dec 4 10:25:35 2025 2025-12-04T10:25:35.2650482Z +-----------------------------------------------------------------------------------------+ 2025-12-04T10:25:35.2650966Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-12-04T10:25:35.2651428Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T10:25:35.2651919Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T10:25:35.2652590Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-12-04T10:25:35.2652984Z | | | MIG M. | 2025-12-04T10:25:35.2653273Z |=========================================+========================+======================| 2025-12-04T10:25:35.2792888Z | 0 NVIDIA L4 On | 00000000:35:00.0 Off | 0 | 2025-12-04T10:25:35.2793337Z | N/A 38C P8 15W / 72W | 0MiB / 23034MiB | 0% Default | 2025-12-04T10:25:35.2793716Z | | | N/A | 2025-12-04T10:25:35.2794079Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T10:25:35.2796784Z 2025-12-04T10:25:35.2796981Z +-----------------------------------------------------------------------------------------+ 2025-12-04T10:25:35.2797411Z | Processes: | 2025-12-04T10:25:35.2797825Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T10:25:35.2798194Z | ID ID Usage | 2025-12-04T10:25:35.2798733Z |=========================================================================================| 2025-12-04T10:25:35.2802076Z | No running processes found | 2025-12-04T10:25:35.2802544Z +-----------------------------------------------------------------------------------------+ 2025-12-04T10:25:35.5176959Z + nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 2025-12-04T10:25:35.5334336Z NVIDIA L4 2025-12-04T10:25:35.5368583Z + NVIDIA_SMI_STATUS=0 2025-12-04T10:25:35.5368812Z + '[' 0 -ne 0 ']' 2025-12-04T10:25:35.5375227Z ++ nvidia-smi --list-gpus 2025-12-04T10:25:35.5377062Z ++ wc -l 2025-12-04T10:25:35.5565228Z + GPU_COUNT=1 2025-12-04T10:25:35.5565591Z + NVIDIA_SMI_STATUS=0 2025-12-04T10:25:35.5565948Z + '[' 0 -ne 0 ']' 2025-12-04T10:25:35.5566269Z + '[' 1 -le 8 ']' 2025-12-04T10:25:35.5566535Z + '[' 1 -ne 1 ']' 2025-12-04T10:25:35.5622583Z Post job cleanup. 2025-12-04T10:25:35.5681594Z Post job cleanup. 2025-12-04T10:25:35.5717421Z Post job cleanup. 2025-12-04T10:25:35.6645075Z [command]/usr/bin/git version 2025-12-04T10:25:35.6703776Z git version 2.50.1 2025-12-04T10:25:35.6737184Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/e1c68203-3523-45c7-98d0-9f6ad815d76d/.gitconfig' 2025-12-04T10:25:35.6751992Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/e1c68203-3523-45c7-98d0-9f6ad815d76d' before making global git config changes 2025-12-04T10:25:35.6752838Z Adding repository directory to the temporary git global config as a safe directory 2025-12-04T10:25:35.6756712Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T10:25:35.6798557Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-12-04T10:25:35.6840110Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-12-04T10:25:35.7215585Z Entering 'android/libs/fbjni' 2025-12-04T10:25:35.7288387Z Entering 'third_party/FP16' 2025-12-04T10:25:35.7358805Z Entering 'third_party/FXdiv' 2025-12-04T10:25:35.7429227Z Entering 'third_party/NNPACK' 2025-12-04T10:25:35.7499375Z Entering 'third_party/NVTX' 2025-12-04T10:25:35.7575665Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T10:25:35.7646966Z Entering 'third_party/XNNPACK' 2025-12-04T10:25:35.7733718Z Entering 'third_party/aiter' 2025-12-04T10:25:35.7803718Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T10:25:35.7884570Z Entering 'third_party/benchmark' 2025-12-04T10:25:35.7961749Z Entering 'third_party/composable_kernel' 2025-12-04T10:25:35.8037799Z Entering 'third_party/cpp-httplib' 2025-12-04T10:25:35.8107069Z Entering 'third_party/cpuinfo' 2025-12-04T10:25:35.8181547Z Entering 'third_party/cudnn_frontend' 2025-12-04T10:25:35.8250054Z Entering 'third_party/cutlass' 2025-12-04T10:25:35.8328126Z Entering 'third_party/fbgemm' 2025-12-04T10:25:35.8400114Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T10:25:35.8468452Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T10:25:35.8550221Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T10:25:35.8619887Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T10:25:35.8697703Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T10:25:35.8766909Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T10:25:35.8838129Z Entering 'third_party/fbgemm/external/json' 2025-12-04T10:25:35.8915997Z Entering 'third_party/flash-attention' 2025-12-04T10:25:35.8989274Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T10:25:35.9064515Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T10:25:35.9147203Z Entering 'third_party/flatbuffers' 2025-12-04T10:25:35.9220770Z Entering 'third_party/fmt' 2025-12-04T10:25:35.9290894Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T10:25:35.9359329Z Entering 'third_party/gloo' 2025-12-04T10:25:35.9429385Z Entering 'third_party/googletest' 2025-12-04T10:25:35.9499058Z Entering 'third_party/ideep' 2025-12-04T10:25:35.9566857Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T10:25:35.9644396Z Entering 'third_party/ittapi' 2025-12-04T10:25:35.9716367Z Entering 'third_party/kineto' 2025-12-04T10:25:35.9786671Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T10:25:35.9856113Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T10:25:35.9929335Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T10:25:35.9998478Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T10:25:36.0069309Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T10:25:36.0137620Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T10:25:36.0211750Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T10:25:36.0282055Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T10:25:36.0352117Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T10:25:36.0422395Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T10:25:36.0490845Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T10:25:36.0558309Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:25:36.0631114Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:25:36.0706774Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T10:25:36.0777064Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T10:25:36.0850962Z Entering 'third_party/kleidiai' 2025-12-04T10:25:36.0921444Z Entering 'third_party/mimalloc' 2025-12-04T10:25:36.0991872Z Entering 'third_party/nlohmann' 2025-12-04T10:25:36.1061171Z Entering 'third_party/onnx' 2025-12-04T10:25:36.1144054Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T10:25:36.1215346Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T10:25:36.1288708Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T10:25:36.1357911Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T10:25:36.1431987Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T10:25:36.1499853Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T10:25:36.1569635Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T10:25:36.1638769Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T10:25:36.1707720Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T10:25:36.1776884Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:25:36.1849960Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:25:36.1921350Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T10:25:36.2011063Z Entering 'third_party/pocketfft' 2025-12-04T10:25:36.2080740Z Entering 'third_party/protobuf' 2025-12-04T10:25:36.2152373Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T10:25:36.2220313Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T10:25:36.2292527Z Entering 'third_party/psimd' 2025-12-04T10:25:36.2360807Z Entering 'third_party/pthreadpool' 2025-12-04T10:25:36.2431234Z Entering 'third_party/pybind11' 2025-12-04T10:25:36.2499815Z Entering 'third_party/python-peachpy' 2025-12-04T10:25:36.2568775Z Entering 'third_party/sleef' 2025-12-04T10:25:36.2638313Z Entering 'third_party/tensorpipe' 2025-12-04T10:25:36.2708022Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T10:25:36.2777644Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T10:25:36.2847439Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T10:25:36.2919850Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T10:25:36.2987525Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T10:25:36.3085716Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-12-04T10:25:36.3109757Z http.https://github.com/.extraheader 2025-12-04T10:25:36.3119330Z [command]/usr/bin/git config --local --unset-all http.https://github.com/.extraheader 2025-12-04T10:25:36.3152245Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-12-04T10:25:36.3511545Z Entering 'android/libs/fbjni' 2025-12-04T10:25:36.3560450Z http.https://github.com/.extraheader 2025-12-04T10:25:36.3602848Z Entering 'third_party/FP16' 2025-12-04T10:25:36.3648589Z http.https://github.com/.extraheader 2025-12-04T10:25:36.3691874Z Entering 'third_party/FXdiv' 2025-12-04T10:25:36.3736946Z http.https://github.com/.extraheader 2025-12-04T10:25:36.3781278Z Entering 'third_party/NNPACK' 2025-12-04T10:25:36.3826480Z http.https://github.com/.extraheader 2025-12-04T10:25:36.3870517Z Entering 'third_party/NVTX' 2025-12-04T10:25:36.3916214Z http.https://github.com/.extraheader 2025-12-04T10:25:36.3960450Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T10:25:36.4006604Z http.https://github.com/.extraheader 2025-12-04T10:25:36.4050610Z Entering 'third_party/XNNPACK' 2025-12-04T10:25:36.4096549Z http.https://github.com/.extraheader 2025-12-04T10:25:36.4154122Z Entering 'third_party/aiter' 2025-12-04T10:25:36.4202098Z http.https://github.com/.extraheader 2025-12-04T10:25:36.4247188Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T10:25:36.4295068Z http.https://github.com/.extraheader 2025-12-04T10:25:36.4348535Z Entering 'third_party/benchmark' 2025-12-04T10:25:36.4395596Z http.https://github.com/.extraheader 2025-12-04T10:25:36.4439325Z Entering 'third_party/composable_kernel' 2025-12-04T10:25:36.4487089Z http.https://github.com/.extraheader 2025-12-04T10:25:36.4536662Z Entering 'third_party/cpp-httplib' 2025-12-04T10:25:36.4586287Z http.https://github.com/.extraheader 2025-12-04T10:25:36.4629206Z Entering 'third_party/cpuinfo' 2025-12-04T10:25:36.4676102Z http.https://github.com/.extraheader 2025-12-04T10:25:36.4720485Z Entering 'third_party/cudnn_frontend' 2025-12-04T10:25:36.4766346Z http.https://github.com/.extraheader 2025-12-04T10:25:36.4810895Z Entering 'third_party/cutlass' 2025-12-04T10:25:36.4856260Z http.https://github.com/.extraheader 2025-12-04T10:25:36.4908906Z Entering 'third_party/fbgemm' 2025-12-04T10:25:36.4955452Z http.https://github.com/.extraheader 2025-12-04T10:25:36.5002527Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T10:25:36.5047285Z http.https://github.com/.extraheader 2025-12-04T10:25:36.5091357Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T10:25:36.5135902Z http.https://github.com/.extraheader 2025-12-04T10:25:36.5188598Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T10:25:36.5234360Z http.https://github.com/.extraheader 2025-12-04T10:25:36.5281705Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T10:25:36.5326219Z http.https://github.com/.extraheader 2025-12-04T10:25:36.5378852Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T10:25:36.5424699Z http.https://github.com/.extraheader 2025-12-04T10:25:36.5469333Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T10:25:36.5514135Z http.https://github.com/.extraheader 2025-12-04T10:25:36.5561588Z Entering 'third_party/fbgemm/external/json' 2025-12-04T10:25:36.5607065Z http.https://github.com/.extraheader 2025-12-04T10:25:36.5654025Z Entering 'third_party/flash-attention' 2025-12-04T10:25:36.5703305Z http.https://github.com/.extraheader 2025-12-04T10:25:36.5750454Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T10:25:36.5794173Z http.https://github.com/.extraheader 2025-12-04T10:25:36.5844072Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T10:25:36.5892141Z http.https://github.com/.extraheader 2025-12-04T10:25:36.5949510Z Entering 'third_party/flatbuffers' 2025-12-04T10:25:36.5994309Z http.https://github.com/.extraheader 2025-12-04T10:25:36.6041451Z Entering 'third_party/fmt' 2025-12-04T10:25:36.6086673Z http.https://github.com/.extraheader 2025-12-04T10:25:36.6129844Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T10:25:36.6175805Z http.https://github.com/.extraheader 2025-12-04T10:25:36.6222305Z Entering 'third_party/gloo' 2025-12-04T10:25:36.6267414Z http.https://github.com/.extraheader 2025-12-04T10:25:36.6311301Z Entering 'third_party/googletest' 2025-12-04T10:25:36.6357066Z http.https://github.com/.extraheader 2025-12-04T10:25:36.6400814Z Entering 'third_party/ideep' 2025-12-04T10:25:36.6445915Z http.https://github.com/.extraheader 2025-12-04T10:25:36.6487925Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T10:25:36.6534595Z http.https://github.com/.extraheader 2025-12-04T10:25:36.6588013Z Entering 'third_party/ittapi' 2025-12-04T10:25:36.6634923Z http.https://github.com/.extraheader 2025-12-04T10:25:36.6679847Z Entering 'third_party/kineto' 2025-12-04T10:25:36.6724502Z http.https://github.com/.extraheader 2025-12-04T10:25:36.6767706Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T10:25:36.6815588Z http.https://github.com/.extraheader 2025-12-04T10:25:36.6858794Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T10:25:36.6905974Z http.https://github.com/.extraheader 2025-12-04T10:25:36.6951190Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T10:25:36.6994059Z http.https://github.com/.extraheader 2025-12-04T10:25:36.7039519Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T10:25:36.7085559Z http.https://github.com/.extraheader 2025-12-04T10:25:36.7129683Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T10:25:36.7176928Z http.https://github.com/.extraheader 2025-12-04T10:25:36.7219259Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T10:25:36.7265036Z http.https://github.com/.extraheader 2025-12-04T10:25:36.7314684Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T10:25:36.7358552Z http.https://github.com/.extraheader 2025-12-04T10:25:36.7403616Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T10:25:36.7453240Z http.https://github.com/.extraheader 2025-12-04T10:25:36.7497505Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T10:25:36.7544311Z http.https://github.com/.extraheader 2025-12-04T10:25:36.7590233Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T10:25:36.7636737Z http.https://github.com/.extraheader 2025-12-04T10:25:36.7682364Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T10:25:36.7727324Z http.https://github.com/.extraheader 2025-12-04T10:25:36.7770297Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:25:36.7816119Z http.https://github.com/.extraheader 2025-12-04T10:25:36.7863468Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:25:36.7908837Z http.https://github.com/.extraheader 2025-12-04T10:25:36.7958782Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T10:25:36.8005687Z http.https://github.com/.extraheader 2025-12-04T10:25:36.8049835Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T10:25:36.8095608Z http.https://github.com/.extraheader 2025-12-04T10:25:36.8142939Z Entering 'third_party/kleidiai' 2025-12-04T10:25:36.8188145Z http.https://github.com/.extraheader 2025-12-04T10:25:36.8231169Z Entering 'third_party/mimalloc' 2025-12-04T10:25:36.8277475Z http.https://github.com/.extraheader 2025-12-04T10:25:36.8320783Z Entering 'third_party/nlohmann' 2025-12-04T10:25:36.8366736Z http.https://github.com/.extraheader 2025-12-04T10:25:36.8411545Z Entering 'third_party/onnx' 2025-12-04T10:25:36.8456089Z http.https://github.com/.extraheader 2025-12-04T10:25:36.8512477Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T10:25:36.8557118Z http.https://github.com/.extraheader 2025-12-04T10:25:36.8604886Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T10:25:36.8652892Z http.https://github.com/.extraheader 2025-12-04T10:25:36.8698901Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T10:25:36.8745610Z http.https://github.com/.extraheader 2025-12-04T10:25:36.8788710Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T10:25:36.8834502Z http.https://github.com/.extraheader 2025-12-04T10:25:36.8878995Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T10:25:36.8924541Z http.https://github.com/.extraheader 2025-12-04T10:25:36.8968796Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T10:25:36.9013918Z http.https://github.com/.extraheader 2025-12-04T10:25:36.9059030Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T10:25:36.9105167Z http.https://github.com/.extraheader 2025-12-04T10:25:36.9148580Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T10:25:36.9194823Z http.https://github.com/.extraheader 2025-12-04T10:25:36.9238813Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T10:25:36.9284831Z http.https://github.com/.extraheader 2025-12-04T10:25:36.9327047Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:25:36.9375033Z http.https://github.com/.extraheader 2025-12-04T10:25:36.9421346Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:25:36.9466861Z http.https://github.com/.extraheader 2025-12-04T10:25:36.9513705Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T10:25:36.9559476Z http.https://github.com/.extraheader 2025-12-04T10:25:36.9620860Z Entering 'third_party/pocketfft' 2025-12-04T10:25:36.9667432Z http.https://github.com/.extraheader 2025-12-04T10:25:36.9710662Z Entering 'third_party/protobuf' 2025-12-04T10:25:36.9756228Z http.https://github.com/.extraheader 2025-12-04T10:25:36.9800850Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T10:25:36.9845782Z http.https://github.com/.extraheader 2025-12-04T10:25:36.9889299Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T10:25:36.9933831Z http.https://github.com/.extraheader 2025-12-04T10:25:36.9982932Z Entering 'third_party/psimd' 2025-12-04T10:25:37.0027392Z http.https://github.com/.extraheader 2025-12-04T10:25:37.0070483Z Entering 'third_party/pthreadpool' 2025-12-04T10:25:37.0115979Z http.https://github.com/.extraheader 2025-12-04T10:25:37.0159141Z Entering 'third_party/pybind11' 2025-12-04T10:25:37.0205419Z http.https://github.com/.extraheader 2025-12-04T10:25:37.0248770Z Entering 'third_party/python-peachpy' 2025-12-04T10:25:37.0295618Z http.https://github.com/.extraheader 2025-12-04T10:25:37.0339095Z Entering 'third_party/sleef' 2025-12-04T10:25:37.0385373Z http.https://github.com/.extraheader 2025-12-04T10:25:37.0429396Z Entering 'third_party/tensorpipe' 2025-12-04T10:25:37.0475910Z http.https://github.com/.extraheader 2025-12-04T10:25:37.0518706Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T10:25:37.0564298Z http.https://github.com/.extraheader 2025-12-04T10:25:37.0608620Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T10:25:37.0654515Z http.https://github.com/.extraheader 2025-12-04T10:25:37.0698981Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T10:25:37.0744787Z http.https://github.com/.extraheader 2025-12-04T10:25:37.0789319Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T10:25:37.0834630Z http.https://github.com/.extraheader 2025-12-04T10:25:37.0878110Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T10:25:37.0924762Z http.https://github.com/.extraheader 2025-12-04T10:25:37.0995111Z [command]/usr/bin/git config --local --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.1028556Z [command]/usr/bin/git submodule foreach --recursive git config --local --show-origin --name-only --get-regexp remote.origin.url 2025-12-04T10:25:37.1388332Z Entering 'android/libs/fbjni' 2025-12-04T10:25:37.1419585Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-12-04T10:25:37.1441340Z Entering 'third_party/FP16' 2025-12-04T10:25:37.1471533Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-12-04T10:25:37.1492002Z Entering 'third_party/FXdiv' 2025-12-04T10:25:37.1522625Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-12-04T10:25:37.1545349Z Entering 'third_party/NNPACK' 2025-12-04T10:25:37.1575561Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-12-04T10:25:37.1598526Z Entering 'third_party/NVTX' 2025-12-04T10:25:37.1629019Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-12-04T10:25:37.1650932Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T10:25:37.1682073Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-12-04T10:25:37.1708271Z Entering 'third_party/XNNPACK' 2025-12-04T10:25:37.1739083Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-12-04T10:25:37.1774680Z Entering 'third_party/aiter' 2025-12-04T10:25:37.1806207Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-12-04T10:25:37.1828001Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T10:25:37.1858560Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-12-04T10:25:37.1891489Z Entering 'third_party/benchmark' 2025-12-04T10:25:37.1923385Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-12-04T10:25:37.1946004Z Entering 'third_party/composable_kernel' 2025-12-04T10:25:37.1979344Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-12-04T10:25:37.2008589Z Entering 'third_party/cpp-httplib' 2025-12-04T10:25:37.2039110Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-12-04T10:25:37.2060949Z Entering 'third_party/cpuinfo' 2025-12-04T10:25:37.2092859Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-12-04T10:25:37.2116089Z Entering 'third_party/cudnn_frontend' 2025-12-04T10:25:37.2150273Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-12-04T10:25:37.2172087Z Entering 'third_party/cutlass' 2025-12-04T10:25:37.2202278Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-12-04T10:25:37.2233533Z Entering 'third_party/fbgemm' 2025-12-04T10:25:37.2265562Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-12-04T10:25:37.2287958Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T10:25:37.2318187Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-12-04T10:25:37.2339721Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T10:25:37.2367900Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-12-04T10:25:37.2397404Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T10:25:37.2427740Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-12-04T10:25:37.2449214Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T10:25:37.2479623Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-12-04T10:25:37.2508880Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T10:25:37.2538839Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-12-04T10:25:37.2559896Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T10:25:37.2595922Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-12-04T10:25:37.2617431Z Entering 'third_party/fbgemm/external/json' 2025-12-04T10:25:37.2648303Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-12-04T10:25:37.2674692Z Entering 'third_party/flash-attention' 2025-12-04T10:25:37.2706126Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-12-04T10:25:37.2728417Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T10:25:37.2759360Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-12-04T10:25:37.2787450Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T10:25:37.2817453Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-12-04T10:25:37.2848651Z Entering 'third_party/flatbuffers' 2025-12-04T10:25:37.2882785Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-12-04T10:25:37.2908203Z Entering 'third_party/fmt' 2025-12-04T10:25:37.2938747Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-12-04T10:25:37.2960572Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T10:25:37.2991697Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-12-04T10:25:37.3015973Z Entering 'third_party/gloo' 2025-12-04T10:25:37.3047012Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-12-04T10:25:37.3069069Z Entering 'third_party/googletest' 2025-12-04T10:25:37.3100810Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:25:37.3121978Z Entering 'third_party/ideep' 2025-12-04T10:25:37.3152718Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-12-04T10:25:37.3171106Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T10:25:37.3200775Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-12-04T10:25:37.3229982Z Entering 'third_party/ittapi' 2025-12-04T10:25:37.3260943Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-12-04T10:25:37.3282312Z Entering 'third_party/kineto' 2025-12-04T10:25:37.3312578Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-12-04T10:25:37.3332504Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T10:25:37.3362955Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-12-04T10:25:37.3382523Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T10:25:37.3411312Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-12-04T10:25:37.3433711Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T10:25:37.3464528Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-12-04T10:25:37.3487167Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T10:25:37.3518025Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-12-04T10:25:37.3539829Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T10:25:37.3570039Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-12-04T10:25:37.3589862Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T10:25:37.3621198Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-12-04T10:25:37.3645016Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T10:25:37.3677799Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-12-04T10:25:37.3699638Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T10:25:37.3729942Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:25:37.3751765Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T10:25:37.3781502Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-12-04T10:25:37.3803479Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T10:25:37.3833257Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-12-04T10:25:37.3853616Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T10:25:37.3883722Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T10:25:37.3903856Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:25:37.3933027Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T10:25:37.3955913Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:25:37.3988366Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T10:25:37.4015989Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T10:25:37.4046587Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-12-04T10:25:37.4068041Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T10:25:37.4098930Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-12-04T10:25:37.4122989Z Entering 'third_party/kleidiai' 2025-12-04T10:25:37.4155851Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-12-04T10:25:37.4177308Z Entering 'third_party/mimalloc' 2025-12-04T10:25:37.4206458Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-12-04T10:25:37.4228930Z Entering 'third_party/nlohmann' 2025-12-04T10:25:37.4258275Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-12-04T10:25:37.4281551Z Entering 'third_party/onnx' 2025-12-04T10:25:37.4311777Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-12-04T10:25:37.4349921Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T10:25:37.4380543Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-12-04T10:25:37.4405885Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T10:25:37.4437527Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-12-04T10:25:37.4459392Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T10:25:37.4489645Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-12-04T10:25:37.4510483Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T10:25:37.4540841Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:25:37.4561887Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T10:25:37.4592103Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-12-04T10:25:37.4612428Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T10:25:37.4642147Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-12-04T10:25:37.4663991Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T10:25:37.4694154Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-12-04T10:25:37.4716348Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T10:25:37.4747758Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-12-04T10:25:37.4768884Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T10:25:37.4799911Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T10:25:37.4818913Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:25:37.4849325Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T10:25:37.4872672Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:25:37.4902730Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T10:25:37.4925614Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T10:25:37.4956637Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-12-04T10:25:37.4998894Z Entering 'third_party/pocketfft' 2025-12-04T10:25:37.5030282Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-12-04T10:25:37.5051071Z Entering 'third_party/protobuf' 2025-12-04T10:25:37.5082519Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-12-04T10:25:37.5104962Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T10:25:37.5135641Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-12-04T10:25:37.5157040Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T10:25:37.5191622Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:25:37.5215703Z Entering 'third_party/psimd' 2025-12-04T10:25:37.5246651Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-12-04T10:25:37.5268415Z Entering 'third_party/pthreadpool' 2025-12-04T10:25:37.5299036Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-12-04T10:25:37.5320127Z Entering 'third_party/pybind11' 2025-12-04T10:25:37.5350441Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-12-04T10:25:37.5372106Z Entering 'third_party/python-peachpy' 2025-12-04T10:25:37.5401861Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-12-04T10:25:37.5424914Z Entering 'third_party/sleef' 2025-12-04T10:25:37.5456087Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-12-04T10:25:37.5479476Z Entering 'third_party/tensorpipe' 2025-12-04T10:25:37.5510702Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-12-04T10:25:37.5530542Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T10:25:37.5560204Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:25:37.5581466Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T10:25:37.5610797Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-12-04T10:25:37.5631475Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T10:25:37.5660130Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-12-04T10:25:37.5681773Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T10:25:37.5713149Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-12-04T10:25:37.5734281Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T10:25:37.5764591Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-12-04T10:25:37.5812272Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.5842898Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.5874430Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.5902853Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.5929862Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.5955885Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.5985426Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6012692Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6039031Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6066213Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6094795Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6123650Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6151802Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6178338Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6204326Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6231164Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6258816Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6286654Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6314150Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6342022Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6368373Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6395909Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6426135Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6452488Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6481625Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6507498Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6533325Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6560232Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6587021Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6612617Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6638010Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6665107Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6693103Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6721785Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6749470Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6778524Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6805884Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6834479Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6861338Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6887713Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6914708Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6941447Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.6974246Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7002743Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7030218Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7055769Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7081661Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7109973Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7136047Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7160762Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7186511Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7213888Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7239487Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7264000Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7291244Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7317687Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7342467Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7368078Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7392743Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7420122Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7445642Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7472774Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7498678Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7524524Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7551140Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7582839Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7609095Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7635256Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7662835Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7689928Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7714631Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7740250Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7764927Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7790137Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7814703Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7840310Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7866049Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7893476Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7922303Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7948554Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.7975897Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:25:37.8090816Z A job completed hook has been configured by the self-hosted runner administrator 2025-12-04T10:25:37.8105417Z ##[group]Run '/home/ec2-user/runner-scripts/after_job.sh' 2025-12-04T10:25:37.8112454Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:25:37.8112846Z ##[endgroup] 2025-12-04T10:25:45.9501960Z Cleaning up orphan processes