2025-12-04T08:55:53.5456040Z Current runner version: '2.330.0' 2025-12-04T08:55:53.5461727Z Runner name: 'i-077765b42bd7d5290' 2025-12-04T08:55:53.5462921Z Runner group name: 'Default' 2025-12-04T08:55:53.5463735Z Machine name: 'ip-10-1-22-180' 2025-12-04T08:55:53.5466721Z ##[group]GITHUB_TOKEN Permissions 2025-12-04T08:55:53.5468849Z Contents: read 2025-12-04T08:55:53.5469370Z Metadata: read 2025-12-04T08:55:53.5469934Z ##[endgroup] 2025-12-04T08:55:53.5471902Z Secret source: Actions 2025-12-04T08:55:53.5472612Z Prepare workflow directory 2025-12-04T08:55:53.5947768Z Prepare all required actions 2025-12-04T08:55:53.5980175Z Getting action download info 2025-12-04T08:55:53.9687238Z Download action repository 'pytorch/test-infra@main' (SHA:39aa74d619174326f4e2fb0e216151c2f29d9ffd) 2025-12-04T08:55:56.4928985Z Download action repository 'pytorch/pytorch@main' (SHA:eabb7ad2128580ef674446027b95bcf4e21e8df3) 2025-12-04T08:56:13.0412902Z Download action repository 'actions/setup-python@a26af69be951a213d495a4c3e4e4022e16d87065' (SHA:a26af69be951a213d495a4c3e4e4022e16d87065) 2025-12-04T08:56:13.4457309Z Download action repository 'aws-actions/configure-aws-credentials@ececac1a45f3b08a01d2dd070d28d111c5fe6722' (SHA:ececac1a45f3b08a01d2dd070d28d111c5fe6722) 2025-12-04T08:56:13.7367114Z Download action repository 'aws-actions/amazon-ecr-login@062b18b96a7aff071d4dc91bc00c4c1a7945b076' (SHA:062b18b96a7aff071d4dc91bc00c4c1a7945b076) 2025-12-04T08:56:13.9348238Z Download action repository 'seemethere/download-artifact-s3@1da556a7aa0a088e3153970611f6c432d58e80e6' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-12-04T08:56:14.1877520Z Download action repository 'seemethere/upload-artifact-s3@baba72d0712b404f646cebe0730933554ebce96a' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-12-04T08:56:14.5021421Z Getting action download info 2025-12-04T08:56:14.6194994Z Download action repository 'actions/checkout@v4' (SHA:34e114876b0b11c390a56381ad16ebd13914f8d5) 2025-12-04T08:56:14.8931580Z Getting action download info 2025-12-04T08:56:15.0348396Z Download action repository 'nick-fields/retry@v3.0.0' (SHA:7152eba30c6575329ac0576536151aca5a72780e) 2025-12-04T08:56:15.2645973Z Getting action download info 2025-12-04T08:56:15.4138898Z Download action repository 'nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482' (SHA:3e91a01664abd3c5cd539100d10d33b9c5b68482) 2025-12-04T08:56:15.6024498Z Getting action download info 2025-12-04T08:56:15.7410250Z Uses: pytorch/pytorch/.github/workflows/_linux-test.yml@refs/heads/main (ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32) 2025-12-04T08:56:15.7414146Z ##[group] Inputs 2025-12-04T08:56:15.7414450Z build-environment: linux-jammy-cuda12.8-py3.10-gcc11 2025-12-04T08:56:15.7421612Z test-matrix: {"include": [{"config": "default", "shard": 1, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 1, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 2, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 2, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 3, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 3, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 4, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 4, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 5, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 5, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "pr_time_benchmarks", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "pr_time_benchmarks", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "libtorch_agnostic_targetting", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "libtorch_agnostic_targetting", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}]} 2025-12-04T08:56:15.7429188Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:56:15.7429802Z sync-tag: 2025-12-04T08:56:15.7430483Z timeout-minutes: 360 2025-12-04T08:56:15.7430692Z use-gha: 2025-12-04T08:56:15.7431020Z dashboard-tag: 2025-12-04T08:56:15.7431326Z s3-bucket: gha-artifacts 2025-12-04T08:56:15.7431650Z aws-role-to-assume: 2025-12-04T08:56:15.7432297Z disable-monitor: false 2025-12-04T08:56:15.7432686Z monitor-log-interval: 5 2025-12-04T08:56:15.7433011Z monitor-data-collect-interval: 1 2025-12-04T08:56:15.7433341Z ##[endgroup] 2025-12-04T08:56:15.7433931Z Complete job name: linux-jammy-cuda12.8-py3.10-gcc11 / test (default, 4, 5, lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests) 2025-12-04T08:56:15.8028048Z A job started hook has been configured by the self-hosted runner administrator 2025-12-04T08:56:15.8120794Z ##[group]Run '/home/ec2-user/runner-scripts/before_job.sh' 2025-12-04T08:56:15.8130577Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:56:15.8131126Z ##[endgroup] 2025-12-04T08:56:17.0609960Z Runner Type: lf.linux.g6.4xlarge.experimental.nvidia.gpu 2025-12-04T08:56:17.0610463Z Instance Type: g6.4xlarge 2025-12-04T08:56:17.0610664Z AMI Name: unknown 2025-12-04T08:56:17.0643049Z AMI ID: ami-08982f1c5bf93d976 2025-12-04T08:56:21.7686014Z ##[group]Run pytorch/test-infra/.github/actions/setup-ssh@main 2025-12-04T08:56:21.7686360Z with: 2025-12-04T08:56:21.7686861Z github-secret: *** 2025-12-04T08:56:21.7687385Z instructions: All testing is done inside the container, to start an interactive session run: docker exec -it $(docker container ps --format '{{.ID}}') bash 2025-12-04T08:56:21.7687936Z activate-with-label: false 2025-12-04T08:56:21.7688147Z label: with-ssh 2025-12-04T08:56:21.7688323Z remove-existing-keys: true 2025-12-04T08:56:21.7688530Z fail-silently: true 2025-12-04T08:56:21.7688700Z env: 2025-12-04T08:56:21.7688847Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:56:21.7689044Z ##[endgroup] 2025-12-04T08:56:21.8749759Z Please see https://github.com/pytorch/pytorch/wiki/Debugging-using-with-ssh-for-Github-Actions for more info. 2025-12-04T08:56:21.8750828Z Not on pull request and ciflow reference could not be extracted, skipping adding ssh keys 2025-12-04T08:56:21.8878839Z ##[group]Run pytorch/pytorch/.github/actions/checkout-pytorch@main 2025-12-04T08:56:21.8879404Z with: 2025-12-04T08:56:21.8879579Z no-sudo: true 2025-12-04T08:56:21.8879760Z submodules: recursive 2025-12-04T08:56:21.8879961Z fetch-depth: 0 2025-12-04T08:56:21.8880139Z env: 2025-12-04T08:56:21.8880301Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:56:21.8880497Z ##[endgroup] 2025-12-04T08:56:21.8942135Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T08:56:21.8942860Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T08:56:21.8954004Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:56:21.8954296Z env: 2025-12-04T08:56:21.8954481Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:56:21.8954713Z ##[endgroup] 2025-12-04T08:56:21.9030892Z ##[group]Run # Use all available CPUs for fetching 2025-12-04T08:56:21.9031218Z # Use all available CPUs for fetching 2025-12-04T08:56:21.9031469Z cd "${GITHUB_WORKSPACE}" 2025-12-04T08:56:21.9031719Z git config --global fetch.parallel 0 2025-12-04T08:56:21.9032009Z git config --global submodule.fetchJobs 0 2025-12-04T08:56:21.9032250Z  2025-12-04T08:56:21.9052113Z # Clean workspace. The default checkout action should also do this, but 2025-12-04T08:56:21.9052517Z # do it here as well just in case 2025-12-04T08:56:21.9052758Z if [[ -d .git ]]; then 2025-12-04T08:56:21.9052977Z  if [ -z "${NO_SUDO}" ]; then 2025-12-04T08:56:21.9053268Z  sudo git clean -ffdx 2025-12-04T08:56:21.9053475Z  else 2025-12-04T08:56:21.9053646Z  git clean -ffdx 2025-12-04T08:56:21.9053838Z  fi 2025-12-04T08:56:21.9053999Z fi 2025-12-04T08:56:21.9060827Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:56:21.9061136Z env: 2025-12-04T08:56:21.9061331Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:56:21.9061566Z NO_SUDO: true 2025-12-04T08:56:21.9061727Z ##[endgroup] 2025-12-04T08:56:21.9171325Z ##[group]Run actions/checkout@v4 2025-12-04T08:56:21.9171552Z with: 2025-12-04T08:56:21.9171749Z ref: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T08:56:21.9172010Z fetch-depth: 0 2025-12-04T08:56:21.9172188Z submodules: recursive 2025-12-04T08:56:21.9172384Z show-progress: false 2025-12-04T08:56:21.9172586Z repository: pytorch/pytorch 2025-12-04T08:56:21.9172924Z token: *** 2025-12-04T08:56:21.9173093Z ssh-strict: true 2025-12-04T08:56:21.9173356Z ssh-user: git 2025-12-04T08:56:21.9173542Z persist-credentials: true 2025-12-04T08:56:21.9173738Z clean: true 2025-12-04T08:56:21.9173923Z sparse-checkout-cone-mode: true 2025-12-04T08:56:21.9174143Z fetch-tags: false 2025-12-04T08:56:21.9174310Z lfs: false 2025-12-04T08:56:21.9174481Z set-safe-directory: true 2025-12-04T08:56:21.9174682Z env: 2025-12-04T08:56:21.9174855Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:56:21.9175050Z ##[endgroup] 2025-12-04T08:56:22.0160853Z Syncing repository: pytorch/pytorch 2025-12-04T08:56:22.0161978Z ##[group]Getting Git version info 2025-12-04T08:56:22.0162372Z Working directory is '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-12-04T08:56:22.0162878Z [command]/usr/bin/git version 2025-12-04T08:56:22.0163081Z git version 2.50.1 2025-12-04T08:56:22.0174440Z ##[endgroup] 2025-12-04T08:56:22.0183214Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/4a69c947-5119-4a9b-a55f-05f1817ed2e8/.gitconfig' 2025-12-04T08:56:22.0201158Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/4a69c947-5119-4a9b-a55f-05f1817ed2e8' before making global git config changes 2025-12-04T08:56:22.0202080Z Adding repository directory to the temporary git global config as a safe directory 2025-12-04T08:56:22.0205929Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T08:56:22.0251223Z Deleting the contents of '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-12-04T08:56:22.0254297Z ##[group]Initializing the repository 2025-12-04T08:56:22.0257700Z [command]/usr/bin/git init /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T08:56:22.0294936Z hint: Using 'master' as the name for the initial branch. This default branch name 2025-12-04T08:56:22.0295587Z hint: is subject to change. To configure the initial branch name to use in all 2025-12-04T08:56:22.0296119Z hint: of your new repositories, which will suppress this warning, call: 2025-12-04T08:56:22.0296489Z hint: 2025-12-04T08:56:22.0296761Z hint: git config --global init.defaultBranch 2025-12-04T08:56:22.0297081Z hint: 2025-12-04T08:56:22.0297380Z hint: Names commonly chosen instead of 'master' are 'main', 'trunk' and 2025-12-04T08:56:22.0297895Z hint: 'development'. The just-created branch can be renamed via this command: 2025-12-04T08:56:22.0298267Z hint: 2025-12-04T08:56:22.0298476Z hint: git branch -m 2025-12-04T08:56:22.0298708Z hint: 2025-12-04T08:56:22.0298988Z hint: Disable this message with "git config set advice.defaultBranchName false" 2025-12-04T08:56:22.0299493Z Initialized empty Git repository in /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/ 2025-12-04T08:56:22.0304804Z [command]/usr/bin/git remote add origin https://github.com/pytorch/pytorch 2025-12-04T08:56:22.0332330Z ##[endgroup] 2025-12-04T08:56:22.0332732Z ##[group]Disabling automatic garbage collection 2025-12-04T08:56:22.0335686Z [command]/usr/bin/git config --local gc.auto 0 2025-12-04T08:56:22.0362928Z ##[endgroup] 2025-12-04T08:56:22.0363297Z ##[group]Setting up auth 2025-12-04T08:56:22.0368175Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-12-04T08:56:22.0395090Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-12-04T08:56:22.0723656Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-12-04T08:56:22.0752332Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-12-04T08:56:22.1067971Z [command]/usr/bin/git config --local --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:56:22.1094254Z [command]/usr/bin/git submodule foreach --recursive git config --local --show-origin --name-only --get-regexp remote.origin.url 2025-12-04T08:56:22.1415937Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-12-04T08:56:22.1470162Z ##[endgroup] 2025-12-04T08:56:22.1470499Z ##[group]Fetching the repository 2025-12-04T08:56:22.1477272Z [command]/usr/bin/git -c protocol.version=2 fetch --prune --no-recurse-submodules origin +refs/heads/*:refs/remotes/origin/* +refs/tags/*:refs/tags/* 2025-12-04T08:57:03.3591302Z From https://github.com/pytorch/pytorch 2025-12-04T08:57:03.3591962Z * [new branch] 2.6.0.dev20241004+ -> origin/2.6.0.dev20241004+ 2025-12-04T08:57:03.3592432Z * [new branch] 2.9.1 -> origin/2.9.1 2025-12-04T08:57:03.3592967Z * [new branch] AaronWang04_addmmfusion_perftest -> origin/AaronWang04_addmmfusion_perftest 2025-12-04T08:57:03.3593522Z * [new branch] Flamefire-patch-1 -> origin/Flamefire-patch-1 2025-12-04T08:57:03.3594926Z * [new branch] HDCharles-2.6.0-release-notes -> origin/HDCharles-2.6.0-release-notes 2025-12-04T08:57:03.3596432Z * [new branch] HOPrintFunc -> origin/HOPrintFunc 2025-12-04T08:57:03.3599105Z * [new branch] IvanKobzarev/stack/1 -> origin/IvanKobzarev/stack/1 2025-12-04T08:57:03.3601644Z * [new branch] NicoshevSVE128 -> origin/NicoshevSVE128 2025-12-04T08:57:03.3602942Z * [new branch] PR-AOTInductorNoneBug -> origin/PR-AOTInductorNoneBug 2025-12-04T08:57:03.3605062Z * [new branch] PR-AOTInductorNoneBugFix -> origin/PR-AOTInductorNoneBugFix 2025-12-04T08:57:03.3606954Z * [new branch] PR-FixConfigsIssue -> origin/PR-FixConfigsIssue 2025-12-04T08:57:03.3608460Z * [new branch] PR-NoneBugFix-viable -> origin/PR-NoneBugFix-viable 2025-12-04T08:57:03.3610087Z * [new branch] PR-ResetToZero -> origin/PR-ResetToZero 2025-12-04T08:57:03.3611740Z * [new branch] Update-Flash-Packaging -> origin/Update-Flash-Packaging 2025-12-04T08:57:03.3613403Z * [new branch] VLA_exp -> origin/VLA_exp 2025-12-04T08:57:03.3615448Z * [new branch] activation_bench -> origin/activation_bench 2025-12-04T08:57:03.3617548Z * [new branch] addmm-heuristic -> origin/addmm-heuristic 2025-12-04T08:57:03.3619676Z * [new branch] adi/onednn_aarch64 -> origin/adi/onednn_aarch64 2025-12-04T08:57:03.3621288Z * [new branch] adi/test -> origin/adi/test 2025-12-04T08:57:03.3622841Z * [new branch] adi/test_bgemm -> origin/adi/test_bgemm 2025-12-04T08:57:03.3624421Z * [new branch] adi/test_m8g -> origin/adi/test_m8g 2025-12-04T08:57:03.3626016Z * [new branch] adi/test_onednn -> origin/adi/test_onednn 2025-12-04T08:57:03.3627619Z * [new branch] adi/test_onednn_v3.9 -> origin/adi/test_onednn_v3.9 2025-12-04T08:57:03.3629218Z * [new branch] adi/test_presve_change -> origin/adi/test_presve_change 2025-12-04T08:57:03.3630767Z * [new branch] adi/test_timm -> origin/adi/test_timm 2025-12-04T08:57:03.3632731Z * [new branch] adi/testpresve_change -> origin/adi/testpresve_change 2025-12-04T08:57:03.3635707Z * [new branch] aditew01/test/vec_bf16 -> origin/aditew01/test/vec_bf16 2025-12-04T08:57:03.3637270Z * [new branch] ah-globalfeedback-hook -> origin/ah-globalfeedback-hook 2025-12-04T08:57:03.3639208Z * [new branch] albanD-patch-1 -> origin/albanD-patch-1 2025-12-04T08:57:03.3640596Z * [new branch] also-surround-shimh -> origin/also-surround-shimh 2025-12-04T08:57:03.3642951Z * [new branch] angelayi/aot_compile -> origin/angelayi/aot_compile 2025-12-04T08:57:03.3644539Z * [new branch] angelayi/aoti_additional_files -> origin/angelayi/aoti_additional_files 2025-12-04T08:57:03.3646073Z * [new branch] angelayi/benchmark -> origin/angelayi/benchmark 2025-12-04T08:57:03.3647716Z * [new branch] angelayi/change_pytree_serialization -> origin/angelayi/change_pytree_serialization 2025-12-04T08:57:03.3649201Z * [new branch] angelayi/cpp_loader -> origin/angelayi/cpp_loader 2025-12-04T08:57:03.3650815Z * [new branch] angelayi/inductor_const -> origin/angelayi/inductor_const 2025-12-04T08:57:03.3652333Z * [new branch] angelayi/lstm -> origin/angelayi/lstm 2025-12-04T08:57:03.3654544Z * [new branch] angelayi/no_so_weight -> origin/angelayi/no_so_weight 2025-12-04T08:57:03.3656479Z * [new branch] angelayi/scan_layers -> origin/angelayi/scan_layers 2025-12-04T08:57:03.3658130Z * [new branch] angelayi/side_eff -> origin/angelayi/side_eff 2025-12-04T08:57:03.3659802Z * [new branch] angelayi/state_dict -> origin/angelayi/state_dict 2025-12-04T08:57:03.3661505Z * [new branch] angelayi/symint_input -> origin/angelayi/symint_input 2025-12-04T08:57:03.3663349Z * [new branch] angelayi/symm_mem -> origin/angelayi/symm_mem 2025-12-04T08:57:03.3664640Z * [new branch] angelayi/test_cpp -> origin/angelayi/test_cpp 2025-12-04T08:57:03.3666432Z * [new branch] angelayi/torch_size -> origin/angelayi/torch_size 2025-12-04T08:57:03.3668017Z * [new branch] annotate_assert -> origin/annotate_assert 2025-12-04T08:57:03.3669729Z * [new branch] annotate_fallback_kernel -> origin/annotate_fallback_kernel 2025-12-04T08:57:03.3671375Z * [new branch] annotation_deepcopy -> origin/annotation_deepcopy 2025-12-04T08:57:03.3673006Z * [new branch] annotation_dynamo -> origin/annotation_dynamo 2025-12-04T08:57:03.3674627Z * [new branch] aot_eager_stack_trace -> origin/aot_eager_stack_trace 2025-12-04T08:57:03.3676271Z * [new branch] aoti-cuda-alloc -> origin/aoti-cuda-alloc 2025-12-04T08:57:03.3677898Z * [new branch] aoti_const_device -> origin/aoti_const_device 2025-12-04T08:57:03.3679506Z * [new branch] aoti_fqn_name_interface -> origin/aoti_fqn_name_interface 2025-12-04T08:57:03.3681099Z * [new branch] aoti_package_weights_binary -> origin/aoti_package_weights_binary 2025-12-04T08:57:03.3682716Z * [new branch] aoti_target_windows -> origin/aoti_target_windows 2025-12-04T08:57:03.3685493Z * [new branch] arsh/feat/inductor_check_profiling -> origin/arsh/feat/inductor_check_profiling 2025-12-04T08:57:03.3687185Z * [new branch] async_tp -> origin/async_tp 2025-12-04T08:57:03.3688852Z * [new branch] atalman-inductor-perf-cu124 -> origin/atalman-inductor-perf-cu124 2025-12-04T08:57:03.3690508Z * [new branch] atalman-inductor-perf-cu124.1 -> origin/atalman-inductor-perf-cu124.1 2025-12-04T08:57:03.3692181Z * [new branch] atalman-patch-2 -> origin/atalman-patch-2 2025-12-04T08:57:03.3694047Z * [new branch] atalman-patch-3 -> origin/atalman-patch-3 2025-12-04T08:57:03.3695702Z * [new branch] atalman-patch-4 -> origin/atalman-patch-4 2025-12-04T08:57:03.3697414Z * [new branch] atalman-patch-5 -> origin/atalman-patch-5 2025-12-04T08:57:03.3699099Z * [new branch] atalman-patch-6 -> origin/atalman-patch-6 2025-12-04T08:57:03.3700728Z * [new branch] atalman-patch-7 -> origin/atalman-patch-7 2025-12-04T08:57:03.3702426Z * [new branch] atalman-patch-8 -> origin/atalman-patch-8 2025-12-04T08:57:03.3704058Z * [new branch] atalman_inductor_2.3.1 -> origin/atalman_inductor_2.3.1 2025-12-04T08:57:03.3706093Z * [new branch] atalman_inductor_2.4.0 -> origin/atalman_inductor_2.4.0 2025-12-04T08:57:03.3707797Z * [new branch] atalman_inductor_2.4.x -> origin/atalman_inductor_2.4.x 2025-12-04T08:57:03.3709485Z * [new branch] attention_benchmarking_clean -> origin/attention_benchmarking_clean 2025-12-04T08:57:03.3711622Z * [new branch] bahuang/dt_fix_scalar_add -> origin/bahuang/dt_fix_scalar_add 2025-12-04T08:57:03.3713216Z * [new branch] bahuang/fix_debug_mode -> origin/bahuang/fix_debug_mode 2025-12-04T08:57:03.3714701Z * [new branch] bahuang/fix_expand -> origin/bahuang/fix_expand 2025-12-04T08:57:03.3716271Z * [new branch] bahuang/test -> origin/bahuang/test 2025-12-04T08:57:03.3718609Z * [new branch] base/1.5 -> origin/base/1.5 2025-12-04T08:57:03.3720490Z * [new branch] batching_sdpa_efficient_attention -> origin/batching_sdpa_efficient_attention 2025-12-04T08:57:03.3722024Z * [new branch] bench_scaled_mm_ops -> origin/bench_scaled_mm_ops 2025-12-04T08:57:03.3723788Z * [new branch] benchmark-updates -> origin/benchmark-updates 2025-12-04T08:57:03.3725392Z * [new branch] benchmarking-script -> origin/benchmarking-script 2025-12-04T08:57:03.3727592Z * [new branch] bertmaher/pinbump26 -> origin/bertmaher/pinbump26 2025-12-04T08:57:03.3729730Z * [new branch] bertrand/cutlass -> origin/bertrand/cutlass 2025-12-04T08:57:03.3731875Z * [new branch] bf/bug-static-input -> origin/bf/bug-static-input 2025-12-04T08:57:03.3733513Z * [new branch] bf/cg-backend -> origin/bf/cg-backend 2025-12-04T08:57:03.3735071Z * [new branch] bf/cg-nccl-test -> origin/bf/cg-nccl-test 2025-12-04T08:57:03.3736593Z * [new branch] bf/cg-remove-check -> origin/bf/cg-remove-check 2025-12-04T08:57:03.3738237Z * [new branch] bf/clean-torchbench-hf -> origin/bf/clean-torchbench-hf 2025-12-04T08:57:03.3739756Z * [new branch] bf/combo-debug-log -> origin/bf/combo-debug-log 2025-12-04T08:57:03.3741331Z * [new branch] bf/cudagraph -> origin/bf/cudagraph 2025-12-04T08:57:03.3743371Z * [new branch] bf/cudagraph-disable-input-mutation -> origin/bf/cudagraph-disable-input-mutation 2025-12-04T08:57:03.3745166Z * [new branch] bf/cudagraph-enable-input-mutation-support-benchmark -> origin/bf/cudagraph-enable-input-mutation-support-benchmark 2025-12-04T08:57:03.3746402Z * [new branch] bf/cudagraph-partition -> origin/bf/cudagraph-partition 2025-12-04T08:57:03.3748187Z * [new branch] bf/donated-buffer-bench -> origin/bf/donated-buffer-bench 2025-12-04T08:57:03.3749859Z * [new branch] bf/dynamo-partition -> origin/bf/dynamo-partition 2025-12-04T08:57:03.3751447Z * [new branch] bf/lite -> origin/bf/lite 2025-12-04T08:57:03.3753097Z * [new branch] bf/pa-non-divisible -> origin/bf/pa-non-divisible 2025-12-04T08:57:03.3754853Z * [new branch] bf/partition-cache-free-symbols -> origin/bf/partition-cache-free-symbols 2025-12-04T08:57:03.3756548Z * [new branch] bf/partition-memory-plan -> origin/bf/partition-memory-plan 2025-12-04T08:57:03.3758168Z * [new branch] bf/partition-move-cpu -> origin/bf/partition-move-cpu 2025-12-04T08:57:03.3759874Z * [new branch] bf/partition-view-fallback -> origin/bf/partition-view-fallback 2025-12-04T08:57:03.3761471Z * [new branch] bf/remove-check-55b0c39d -> origin/bf/remove-check-55b0c39d 2025-12-04T08:57:03.3763049Z * [new branch] bf/timm-nov-26-2025 -> origin/bf/timm-nov-26-2025 2025-12-04T08:57:03.3764775Z * [new branch] bf/transformer-pin-4-57-3 -> origin/bf/transformer-pin-4-57-3 2025-12-04T08:57:03.3766406Z * [new branch] bisect_perf_hf_T5_3acc6eac492 -> origin/bisect_perf_hf_T5_3acc6eac492 2025-12-04T08:57:03.3767982Z * [new branch] bisect_perf_hf_T5_3fcf66f61fb -> origin/bisect_perf_hf_T5_3fcf66f61fb 2025-12-04T08:57:03.3769596Z * [new branch] bisect_perf_hf_T5_4009d154129 -> origin/bisect_perf_hf_T5_4009d154129 2025-12-04T08:57:03.3771175Z * [new branch] bisect_perf_hf_T5_40d0740e73d -> origin/bisect_perf_hf_T5_40d0740e73d 2025-12-04T08:57:03.3772787Z * [new branch] bisect_perf_hf_T5_5268754e -> origin/bisect_perf_hf_T5_5268754e 2025-12-04T08:57:03.3774840Z * [new branch] bisect_perf_hf_T5_7d89a8d385c -> origin/bisect_perf_hf_T5_7d89a8d385c 2025-12-04T08:57:03.3776453Z * [new branch] bisect_perf_hf_T5_b7a25c1ee7c -> origin/bisect_perf_hf_T5_b7a25c1ee7c 2025-12-04T08:57:03.3777987Z * [new branch] bisect_perf_hf_T5_c25b201583f -> origin/bisect_perf_hf_T5_c25b201583f 2025-12-04T08:57:03.3779628Z * [new branch] bisect_perf_hf_T5_c93e57efac0 -> origin/bisect_perf_hf_T5_c93e57efac0 2025-12-04T08:57:03.3781405Z * [new branch] bisect_perf_hf_T5_ca9813ea149 -> origin/bisect_perf_hf_T5_ca9813ea149 2025-12-04T08:57:03.3782740Z * [new branch] bisect_perf_hf_T5_d65f194a -> origin/bisect_perf_hf_T5_d65f194a 2025-12-04T08:57:03.3784534Z * [new branch] bisect_perf_hf_T5_da94ab0b -> origin/bisect_perf_hf_T5_da94ab0b 2025-12-04T08:57:03.3785998Z * [new branch] bisect_perf_hf_T5_da94ab0b_new -> origin/bisect_perf_hf_T5_da94ab0b_new 2025-12-04T08:57:03.3787600Z * [new branch] bisect_perf_hf_T5_db4e8a1d8a8 -> origin/bisect_perf_hf_T5_db4e8a1d8a8 2025-12-04T08:57:03.3789104Z * [new branch] bisect_perf_hf_T5_e0d97e936a2 -> origin/bisect_perf_hf_T5_e0d97e936a2 2025-12-04T08:57:03.3790782Z * [new branch] bisect_perf_hf_T5_f23621ec563 -> origin/bisect_perf_hf_T5_f23621ec563 2025-12-04T08:57:03.3793073Z * [new branch] brister/fx_device_type -> origin/brister/fx_device_type 2025-12-04T08:57:03.3794705Z * [new branch] brister/test_inductor_all_fx -> origin/brister/test_inductor_all_fx 2025-12-04T08:57:03.3796309Z * [new branch] brister/tiled_reduction_no_numel_check -> origin/brister/tiled_reduction_no_numel_check 2025-12-04T08:57:03.3797833Z * [new branch] bwd-backup -> origin/bwd-backup 2025-12-04T08:57:03.3799641Z * [new branch] c57382a49 -> origin/c57382a49 2025-12-04T08:57:03.3801194Z * [new branch] ca_0431d47eaa -> origin/ca_0431d47eaa 2025-12-04T08:57:03.3802753Z * [new branch] ca_fix_0431d47eaa -> origin/ca_fix_0431d47eaa 2025-12-04T08:57:03.3804987Z * [new branch] camyllh/test_setup_hooks_push -> origin/camyllh/test_setup_hooks_push 2025-12-04T08:57:03.3806982Z * [new branch] cccclai-patch-1 -> origin/cccclai-patch-1 2025-12-04T08:57:03.3808736Z * [new branch] cherry-pick-159969-by-pytorch_bot_bot_ -> origin/cherry-pick-159969-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3810378Z * [new branch] cherry-pick-160586-by-pytorch_bot_bot_ -> origin/cherry-pick-160586-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3812097Z * [new branch] cherry-pick-162208-by-pytorch_bot_bot_ -> origin/cherry-pick-162208-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3813850Z * [new branch] cherry-pick-163169-by-pytorch_bot_bot_ -> origin/cherry-pick-163169-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3815568Z * [new branch] cherry-pick-165086-by-pytorch_bot_bot_ -> origin/cherry-pick-165086-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3817322Z * [new branch] cherry-pick-165514-by-pytorch_bot_bot_ -> origin/cherry-pick-165514-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3818980Z * [new branch] cherry-pick-165601-by-pytorch_bot_bot_ -> origin/cherry-pick-165601-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3820601Z * [new branch] cherry-pick-165667-by-pytorch_bot_bot_ -> origin/cherry-pick-165667-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3822334Z * [new branch] cherry-pick-165815-by-pytorch_bot_bot_ -> origin/cherry-pick-165815-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3824039Z * [new branch] cherry-pick-165922-by-pytorch_bot_bot_ -> origin/cherry-pick-165922-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3825707Z * [new branch] cherry-pick-166148-by-pytorch_bot_bot_ -> origin/cherry-pick-166148-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3827292Z * [new branch] cherry-pick-166181-by-pytorch_bot_bot_ -> origin/cherry-pick-166181-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3828912Z * [new branch] cherry-pick-166404-by-pytorch_bot_bot_ -> origin/cherry-pick-166404-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3830573Z * [new branch] cherry-pick-166427-by-pytorch_bot_bot_ -> origin/cherry-pick-166427-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3832398Z * [new branch] cherry-pick-166480-by-pytorch_bot_bot_ -> origin/cherry-pick-166480-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3833725Z * [new branch] cherry-pick-166570-by-pytorch_bot_bot_ -> origin/cherry-pick-166570-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3835534Z * [new branch] cherry-pick-166993-by-pytorch_bot_bot_ -> origin/cherry-pick-166993-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3837153Z * [new branch] cherry-pick-167111-by-pytorch_bot_bot_ -> origin/cherry-pick-167111-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3838863Z * [new branch] cherry-pick-167478-by-pytorch_bot_bot_ -> origin/cherry-pick-167478-by-pytorch_bot_bot_ 2025-12-04T08:57:03.3840441Z * [new branch] cherry_pick_166036_166040 -> origin/cherry_pick_166036_166040 2025-12-04T08:57:03.3842088Z * [new branch] cherry_pick_166457 -> origin/cherry_pick_166457 2025-12-04T08:57:03.3843796Z * [new branch] cherrypick_166338 -> origin/cherrypick_166338 2025-12-04T08:57:03.3845430Z * [new branch] cherrypick_166458 -> origin/cherrypick_166458 2025-12-04T08:57:03.3847026Z * [new branch] cherrypick_166586 -> origin/cherrypick_166586 2025-12-04T08:57:03.3848620Z * [new branch] cherrypick_166956 -> origin/cherrypick_166956 2025-12-04T08:57:03.3850242Z * [new branch] ci_attn -> origin/ci_attn 2025-12-04T08:57:03.3851836Z * [new branch] codex-testing -> origin/codex-testing 2025-12-04T08:57:03.3854697Z * [new branch] codex/add-check_memory_overlap-helper-functions -> origin/codex/add-check_memory_overlap-helper-functions 2025-12-04T08:57:03.3856004Z * [new branch] codex/fix-issue-121219-in-pytorch -> origin/codex/fix-issue-121219-in-pytorch 2025-12-04T08:57:03.3858206Z * [new branch] codex/investigate-segfaults-in-get_tensor_storage_id -> origin/codex/investigate-segfaults-in-get_tensor_storage_id 2025-12-04T08:57:03.3860120Z * [new branch] codex/refactor-lintrunner-config-to-use-uv-run -> origin/codex/refactor-lintrunner-config-to-use-uv-run 2025-12-04T08:57:03.3861383Z * [new branch] compatiblpy39util -> origin/compatiblpy39util 2025-12-04T08:57:03.3863290Z * [new branch] cond_hop_device -> origin/cond_hop_device 2025-12-04T08:57:03.3865044Z * [new branch] context_test -> origin/context_test 2025-12-04T08:57:03.3867335Z * [new branch] copilot/code-style-cleanup-python-pip -> origin/copilot/code-style-cleanup-python-pip 2025-12-04T08:57:03.3869443Z * [new branch] cpio/fix_new_ami_tests -> origin/cpio/fix_new_ami_tests 2025-12-04T08:57:03.3871127Z * [new branch] cpp-docs-dependency-upgrade -> origin/cpp-docs-dependency-upgrade 2025-12-04T08:57:03.3873401Z * [new branch] csl/always_produce_xml -> origin/csl/always_produce_xml 2025-12-04T08:57:03.3875217Z * [new branch] csl/build_test_more_procs -> origin/csl/build_test_more_procs 2025-12-04T08:57:03.3876754Z * [new branch] csl/build_test_more_procs2 -> origin/csl/build_test_more_procs2 2025-12-04T08:57:03.3878299Z * [new branch] csl/clean_up -> origin/csl/clean_up 2025-12-04T08:57:03.3879928Z * [new branch] csl/fix_retry_segfault_exit -> origin/csl/fix_retry_segfault_exit 2025-12-04T08:57:03.3881402Z * [new branch] csl/katex -> origin/csl/katex 2025-12-04T08:57:03.3883241Z * [new branch] csl/larger_runner -> origin/csl/larger_runner 2025-12-04T08:57:03.3885158Z * [new branch] csl/lint_testing -> origin/csl/lint_testing 2025-12-04T08:57:03.3887067Z * [new branch] csl/lint_thing -> origin/csl/lint_thing 2025-12-04T08:57:03.3888676Z * [new branch] csl/lintrunner_stuff -> origin/csl/lintrunner_stuff 2025-12-04T08:57:03.3890405Z * [new branch] csl/manually_gen_json -> origin/csl/manually_gen_json 2025-12-04T08:57:03.3892061Z * [new branch] csl/mps_sharding -> origin/csl/mps_sharding 2025-12-04T08:57:03.3893720Z * [new branch] csl/multistage_docker -> origin/csl/multistage_docker 2025-12-04T08:57:03.3895385Z * [new branch] csl/print_timing -> origin/csl/print_timing 2025-12-04T08:57:03.3896986Z * [new branch] csl/remove_experiment -> origin/csl/remove_experiment 2025-12-04T08:57:03.3898670Z * [new branch] csl/remove_maybe_unused_var -> origin/csl/remove_maybe_unused_var 2025-12-04T08:57:03.3900401Z * [new branch] csl/remove_repo_specific_autolabel -> origin/csl/remove_repo_specific_autolabel 2025-12-04T08:57:03.3901999Z * [new branch] csl/remove_run_parallel -> origin/csl/remove_run_parallel 2025-12-04T08:57:03.3903511Z * [new branch] csl/remove_unused_vars -> origin/csl/remove_unused_vars 2025-12-04T08:57:03.3905350Z * [new branch] csl/revert_open -> origin/csl/revert_open 2025-12-04T08:57:03.3907487Z * [new branch] csl/skip_build -> origin/csl/skip_build 2025-12-04T08:57:03.3909148Z * [new branch] csl/smaller_avx_amx_runenrs -> origin/csl/smaller_avx_amx_runenrs 2025-12-04T08:57:03.3910690Z * [new branch] csl/td_job_level -> origin/csl/td_job_level 2025-12-04T08:57:03.3912318Z * [new branch] csl/test_cuda_build_large_runner -> origin/csl/test_cuda_build_large_runner 2025-12-04T08:57:03.3914054Z * [new branch] csl/test_owners_autograd_dispatch_nn -> origin/csl/test_owners_autograd_dispatch_nn 2025-12-04T08:57:03.3915621Z * [new branch] csl/test_owners_higher_confidence -> origin/csl/test_owners_higher_confidence 2025-12-04T08:57:03.3917216Z * [new branch] csl/upload_json_running -> origin/csl/upload_json_running 2025-12-04T08:57:03.3918772Z * [new branch] csl/win_sccache -> origin/csl/win_sccache 2025-12-04T08:57:03.3920345Z * [new branch] csl/xml_stuff -> origin/csl/xml_stuff 2025-12-04T08:57:03.3922003Z * [new branch] cublasrelax2 -> origin/cublasrelax2 2025-12-04T08:57:03.3923589Z * [new branch] cuda_mempool -> origin/cuda_mempool 2025-12-04T08:57:03.3925205Z * [new branch] custom_lowering_dict -> origin/custom_lowering_dict 2025-12-04T08:57:03.3927325Z * [new branch] d4l3k/debug_plane_frtrace -> origin/d4l3k/debug_plane_frtrace 2025-12-04T08:57:03.3929496Z * [new branch] daxia6/2.8o3 -> origin/daxia6/2.8o3 2025-12-04T08:57:03.3931059Z * [new branch] debug-guard -> origin/debug-guard 2025-12-04T08:57:03.3932742Z * [new branch] delete-quant-docs -> origin/delete-quant-docs 2025-12-04T08:57:03.3937649Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 2025-12-04T08:57:03.3939308Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.1 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.1 2025-12-04T08:57:03.3941380Z * [new branch] desertfire/test_cpp_wrapper -> origin/desertfire/test_cpp_wrapper 2025-12-04T08:57:03.3943046Z * [new branch] desertfire/triton-cpu-for-aarch64 -> origin/desertfire/triton-cpu-for-aarch64 2025-12-04T08:57:03.3945793Z * [new branch] dev/dhruva/flex_attn_opt -> origin/dev/dhruva/flex_attn_opt 2025-12-04T08:57:03.3948083Z * [new branch] dev/joona/MPSNDArrayAdd -> origin/dev/joona/MPSNDArrayAdd 2025-12-04T08:57:03.3949718Z * [new branch] dev/joona/Unranked -> origin/dev/joona/Unranked 2025-12-04T08:57:03.3951701Z * [new branch] dev/joona/cat -> origin/dev/joona/cat 2025-12-04T08:57:03.3953287Z * [new branch] dev/joona/embeddingbag -> origin/dev/joona/embeddingbag 2025-12-04T08:57:03.3955146Z * [new branch] dev/joona/fix_sdpa_memtest -> origin/dev/joona/fix_sdpa_memtest 2025-12-04T08:57:03.3956916Z * [new branch] dev/joona/getTensorsString -> origin/dev/joona/getTensorsString 2025-12-04T08:57:03.3958762Z * [new branch] dev/joona/mps_linear_macos14 -> origin/dev/joona/mps_linear_macos14 2025-12-04T08:57:03.3960855Z * [new branch] dev/joona/scalar_clamp -> origin/dev/joona/scalar_clamp 2025-12-04T08:57:03.3962952Z * [new branch] dev/joona/sdpa -> origin/dev/joona/sdpa 2025-12-04T08:57:03.3965613Z * [new branch] dev/joona/sdpa_api -> origin/dev/joona/sdpa_api 2025-12-04T08:57:03.3967384Z * [new branch] dev/joona/type_inf -> origin/dev/joona/type_inf 2025-12-04T08:57:03.3969194Z * [new branch] dev/joona/ulpAssertClose -> origin/dev/joona/ulpAssertClose 2025-12-04T08:57:03.3970964Z * [new branch] dev/joona/upsize3d -> origin/dev/joona/upsize3d 2025-12-04T08:57:03.3972513Z * [new branch] disp_counter -> origin/disp_counter 2025-12-04T08:57:03.3974321Z * [new branch] divyanshk-patch-1 -> origin/divyanshk-patch-1 2025-12-04T08:57:03.3975857Z * [new branch] docs -> origin/docs 2025-12-04T08:57:03.3977603Z * [new branch] documentation -> origin/documentation 2025-12-04T08:57:03.3979254Z * [new branch] eager_model_benchmarks -> origin/eager_model_benchmarks 2025-12-04T08:57:03.3981532Z * [new branch] embg/test_inductor_ci_control -> origin/embg/test_inductor_ci_control 2025-12-04T08:57:03.3983016Z * [new branch] embg/triton_l2_prefetch_128B -> origin/embg/triton_l2_prefetch_128B 2025-12-04T08:57:03.3984362Z * [new branch] embg/triton_l2_prefetch_256B -> origin/embg/triton_l2_prefetch_256B 2025-12-04T08:57:03.3986034Z * [new branch] eqy-patch-1 -> origin/eqy-patch-1 2025-12-04T08:57:03.3987689Z * [new branch] eqy-patch-2 -> origin/eqy-patch-2 2025-12-04T08:57:03.3989352Z * [new branch] eqy-patch-3 -> origin/eqy-patch-3 2025-12-04T08:57:03.3990953Z * [new branch] eqy-patch-4 -> origin/eqy-patch-4 2025-12-04T08:57:03.3992637Z * [new branch] eqy-patch-5 -> origin/eqy-patch-5 2025-12-04T08:57:03.3994292Z * [new branch] eqy-patch-6 -> origin/eqy-patch-6 2025-12-04T08:57:03.3996400Z * [new branch] exclamaforte/amd-ma -> origin/exclamaforte/amd-ma 2025-12-04T08:57:03.3998024Z * [new branch] exclamaforte/combo-kernels-perf-run -> origin/exclamaforte/combo-kernels-perf-run 2025-12-04T08:57:03.3999326Z * [new branch] exclamaforte/do_bench_refactor -> origin/exclamaforte/do_bench_refactor 2025-12-04T08:57:03.4001101Z * [new branch] exclamaforte/enable-mem-dep-fusion -> origin/exclamaforte/enable-mem-dep-fusion 2025-12-04T08:57:03.4002700Z * [new branch] exclamaforte/fix-exhaustive-autotuning -> origin/exclamaforte/fix-exhaustive-autotuning 2025-12-04T08:57:03.4004715Z * [new branch] exclamaforte/fix-trace-parsing-fx-svg -> origin/exclamaforte/fix-trace-parsing-fx-svg 2025-12-04T08:57:03.4006854Z * [new branch] exclamaforte/force-pointwise-cat-perf-run -> origin/exclamaforte/force-pointwise-cat-perf-run 2025-12-04T08:57:03.4008385Z * [new branch] exclamaforte/fusion-data -> origin/exclamaforte/fusion-data 2025-12-04T08:57:03.4010071Z * [new branch] exclamaforte/gemm-benchmark-run -> origin/exclamaforte/gemm-benchmark-run 2025-12-04T08:57:03.4011882Z * [new branch] exclamaforte/gemm-export-model -> origin/exclamaforte/gemm-export-model 2025-12-04T08:57:03.4013300Z * [new branch] exclamaforte/gemm-model -> origin/exclamaforte/gemm-model 2025-12-04T08:57:03.4015197Z * [new branch] exclamaforte/gemm-model-all-data-collection -> origin/exclamaforte/gemm-model-all-data-collection 2025-12-04T08:57:03.4016511Z * [new branch] exclamaforte/gemm-to-amd -> origin/exclamaforte/gemm-to-amd 2025-12-04T08:57:03.4018246Z * [new branch] exclamaforte/just-gemm-model -> origin/exclamaforte/just-gemm-model 2025-12-04T08:57:03.4019980Z * [new branch] exclamaforte/just-gemm-model-no-refactor -> origin/exclamaforte/just-gemm-model-no-refactor 2025-12-04T08:57:03.4021548Z * [new branch] exclamaforte/profile-diff-algo -> origin/exclamaforte/profile-diff-algo 2025-12-04T08:57:03.4023192Z * [new branch] exclamaforte/profiler-visualization -> origin/exclamaforte/profiler-visualization 2025-12-04T08:57:03.4024865Z * [new branch] exclamaforte/test_cpp_wrapper_mode -> origin/exclamaforte/test_cpp_wrapper_mode 2025-12-04T08:57:03.4026539Z * [new branch] exclamaforte/update-autotune-configs -> origin/exclamaforte/update-autotune-configs 2025-12-04T08:57:03.4028140Z * [new branch] exclamaforte/update-autotune-configs-2 -> origin/exclamaforte/update-autotune-configs-2 2025-12-04T08:57:03.4029685Z * [new branch] exec -> origin/exec 2025-12-04T08:57:03.4031516Z * [new branch] experimental-mosaic -> origin/experimental-mosaic 2025-12-04T08:57:03.4033187Z * [new branch] export-D61047529 -> origin/export-D61047529 2025-12-04T08:57:03.4034813Z * [new branch] export-D71412006 -> origin/export-D71412006 2025-12-04T08:57:03.4036577Z * [new branch] export-D73042989 -> origin/export-D73042989 2025-12-04T08:57:03.4038125Z * [new branch] export-D78957093 -> origin/export-D78957093 2025-12-04T08:57:03.4039689Z * [new branch] export-D78996107 -> origin/export-D78996107 2025-12-04T08:57:03.4041336Z * [new branch] export-D80823877 -> origin/export-D80823877 2025-12-04T08:57:03.4043050Z * [new branch] export-D80958642 -> origin/export-D80958642 2025-12-04T08:57:03.4044642Z * [new branch] export-D81054193 -> origin/export-D81054193 2025-12-04T08:57:03.4046181Z * [new branch] export-D81204584 -> origin/export-D81204584 2025-12-04T08:57:03.4047818Z * [new branch] export-D81429090 -> origin/export-D81429090 2025-12-04T08:57:03.4049549Z * [new branch] export-D82250826 -> origin/export-D82250826 2025-12-04T08:57:03.4051103Z * [new branch] export-D82253817 -> origin/export-D82253817 2025-12-04T08:57:03.4052728Z * [new branch] export-D83541846 -> origin/export-D83541846 2025-12-04T08:57:03.4054531Z * [new branch] export-D83627170 -> origin/export-D83627170 2025-12-04T08:57:03.4056191Z * [new branch] export-D83766701 -> origin/export-D83766701 2025-12-04T08:57:03.4057793Z * [new branch] export-D83768878 -> origin/export-D83768878 2025-12-04T08:57:03.4059467Z * [new branch] export-D83769447 -> origin/export-D83769447 2025-12-04T08:57:03.4061003Z * [new branch] export-D84089824 -> origin/export-D84089824 2025-12-04T08:57:03.4062556Z * [new branch] export-D84213020 -> origin/export-D84213020 2025-12-04T08:57:03.4064599Z * [new branch] export-D84373821 -> origin/export-D84373821 2025-12-04T08:57:03.4066210Z * [new branch] export-D84612194 -> origin/export-D84612194 2025-12-04T08:57:03.4067954Z * [new branch] export-D84890985 -> origin/export-D84890985 2025-12-04T08:57:03.4069504Z * [new branch] export-D85122326 -> origin/export-D85122326 2025-12-04T08:57:03.4071287Z * [new branch] export-D86256198 -> origin/export-D86256198 2025-12-04T08:57:03.4072856Z * [new branch] export-D86460608 -> origin/export-D86460608 2025-12-04T08:57:03.4074501Z * [new branch] export-D86474796 -> origin/export-D86474796 2025-12-04T08:57:03.4076205Z * [new branch] export-D86712396 -> origin/export-D86712396 2025-12-04T08:57:03.4077808Z * [new branch] export-D87022129 -> origin/export-D87022129 2025-12-04T08:57:03.4079528Z * [new branch] export-D87838959 -> origin/export-D87838959 2025-12-04T08:57:03.4081162Z * [new branch] export-D88319437 -> origin/export-D88319437 2025-12-04T08:57:03.4082928Z * [new branch] exported-model-train-idempotent -> origin/exported-model-train-idempotent 2025-12-04T08:57:03.4084571Z * [new branch] ezyang-titan-october -> origin/ezyang-titan-october 2025-12-04T08:57:03.4086637Z * [new branch] ezyang-titan-october2 -> origin/ezyang-titan-october2 2025-12-04T08:57:03.4088199Z * [new branch] ezyang-war -> origin/ezyang-war 2025-12-04T08:57:03.4090334Z * [new branch] ezyang/wip-aot-descriptors -> origin/ezyang/wip-aot-descriptors 2025-12-04T08:57:03.4091902Z * [new branch] fa_u8_brgemm -> origin/fa_u8_brgemm 2025-12-04T08:57:03.4094233Z * [new branch] fadeputr/sequence_fbgemm -> origin/fadeputr/sequence_fbgemm 2025-12-04T08:57:03.4095896Z * [new branch] fastmath_baseline -> origin/fastmath_baseline 2025-12-04T08:57:03.4098053Z * [new branch] fbcode/warm -> origin/fbcode/warm 2025-12-04T08:57:03.4099709Z * [new branch] fca -> origin/fca 2025-12-04T08:57:03.4101714Z * [new branch] fca2_ca5984c -> origin/fca2_ca5984c 2025-12-04T08:57:03.4103311Z * [new branch] fca5 -> origin/fca5 2025-12-04T08:57:03.4105994Z * [new branch] feature/justknobs-cpp -> origin/feature/justknobs-cpp 2025-12-04T08:57:03.4107604Z * [new branch] feature/numa-forkserver -> origin/feature/numa-forkserver 2025-12-04T08:57:03.4109564Z * [new branch] ffast_math_baseline -> origin/ffast_math_baseline 2025-12-04T08:57:03.4111129Z * [new branch] ffast_math_target -> origin/ffast_math_target 2025-12-04T08:57:03.4113304Z * [new branch] findhao/base_commit -> origin/findhao/base_commit 2025-12-04T08:57:03.4114863Z * [new branch] findhao/base_commit1 -> origin/findhao/base_commit1 2025-12-04T08:57:03.4116348Z * [new branch] findhao/multistream2 -> origin/findhao/multistream2 2025-12-04T08:57:03.4117883Z * [new branch] findhao/multistream5 -> origin/findhao/multistream5 2025-12-04T08:57:03.4119406Z * [new branch] findhao/multistream6 -> origin/findhao/multistream6 2025-12-04T08:57:03.4120905Z * [new branch] findhao/operatorbench3 -> origin/findhao/operatorbench3 2025-12-04T08:57:03.4122418Z * [new branch] findhao/operatorbench5 -> origin/findhao/operatorbench5 2025-12-04T08:57:03.4123994Z * [new branch] findhao/tritonparse -> origin/findhao/tritonparse 2025-12-04T08:57:03.4125659Z * [new branch] fix-ck-gemm-template-format -> origin/fix-ck-gemm-template-format 2025-12-04T08:57:03.4127243Z * [new branch] fix-config-ignore -> origin/fix-config-ignore 2025-12-04T08:57:03.4128832Z * [new branch] fix-dict-guard -> origin/fix-dict-guard 2025-12-04T08:57:03.4130724Z * [new branch] fix_addmm_issue -> origin/fix_addmm_issue 2025-12-04T08:57:03.4132221Z * [new branch] fix_amd_missing_cluster_dims -> origin/fix_amd_missing_cluster_dims 2025-12-04T08:57:03.4133885Z * [new branch] fix_bench_bwd_pass -> origin/fix_bench_bwd_pass 2025-12-04T08:57:03.4135497Z * [new branch] fix_mem_profiler_config -> origin/fix_mem_profiler_config 2025-12-04T08:57:03.4137123Z * [new branch] fix_nvrtc_discovery -> origin/fix_nvrtc_discovery 2025-12-04T08:57:03.4138672Z * [new branch] fix_op_runner -> origin/fix_op_runner 2025-12-04T08:57:03.4140280Z * [new branch] fix_ubn_159469 -> origin/fix_ubn_159469 2025-12-04T08:57:03.4141959Z * [new branch] fixes-triage -> origin/fixes-triage 2025-12-04T08:57:03.4143555Z * [new branch] fixflashinfer -> origin/fixflashinfer 2025-12-04T08:57:03.4145159Z * [new branch] flash_decoding_cpu -> origin/flash_decoding_cpu 2025-12-04T08:57:03.4146708Z * [new branch] flex-flash -> origin/flex-flash 2025-12-04T08:57:03.4148393Z * [new branch] flex_attention_functorch_grad -> origin/flex_attention_functorch_grad 2025-12-04T08:57:03.4149959Z * [new branch] flex_flash -> origin/flex_flash 2025-12-04T08:57:03.4152295Z * [new branch] fmassa/fix_memeff_sharding_rule -> origin/fmassa/fix_memeff_sharding_rule 2025-12-04T08:57:03.4153858Z * [new branch] fmassa/tests_comm_compute_scheduler -> origin/fmassa/tests_comm_compute_scheduler 2025-12-04T08:57:03.4155395Z * [new branch] forkserver_fix -> origin/forkserver_fix 2025-12-04T08:57:03.4157018Z * [new branch] fsdp2_trace_rules -> origin/fsdp2_trace_rules 2025-12-04T08:57:03.4158642Z * [new branch] fx_cpp -> origin/fx_cpp 2025-12-04T08:57:03.4160900Z * [new branch] fy/fix-win -> origin/fy/fix-win 2025-12-04T08:57:03.4162565Z * [new branch] galv-patch-1 -> origin/galv-patch-1 2025-12-04T08:57:03.4164874Z * [new branch] galv/cudagraphs-conditional-nodes-4 -> origin/galv/cudagraphs-conditional-nodes-4 2025-12-04T08:57:03.4166996Z * [new branch] georgehong/cmakelists-patch -> origin/georgehong/cmakelists-patch 2025-12-04T08:57:03.4170155Z * [new branch] gh/AlnisM/1/base -> origin/gh/AlnisM/1/base 2025-12-04T08:57:03.4171736Z * [new branch] gh/AlnisM/1/head -> origin/gh/AlnisM/1/head 2025-12-04T08:57:03.4174577Z * [new branch] gh/EikanWang/67/base -> origin/gh/EikanWang/67/base 2025-12-04T08:57:03.4175949Z * [new branch] gh/EikanWang/67/head -> origin/gh/EikanWang/67/head 2025-12-04T08:57:03.4178941Z * [new branch] gh/Gasoonjia/1/base -> origin/gh/Gasoonjia/1/base 2025-12-04T08:57:03.4180362Z * [new branch] gh/Gasoonjia/1/head -> origin/gh/Gasoonjia/1/head 2025-12-04T08:57:03.4183076Z * [new branch] gh/H-Huang/131/base -> origin/gh/H-Huang/131/base 2025-12-04T08:57:03.4184629Z * [new branch] gh/H-Huang/131/head -> origin/gh/H-Huang/131/head 2025-12-04T08:57:03.4186237Z * [new branch] gh/H-Huang/131/orig -> origin/gh/H-Huang/131/orig 2025-12-04T08:57:03.4188320Z * [new branch] gh/H-Huang/132/base -> origin/gh/H-Huang/132/base 2025-12-04T08:57:03.4189899Z * [new branch] gh/H-Huang/132/head -> origin/gh/H-Huang/132/head 2025-12-04T08:57:03.4191454Z * [new branch] gh/H-Huang/132/orig -> origin/gh/H-Huang/132/orig 2025-12-04T08:57:03.4193608Z * [new branch] gh/H-Huang/180/base -> origin/gh/H-Huang/180/base 2025-12-04T08:57:03.4195193Z * [new branch] gh/H-Huang/180/head -> origin/gh/H-Huang/180/head 2025-12-04T08:57:03.4196626Z * [new branch] gh/H-Huang/180/orig -> origin/gh/H-Huang/180/orig 2025-12-04T08:57:03.4198719Z * [new branch] gh/H-Huang/182/base -> origin/gh/H-Huang/182/base 2025-12-04T08:57:03.4200268Z * [new branch] gh/H-Huang/182/head -> origin/gh/H-Huang/182/head 2025-12-04T08:57:03.4201855Z * [new branch] gh/H-Huang/182/orig -> origin/gh/H-Huang/182/orig 2025-12-04T08:57:03.4204041Z * [new branch] gh/H-Huang/226/base -> origin/gh/H-Huang/226/base 2025-12-04T08:57:03.4206102Z * [new branch] gh/H-Huang/226/head -> origin/gh/H-Huang/226/head 2025-12-04T08:57:03.4207682Z * [new branch] gh/H-Huang/226/orig -> origin/gh/H-Huang/226/orig 2025-12-04T08:57:03.4209765Z * [new branch] gh/H-Huang/228/base -> origin/gh/H-Huang/228/base 2025-12-04T08:57:03.4211344Z * [new branch] gh/H-Huang/228/head -> origin/gh/H-Huang/228/head 2025-12-04T08:57:03.4212906Z * [new branch] gh/H-Huang/228/orig -> origin/gh/H-Huang/228/orig 2025-12-04T08:57:03.4215600Z * [new branch] gh/IvanKobzarev/150/base -> origin/gh/IvanKobzarev/150/base 2025-12-04T08:57:03.4217114Z * [new branch] gh/IvanKobzarev/150/head -> origin/gh/IvanKobzarev/150/head 2025-12-04T08:57:03.4218650Z * [new branch] gh/IvanKobzarev/150/orig -> origin/gh/IvanKobzarev/150/orig 2025-12-04T08:57:03.4220845Z * [new branch] gh/IvanKobzarev/157/base -> origin/gh/IvanKobzarev/157/base 2025-12-04T08:57:03.4222414Z * [new branch] gh/IvanKobzarev/157/head -> origin/gh/IvanKobzarev/157/head 2025-12-04T08:57:03.4224007Z * [new branch] gh/IvanKobzarev/157/orig -> origin/gh/IvanKobzarev/157/orig 2025-12-04T08:57:03.4226298Z * [new branch] gh/IvanKobzarev/159/base -> origin/gh/IvanKobzarev/159/base 2025-12-04T08:57:03.4227904Z * [new branch] gh/IvanKobzarev/159/head -> origin/gh/IvanKobzarev/159/head 2025-12-04T08:57:03.4229491Z * [new branch] gh/IvanKobzarev/159/orig -> origin/gh/IvanKobzarev/159/orig 2025-12-04T08:57:03.4231798Z * [new branch] gh/IvanKobzarev/162/base -> origin/gh/IvanKobzarev/162/base 2025-12-04T08:57:03.4233516Z * [new branch] gh/IvanKobzarev/162/head -> origin/gh/IvanKobzarev/162/head 2025-12-04T08:57:03.4235057Z * [new branch] gh/IvanKobzarev/162/orig -> origin/gh/IvanKobzarev/162/orig 2025-12-04T08:57:03.4237280Z * [new branch] gh/IvanKobzarev/163/base -> origin/gh/IvanKobzarev/163/base 2025-12-04T08:57:03.4238795Z * [new branch] gh/IvanKobzarev/163/head -> origin/gh/IvanKobzarev/163/head 2025-12-04T08:57:03.4240346Z * [new branch] gh/IvanKobzarev/163/orig -> origin/gh/IvanKobzarev/163/orig 2025-12-04T08:57:03.4242537Z * [new branch] gh/IvanKobzarev/166/base -> origin/gh/IvanKobzarev/166/base 2025-12-04T08:57:03.4244121Z * [new branch] gh/IvanKobzarev/166/head -> origin/gh/IvanKobzarev/166/head 2025-12-04T08:57:03.4245791Z * [new branch] gh/IvanKobzarev/166/orig -> origin/gh/IvanKobzarev/166/orig 2025-12-04T08:57:03.4247958Z * [new branch] gh/IvanKobzarev/167/base -> origin/gh/IvanKobzarev/167/base 2025-12-04T08:57:03.4249432Z * [new branch] gh/IvanKobzarev/167/head -> origin/gh/IvanKobzarev/167/head 2025-12-04T08:57:03.4251068Z * [new branch] gh/IvanKobzarev/167/orig -> origin/gh/IvanKobzarev/167/orig 2025-12-04T08:57:03.4253248Z * [new branch] gh/IvanKobzarev/168/base -> origin/gh/IvanKobzarev/168/base 2025-12-04T08:57:03.4254897Z * [new branch] gh/IvanKobzarev/168/head -> origin/gh/IvanKobzarev/168/head 2025-12-04T08:57:03.4256643Z * [new branch] gh/IvanKobzarev/168/orig -> origin/gh/IvanKobzarev/168/orig 2025-12-04T08:57:03.4258631Z * [new branch] gh/IvanKobzarev/169/base -> origin/gh/IvanKobzarev/169/base 2025-12-04T08:57:03.4260173Z * [new branch] gh/IvanKobzarev/169/head -> origin/gh/IvanKobzarev/169/head 2025-12-04T08:57:03.4261700Z * [new branch] gh/IvanKobzarev/169/orig -> origin/gh/IvanKobzarev/169/orig 2025-12-04T08:57:03.4263666Z * [new branch] gh/IvanKobzarev/170/base -> origin/gh/IvanKobzarev/170/base 2025-12-04T08:57:03.4265254Z * [new branch] gh/IvanKobzarev/170/head -> origin/gh/IvanKobzarev/170/head 2025-12-04T08:57:03.4266872Z * [new branch] gh/IvanKobzarev/170/orig -> origin/gh/IvanKobzarev/170/orig 2025-12-04T08:57:03.4269130Z * [new branch] gh/IvanKobzarev/171/base -> origin/gh/IvanKobzarev/171/base 2025-12-04T08:57:03.4270654Z * [new branch] gh/IvanKobzarev/171/head -> origin/gh/IvanKobzarev/171/head 2025-12-04T08:57:03.4272309Z * [new branch] gh/IvanKobzarev/171/orig -> origin/gh/IvanKobzarev/171/orig 2025-12-04T08:57:03.4274512Z * [new branch] gh/IvanKobzarev/172/base -> origin/gh/IvanKobzarev/172/base 2025-12-04T08:57:03.4276120Z * [new branch] gh/IvanKobzarev/172/head -> origin/gh/IvanKobzarev/172/head 2025-12-04T08:57:03.4277686Z * [new branch] gh/IvanKobzarev/172/orig -> origin/gh/IvanKobzarev/172/orig 2025-12-04T08:57:03.4279893Z * [new branch] gh/IvanKobzarev/173/base -> origin/gh/IvanKobzarev/173/base 2025-12-04T08:57:03.4281429Z * [new branch] gh/IvanKobzarev/173/head -> origin/gh/IvanKobzarev/173/head 2025-12-04T08:57:03.4283097Z * [new branch] gh/IvanKobzarev/173/orig -> origin/gh/IvanKobzarev/173/orig 2025-12-04T08:57:03.4285402Z * [new branch] gh/IvanKobzarev/174/base -> origin/gh/IvanKobzarev/174/base 2025-12-04T08:57:03.4287071Z * [new branch] gh/IvanKobzarev/174/head -> origin/gh/IvanKobzarev/174/head 2025-12-04T08:57:03.4288698Z * [new branch] gh/IvanKobzarev/174/orig -> origin/gh/IvanKobzarev/174/orig 2025-12-04T08:57:03.4290770Z * [new branch] gh/IvanKobzarev/175/base -> origin/gh/IvanKobzarev/175/base 2025-12-04T08:57:03.4292467Z * [new branch] gh/IvanKobzarev/175/head -> origin/gh/IvanKobzarev/175/head 2025-12-04T08:57:03.4294221Z * [new branch] gh/IvanKobzarev/175/orig -> origin/gh/IvanKobzarev/175/orig 2025-12-04T08:57:03.4296548Z * [new branch] gh/IvanKobzarev/176/base -> origin/gh/IvanKobzarev/176/base 2025-12-04T08:57:03.4298101Z * [new branch] gh/IvanKobzarev/176/head -> origin/gh/IvanKobzarev/176/head 2025-12-04T08:57:03.4299679Z * [new branch] gh/IvanKobzarev/176/orig -> origin/gh/IvanKobzarev/176/orig 2025-12-04T08:57:03.4302144Z * [new branch] gh/IvanKobzarev/177/base -> origin/gh/IvanKobzarev/177/base 2025-12-04T08:57:03.4303724Z * [new branch] gh/IvanKobzarev/177/head -> origin/gh/IvanKobzarev/177/head 2025-12-04T08:57:03.4305331Z * [new branch] gh/IvanKobzarev/177/orig -> origin/gh/IvanKobzarev/177/orig 2025-12-04T08:57:03.4307868Z * [new branch] gh/IvanKobzarev/178/base -> origin/gh/IvanKobzarev/178/base 2025-12-04T08:57:03.4309493Z * [new branch] gh/IvanKobzarev/178/head -> origin/gh/IvanKobzarev/178/head 2025-12-04T08:57:03.4311123Z * [new branch] gh/IvanKobzarev/178/orig -> origin/gh/IvanKobzarev/178/orig 2025-12-04T08:57:03.4313343Z * [new branch] gh/IvanKobzarev/179/base -> origin/gh/IvanKobzarev/179/base 2025-12-04T08:57:03.4314881Z * [new branch] gh/IvanKobzarev/179/head -> origin/gh/IvanKobzarev/179/head 2025-12-04T08:57:03.4316443Z * [new branch] gh/IvanKobzarev/179/orig -> origin/gh/IvanKobzarev/179/orig 2025-12-04T08:57:03.4318880Z * [new branch] gh/IvanKobzarev/180/base -> origin/gh/IvanKobzarev/180/base 2025-12-04T08:57:03.4320408Z * [new branch] gh/IvanKobzarev/180/head -> origin/gh/IvanKobzarev/180/head 2025-12-04T08:57:03.4321968Z * [new branch] gh/IvanKobzarev/180/orig -> origin/gh/IvanKobzarev/180/orig 2025-12-04T08:57:03.4324648Z * [new branch] gh/IvanKobzarev/181/base -> origin/gh/IvanKobzarev/181/base 2025-12-04T08:57:03.4326281Z * [new branch] gh/IvanKobzarev/181/head -> origin/gh/IvanKobzarev/181/head 2025-12-04T08:57:03.4327896Z * [new branch] gh/IvanKobzarev/181/orig -> origin/gh/IvanKobzarev/181/orig 2025-12-04T08:57:03.4330188Z * [new branch] gh/IvanKobzarev/182/base -> origin/gh/IvanKobzarev/182/base 2025-12-04T08:57:03.4331712Z * [new branch] gh/IvanKobzarev/182/head -> origin/gh/IvanKobzarev/182/head 2025-12-04T08:57:03.4333362Z * [new branch] gh/IvanKobzarev/182/orig -> origin/gh/IvanKobzarev/182/orig 2025-12-04T08:57:03.4335729Z * [new branch] gh/IvanKobzarev/183/base -> origin/gh/IvanKobzarev/183/base 2025-12-04T08:57:03.4337308Z * [new branch] gh/IvanKobzarev/183/head -> origin/gh/IvanKobzarev/183/head 2025-12-04T08:57:03.4338903Z * [new branch] gh/IvanKobzarev/183/orig -> origin/gh/IvanKobzarev/183/orig 2025-12-04T08:57:03.4341310Z * [new branch] gh/IvanKobzarev/184/base -> origin/gh/IvanKobzarev/184/base 2025-12-04T08:57:03.4342960Z * [new branch] gh/IvanKobzarev/184/head -> origin/gh/IvanKobzarev/184/head 2025-12-04T08:57:03.4344813Z * [new branch] gh/IvanKobzarev/184/orig -> origin/gh/IvanKobzarev/184/orig 2025-12-04T08:57:03.4347410Z * [new branch] gh/NikhilAPatel/1/base -> origin/gh/NikhilAPatel/1/base 2025-12-04T08:57:03.4349127Z * [new branch] gh/NikhilAPatel/1/head -> origin/gh/NikhilAPatel/1/head 2025-12-04T08:57:03.4351141Z * [new branch] gh/NikhilAPatel/2/base -> origin/gh/NikhilAPatel/2/base 2025-12-04T08:57:03.4352646Z * [new branch] gh/NikhilAPatel/2/head -> origin/gh/NikhilAPatel/2/head 2025-12-04T08:57:03.4355060Z * [new branch] gh/NikhilAPatel/4/base -> origin/gh/NikhilAPatel/4/base 2025-12-04T08:57:03.4356668Z * [new branch] gh/NikhilAPatel/4/head -> origin/gh/NikhilAPatel/4/head 2025-12-04T08:57:03.4358819Z * [new branch] gh/NikhilAPatel/5/base -> origin/gh/NikhilAPatel/5/base 2025-12-04T08:57:03.4360340Z * [new branch] gh/NikhilAPatel/5/head -> origin/gh/NikhilAPatel/5/head 2025-12-04T08:57:03.4361907Z * [new branch] gh/NikhilAPatel/5/orig -> origin/gh/NikhilAPatel/5/orig 2025-12-04T08:57:03.4364473Z * [new branch] gh/PaliC/17/base -> origin/gh/PaliC/17/base 2025-12-04T08:57:03.4366020Z * [new branch] gh/PaliC/17/head -> origin/gh/PaliC/17/head 2025-12-04T08:57:03.4367621Z * [new branch] gh/PaliC/17/orig -> origin/gh/PaliC/17/orig 2025-12-04T08:57:03.4369791Z * [new branch] gh/PaliC/18/base -> origin/gh/PaliC/18/base 2025-12-04T08:57:03.4371322Z * [new branch] gh/PaliC/18/head -> origin/gh/PaliC/18/head 2025-12-04T08:57:03.4386518Z * [new branch] gh/PaliC/18/orig -> origin/gh/PaliC/18/orig 2025-12-04T08:57:03.4387046Z * [new branch] gh/PaliC/20/base -> origin/gh/PaliC/20/base 2025-12-04T08:57:03.4387452Z * [new branch] gh/PaliC/20/head -> origin/gh/PaliC/20/head 2025-12-04T08:57:03.4387814Z * [new branch] gh/PaliC/20/orig -> origin/gh/PaliC/20/orig 2025-12-04T08:57:03.4388166Z * [new branch] gh/PaliC/21/base -> origin/gh/PaliC/21/base 2025-12-04T08:57:03.4388535Z * [new branch] gh/PaliC/21/head -> origin/gh/PaliC/21/head 2025-12-04T08:57:03.4389026Z * [new branch] gh/PaliC/21/orig -> origin/gh/PaliC/21/orig 2025-12-04T08:57:03.4389477Z * [new branch] gh/PaliC/23/base -> origin/gh/PaliC/23/base 2025-12-04T08:57:03.4389831Z * [new branch] gh/PaliC/23/head -> origin/gh/PaliC/23/head 2025-12-04T08:57:03.4390202Z * [new branch] gh/PaliC/23/orig -> origin/gh/PaliC/23/orig 2025-12-04T08:57:03.4390795Z * [new branch] gh/PaliC/24/base -> origin/gh/PaliC/24/base 2025-12-04T08:57:03.4392527Z * [new branch] gh/PaliC/24/head -> origin/gh/PaliC/24/head 2025-12-04T08:57:03.4394032Z * [new branch] gh/PaliC/24/orig -> origin/gh/PaliC/24/orig 2025-12-04T08:57:03.4396058Z * [new branch] gh/PaliC/25/head -> origin/gh/PaliC/25/head 2025-12-04T08:57:03.4397593Z * [new branch] gh/PaliC/25/next -> origin/gh/PaliC/25/next 2025-12-04T08:57:03.4399213Z * [new branch] gh/PaliC/25/orig -> origin/gh/PaliC/25/orig 2025-12-04T08:57:03.4401390Z * [new branch] gh/PaliC/26/head -> origin/gh/PaliC/26/head 2025-12-04T08:57:03.4402788Z * [new branch] gh/PaliC/26/next -> origin/gh/PaliC/26/next 2025-12-04T08:57:03.4404281Z * [new branch] gh/PaliC/26/orig -> origin/gh/PaliC/26/orig 2025-12-04T08:57:03.4407513Z * [new branch] gh/PaliC/27/next -> origin/gh/PaliC/27/next 2025-12-04T08:57:03.4409592Z * [new branch] gh/PaliC/28/head -> origin/gh/PaliC/28/head 2025-12-04T08:57:03.4411037Z * [new branch] gh/PaliC/28/next -> origin/gh/PaliC/28/next 2025-12-04T08:57:03.4412616Z * [new branch] gh/PaliC/28/orig -> origin/gh/PaliC/28/orig 2025-12-04T08:57:03.4415041Z * [new branch] gh/PaliC/29/head -> origin/gh/PaliC/29/head 2025-12-04T08:57:03.4416425Z * [new branch] gh/PaliC/29/next -> origin/gh/PaliC/29/next 2025-12-04T08:57:03.4417975Z * [new branch] gh/PaliC/29/orig -> origin/gh/PaliC/29/orig 2025-12-04T08:57:03.4420127Z * [new branch] gh/PaliC/30/head -> origin/gh/PaliC/30/head 2025-12-04T08:57:03.4421596Z * [new branch] gh/PaliC/30/next -> origin/gh/PaliC/30/next 2025-12-04T08:57:03.4423207Z * [new branch] gh/PaliC/30/orig -> origin/gh/PaliC/30/orig 2025-12-04T08:57:03.4425306Z * [new branch] gh/PaliC/31/head -> origin/gh/PaliC/31/head 2025-12-04T08:57:03.4426674Z * [new branch] gh/PaliC/31/next -> origin/gh/PaliC/31/next 2025-12-04T08:57:03.4428419Z * [new branch] gh/PaliC/31/orig -> origin/gh/PaliC/31/orig 2025-12-04T08:57:03.4430941Z * [new branch] gh/PaulZhang12/25/base -> origin/gh/PaulZhang12/25/base 2025-12-04T08:57:03.4432649Z * [new branch] gh/PaulZhang12/25/head -> origin/gh/PaulZhang12/25/head 2025-12-04T08:57:03.4434235Z * [new branch] gh/PaulZhang12/25/orig -> origin/gh/PaulZhang12/25/orig 2025-12-04T08:57:03.4436374Z * [new branch] gh/PaulZhang12/28/base -> origin/gh/PaulZhang12/28/base 2025-12-04T08:57:03.4437972Z * [new branch] gh/PaulZhang12/28/head -> origin/gh/PaulZhang12/28/head 2025-12-04T08:57:03.4439554Z * [new branch] gh/PaulZhang12/28/orig -> origin/gh/PaulZhang12/28/orig 2025-12-04T08:57:03.4441828Z * [new branch] gh/PaulZhang12/31/base -> origin/gh/PaulZhang12/31/base 2025-12-04T08:57:03.4443374Z * [new branch] gh/PaulZhang12/31/head -> origin/gh/PaulZhang12/31/head 2025-12-04T08:57:03.4444974Z * [new branch] gh/PaulZhang12/31/orig -> origin/gh/PaulZhang12/31/orig 2025-12-04T08:57:03.4447932Z * [new branch] gh/PaulZhang12/37/base -> origin/gh/PaulZhang12/37/base 2025-12-04T08:57:03.4449207Z * [new branch] gh/PaulZhang12/37/head -> origin/gh/PaulZhang12/37/head 2025-12-04T08:57:03.4450351Z * [new branch] gh/PaulZhang12/37/orig -> origin/gh/PaulZhang12/37/orig 2025-12-04T08:57:03.4452672Z * [new branch] gh/PaulZhang12/40/base -> origin/gh/PaulZhang12/40/base 2025-12-04T08:57:03.4454197Z * [new branch] gh/PaulZhang12/40/head -> origin/gh/PaulZhang12/40/head 2025-12-04T08:57:03.4455830Z * [new branch] gh/PaulZhang12/40/orig -> origin/gh/PaulZhang12/40/orig 2025-12-04T08:57:03.4458031Z * [new branch] gh/PaulZhang12/42/base -> origin/gh/PaulZhang12/42/base 2025-12-04T08:57:03.4459589Z * [new branch] gh/PaulZhang12/42/head -> origin/gh/PaulZhang12/42/head 2025-12-04T08:57:03.4461721Z * [new branch] gh/PaulZhang12/43/base -> origin/gh/PaulZhang12/43/base 2025-12-04T08:57:03.4463408Z * [new branch] gh/PaulZhang12/43/head -> origin/gh/PaulZhang12/43/head 2025-12-04T08:57:03.4464819Z * [new branch] gh/PaulZhang12/43/orig -> origin/gh/PaulZhang12/43/orig 2025-12-04T08:57:03.4466966Z * [new branch] gh/PaulZhang12/44/base -> origin/gh/PaulZhang12/44/base 2025-12-04T08:57:03.4468366Z * [new branch] gh/PaulZhang12/44/head -> origin/gh/PaulZhang12/44/head 2025-12-04T08:57:03.4470707Z * [new branch] gh/PaulZhang12/45/base -> origin/gh/PaulZhang12/45/base 2025-12-04T08:57:03.4472046Z * [new branch] gh/PaulZhang12/45/head -> origin/gh/PaulZhang12/45/head 2025-12-04T08:57:03.4473710Z * [new branch] gh/PaulZhang12/45/orig -> origin/gh/PaulZhang12/45/orig 2025-12-04T08:57:03.4475852Z * [new branch] gh/PaulZhang12/46/base -> origin/gh/PaulZhang12/46/base 2025-12-04T08:57:03.4477419Z * [new branch] gh/PaulZhang12/46/head -> origin/gh/PaulZhang12/46/head 2025-12-04T08:57:03.4478996Z * [new branch] gh/PaulZhang12/46/orig -> origin/gh/PaulZhang12/46/orig 2025-12-04T08:57:03.4481107Z * [new branch] gh/PaulZhang12/47/base -> origin/gh/PaulZhang12/47/base 2025-12-04T08:57:03.4482842Z * [new branch] gh/PaulZhang12/47/head -> origin/gh/PaulZhang12/47/head 2025-12-04T08:57:03.4484397Z * [new branch] gh/PaulZhang12/47/orig -> origin/gh/PaulZhang12/47/orig 2025-12-04T08:57:03.4486341Z * [new branch] gh/PaulZhang12/48/base -> origin/gh/PaulZhang12/48/base 2025-12-04T08:57:03.4487918Z * [new branch] gh/PaulZhang12/48/head -> origin/gh/PaulZhang12/48/head 2025-12-04T08:57:03.4489491Z * [new branch] gh/PaulZhang12/48/orig -> origin/gh/PaulZhang12/48/orig 2025-12-04T08:57:03.4491969Z * [new branch] gh/SamGinzburg/11/base -> origin/gh/SamGinzburg/11/base 2025-12-04T08:57:03.4493642Z * [new branch] gh/SamGinzburg/11/head -> origin/gh/SamGinzburg/11/head 2025-12-04T08:57:03.4496384Z * [new branch] gh/SherlockNoMad/1/base -> origin/gh/SherlockNoMad/1/base 2025-12-04T08:57:03.4497892Z * [new branch] gh/SherlockNoMad/1/head -> origin/gh/SherlockNoMad/1/head 2025-12-04T08:57:03.4500072Z * [new branch] gh/SherlockNoMad/10/base -> origin/gh/SherlockNoMad/10/base 2025-12-04T08:57:03.4501686Z * [new branch] gh/SherlockNoMad/10/head -> origin/gh/SherlockNoMad/10/head 2025-12-04T08:57:03.4503415Z * [new branch] gh/SherlockNoMad/10/orig -> origin/gh/SherlockNoMad/10/orig 2025-12-04T08:57:03.4505835Z * [new branch] gh/SherlockNoMad/11/base -> origin/gh/SherlockNoMad/11/base 2025-12-04T08:57:03.4507154Z * [new branch] gh/SherlockNoMad/11/head -> origin/gh/SherlockNoMad/11/head 2025-12-04T08:57:03.4508984Z * [new branch] gh/SherlockNoMad/11/orig -> origin/gh/SherlockNoMad/11/orig 2025-12-04T08:57:03.4511205Z * [new branch] gh/SherlockNoMad/12/base -> origin/gh/SherlockNoMad/12/base 2025-12-04T08:57:03.4512333Z * [new branch] gh/SherlockNoMad/12/head -> origin/gh/SherlockNoMad/12/head 2025-12-04T08:57:03.4514432Z * [new branch] gh/SherlockNoMad/12/orig -> origin/gh/SherlockNoMad/12/orig 2025-12-04T08:57:03.4516637Z * [new branch] gh/SherlockNoMad/15/base -> origin/gh/SherlockNoMad/15/base 2025-12-04T08:57:03.4518210Z * [new branch] gh/SherlockNoMad/15/head -> origin/gh/SherlockNoMad/15/head 2025-12-04T08:57:03.4519754Z * [new branch] gh/SherlockNoMad/15/orig -> origin/gh/SherlockNoMad/15/orig 2025-12-04T08:57:03.4521893Z * [new branch] gh/SherlockNoMad/17/base -> origin/gh/SherlockNoMad/17/base 2025-12-04T08:57:03.4523437Z * [new branch] gh/SherlockNoMad/17/head -> origin/gh/SherlockNoMad/17/head 2025-12-04T08:57:03.4524957Z * [new branch] gh/SherlockNoMad/17/orig -> origin/gh/SherlockNoMad/17/orig 2025-12-04T08:57:03.4527181Z * [new branch] gh/SherlockNoMad/18/base -> origin/gh/SherlockNoMad/18/base 2025-12-04T08:57:03.4528979Z * [new branch] gh/SherlockNoMad/18/head -> origin/gh/SherlockNoMad/18/head 2025-12-04T08:57:03.4530645Z * [new branch] gh/SherlockNoMad/18/orig -> origin/gh/SherlockNoMad/18/orig 2025-12-04T08:57:03.4532598Z * [new branch] gh/SherlockNoMad/19/base -> origin/gh/SherlockNoMad/19/base 2025-12-04T08:57:03.4534707Z * [new branch] gh/SherlockNoMad/19/head -> origin/gh/SherlockNoMad/19/head 2025-12-04T08:57:03.4536313Z * [new branch] gh/SherlockNoMad/19/orig -> origin/gh/SherlockNoMad/19/orig 2025-12-04T08:57:03.4538247Z * [new branch] gh/SherlockNoMad/2/base -> origin/gh/SherlockNoMad/2/base 2025-12-04T08:57:03.4539728Z * [new branch] gh/SherlockNoMad/2/head -> origin/gh/SherlockNoMad/2/head 2025-12-04T08:57:03.4541716Z * [new branch] gh/SherlockNoMad/20/base -> origin/gh/SherlockNoMad/20/base 2025-12-04T08:57:03.4543591Z * [new branch] gh/SherlockNoMad/20/head -> origin/gh/SherlockNoMad/20/head 2025-12-04T08:57:03.4545205Z * [new branch] gh/SherlockNoMad/20/orig -> origin/gh/SherlockNoMad/20/orig 2025-12-04T08:57:03.4547489Z * [new branch] gh/SherlockNoMad/21/base -> origin/gh/SherlockNoMad/21/base 2025-12-04T08:57:03.4549172Z * [new branch] gh/SherlockNoMad/21/head -> origin/gh/SherlockNoMad/21/head 2025-12-04T08:57:03.4550667Z * [new branch] gh/SherlockNoMad/21/orig -> origin/gh/SherlockNoMad/21/orig 2025-12-04T08:57:03.4552641Z * [new branch] gh/SherlockNoMad/3/base -> origin/gh/SherlockNoMad/3/base 2025-12-04T08:57:03.4554243Z * [new branch] gh/SherlockNoMad/3/head -> origin/gh/SherlockNoMad/3/head 2025-12-04T08:57:03.4556278Z * [new branch] gh/SherlockNoMad/4/base -> origin/gh/SherlockNoMad/4/base 2025-12-04T08:57:03.4557774Z * [new branch] gh/SherlockNoMad/4/head -> origin/gh/SherlockNoMad/4/head 2025-12-04T08:57:03.4559736Z * [new branch] gh/SherlockNoMad/5/base -> origin/gh/SherlockNoMad/5/base 2025-12-04T08:57:03.4561286Z * [new branch] gh/SherlockNoMad/5/head -> origin/gh/SherlockNoMad/5/head 2025-12-04T08:57:03.4564302Z * [new branch] gh/Sidharth123-cpu/24/base -> origin/gh/Sidharth123-cpu/24/base 2025-12-04T08:57:03.4566409Z * [new branch] gh/Sidharth123-cpu/25/base -> origin/gh/Sidharth123-cpu/25/base 2025-12-04T08:57:03.4568382Z * [new branch] gh/Sidharth123-cpu/26/base -> origin/gh/Sidharth123-cpu/26/base 2025-12-04T08:57:03.4570536Z * [new branch] gh/Sidharth123-cpu/27/base -> origin/gh/Sidharth123-cpu/27/base 2025-12-04T08:57:03.4573291Z * [new branch] gh/StrongerXi/1/base -> origin/gh/StrongerXi/1/base 2025-12-04T08:57:03.4575121Z * [new branch] gh/StrongerXi/1/head -> origin/gh/StrongerXi/1/head 2025-12-04T08:57:03.4577175Z * [new branch] gh/StrongerXi/71/base -> origin/gh/StrongerXi/71/base 2025-12-04T08:57:03.4578696Z * [new branch] gh/StrongerXi/71/head -> origin/gh/StrongerXi/71/head 2025-12-04T08:57:03.4580714Z * [new branch] gh/StrongerXi/72/base -> origin/gh/StrongerXi/72/base 2025-12-04T08:57:03.4582242Z * [new branch] gh/StrongerXi/72/head -> origin/gh/StrongerXi/72/head 2025-12-04T08:57:03.4584320Z * [new branch] gh/StrongerXi/73/base -> origin/gh/StrongerXi/73/base 2025-12-04T08:57:03.4585884Z * [new branch] gh/StrongerXi/73/head -> origin/gh/StrongerXi/73/head 2025-12-04T08:57:03.4587486Z * [new branch] gh/StrongerXi/73/orig -> origin/gh/StrongerXi/73/orig 2025-12-04T08:57:03.4590137Z * [new branch] gh/XilunWu/160/base -> origin/gh/XilunWu/160/base 2025-12-04T08:57:03.4591691Z * [new branch] gh/XilunWu/160/head -> origin/gh/XilunWu/160/head 2025-12-04T08:57:03.4593316Z * [new branch] gh/XilunWu/160/orig -> origin/gh/XilunWu/160/orig 2025-12-04T08:57:03.4595324Z * [new branch] gh/XilunWu/163/base -> origin/gh/XilunWu/163/base 2025-12-04T08:57:03.4596939Z * [new branch] gh/XilunWu/163/head -> origin/gh/XilunWu/163/head 2025-12-04T08:57:03.4598544Z * [new branch] gh/XilunWu/163/orig -> origin/gh/XilunWu/163/orig 2025-12-04T08:57:03.4600705Z * [new branch] gh/XilunWu/168/base -> origin/gh/XilunWu/168/base 2025-12-04T08:57:03.4602249Z * [new branch] gh/XilunWu/168/head -> origin/gh/XilunWu/168/head 2025-12-04T08:57:03.4603844Z * [new branch] gh/XilunWu/168/orig -> origin/gh/XilunWu/168/orig 2025-12-04T08:57:03.4606384Z * [new branch] gh/XilunWu/169/base -> origin/gh/XilunWu/169/base 2025-12-04T08:57:03.4607903Z * [new branch] gh/XilunWu/169/head -> origin/gh/XilunWu/169/head 2025-12-04T08:57:03.4609477Z * [new branch] gh/XilunWu/169/orig -> origin/gh/XilunWu/169/orig 2025-12-04T08:57:03.4611427Z * [new branch] gh/XilunWu/170/base -> origin/gh/XilunWu/170/base 2025-12-04T08:57:03.4613005Z * [new branch] gh/XilunWu/170/head -> origin/gh/XilunWu/170/head 2025-12-04T08:57:03.4614662Z * [new branch] gh/XilunWu/170/orig -> origin/gh/XilunWu/170/orig 2025-12-04T08:57:03.4616829Z * [new branch] gh/XilunWu/171/base -> origin/gh/XilunWu/171/base 2025-12-04T08:57:03.4618466Z * [new branch] gh/XilunWu/171/head -> origin/gh/XilunWu/171/head 2025-12-04T08:57:03.4620020Z * [new branch] gh/XilunWu/171/orig -> origin/gh/XilunWu/171/orig 2025-12-04T08:57:03.4622070Z * [new branch] gh/XilunWu/173/base -> origin/gh/XilunWu/173/base 2025-12-04T08:57:03.4623777Z * [new branch] gh/XilunWu/173/head -> origin/gh/XilunWu/173/head 2025-12-04T08:57:03.4625336Z * [new branch] gh/XilunWu/173/orig -> origin/gh/XilunWu/173/orig 2025-12-04T08:57:03.4627464Z * [new branch] gh/XilunWu/175/base -> origin/gh/XilunWu/175/base 2025-12-04T08:57:03.4629005Z * [new branch] gh/XilunWu/175/head -> origin/gh/XilunWu/175/head 2025-12-04T08:57:03.4630627Z * [new branch] gh/XilunWu/175/orig -> origin/gh/XilunWu/175/orig 2025-12-04T08:57:03.4632825Z * [new branch] gh/XilunWu/176/base -> origin/gh/XilunWu/176/base 2025-12-04T08:57:03.4634645Z * [new branch] gh/XilunWu/176/head -> origin/gh/XilunWu/176/head 2025-12-04T08:57:03.4636334Z * [new branch] gh/XilunWu/176/orig -> origin/gh/XilunWu/176/orig 2025-12-04T08:57:03.4639085Z * [new branch] gh/XuehaiPan/14/base -> origin/gh/XuehaiPan/14/base 2025-12-04T08:57:03.4640540Z * [new branch] gh/XuehaiPan/14/head -> origin/gh/XuehaiPan/14/head 2025-12-04T08:57:03.4642143Z * [new branch] gh/XuehaiPan/14/orig -> origin/gh/XuehaiPan/14/orig 2025-12-04T08:57:03.4644260Z * [new branch] gh/XuehaiPan/179/base -> origin/gh/XuehaiPan/179/base 2025-12-04T08:57:03.4645837Z * [new branch] gh/XuehaiPan/179/head -> origin/gh/XuehaiPan/179/head 2025-12-04T08:57:03.4647506Z * [new branch] gh/XuehaiPan/179/orig -> origin/gh/XuehaiPan/179/orig 2025-12-04T08:57:03.4649541Z * [new branch] gh/XuehaiPan/249/base -> origin/gh/XuehaiPan/249/base 2025-12-04T08:57:03.4651100Z * [new branch] gh/XuehaiPan/249/head -> origin/gh/XuehaiPan/249/head 2025-12-04T08:57:03.4652766Z * [new branch] gh/XuehaiPan/249/orig -> origin/gh/XuehaiPan/249/orig 2025-12-04T08:57:03.4654939Z * [new branch] gh/XuehaiPan/253/base -> origin/gh/XuehaiPan/253/base 2025-12-04T08:57:03.4656531Z * [new branch] gh/XuehaiPan/253/head -> origin/gh/XuehaiPan/253/head 2025-12-04T08:57:03.4658134Z * [new branch] gh/XuehaiPan/253/orig -> origin/gh/XuehaiPan/253/orig 2025-12-04T08:57:03.4660267Z * [new branch] gh/XuehaiPan/254/base -> origin/gh/XuehaiPan/254/base 2025-12-04T08:57:03.4661871Z * [new branch] gh/XuehaiPan/254/head -> origin/gh/XuehaiPan/254/head 2025-12-04T08:57:03.4663465Z * [new branch] gh/XuehaiPan/254/orig -> origin/gh/XuehaiPan/254/orig 2025-12-04T08:57:03.4665441Z * [new branch] gh/XuehaiPan/255/base -> origin/gh/XuehaiPan/255/base 2025-12-04T08:57:03.4667062Z * [new branch] gh/XuehaiPan/255/head -> origin/gh/XuehaiPan/255/head 2025-12-04T08:57:03.4668647Z * [new branch] gh/XuehaiPan/255/orig -> origin/gh/XuehaiPan/255/orig 2025-12-04T08:57:03.4670733Z * [new branch] gh/XuehaiPan/271/base -> origin/gh/XuehaiPan/271/base 2025-12-04T08:57:03.4672351Z * [new branch] gh/XuehaiPan/271/head -> origin/gh/XuehaiPan/271/head 2025-12-04T08:57:03.4674050Z * [new branch] gh/XuehaiPan/271/orig -> origin/gh/XuehaiPan/271/orig 2025-12-04T08:57:03.4676043Z * [new branch] gh/XuehaiPan/343/base -> origin/gh/XuehaiPan/343/base 2025-12-04T08:57:03.4677629Z * [new branch] gh/XuehaiPan/343/head -> origin/gh/XuehaiPan/343/head 2025-12-04T08:57:03.4679184Z * [new branch] gh/XuehaiPan/343/orig -> origin/gh/XuehaiPan/343/orig 2025-12-04T08:57:03.4681317Z * [new branch] gh/XuehaiPan/347/base -> origin/gh/XuehaiPan/347/base 2025-12-04T08:57:03.4682919Z * [new branch] gh/XuehaiPan/347/head -> origin/gh/XuehaiPan/347/head 2025-12-04T08:57:03.4684444Z * [new branch] gh/XuehaiPan/347/orig -> origin/gh/XuehaiPan/347/orig 2025-12-04T08:57:03.4686637Z * [new branch] gh/XuehaiPan/348/base -> origin/gh/XuehaiPan/348/base 2025-12-04T08:57:03.4688164Z * [new branch] gh/XuehaiPan/348/head -> origin/gh/XuehaiPan/348/head 2025-12-04T08:57:03.4689724Z * [new branch] gh/XuehaiPan/348/orig -> origin/gh/XuehaiPan/348/orig 2025-12-04T08:57:03.4691790Z * [new branch] gh/XuehaiPan/350/base -> origin/gh/XuehaiPan/350/base 2025-12-04T08:57:03.4693434Z * [new branch] gh/XuehaiPan/350/head -> origin/gh/XuehaiPan/350/head 2025-12-04T08:57:03.4695058Z * [new branch] gh/XuehaiPan/350/orig -> origin/gh/XuehaiPan/350/orig 2025-12-04T08:57:03.4697183Z * [new branch] gh/XuehaiPan/365/base -> origin/gh/XuehaiPan/365/base 2025-12-04T08:57:03.4698782Z * [new branch] gh/XuehaiPan/365/head -> origin/gh/XuehaiPan/365/head 2025-12-04T08:57:03.4700486Z * [new branch] gh/XuehaiPan/365/orig -> origin/gh/XuehaiPan/365/orig 2025-12-04T08:57:03.4702584Z * [new branch] gh/XuehaiPan/366/base -> origin/gh/XuehaiPan/366/base 2025-12-04T08:57:03.4704195Z * [new branch] gh/XuehaiPan/366/head -> origin/gh/XuehaiPan/366/head 2025-12-04T08:57:03.4706601Z * [new branch] gh/XuehaiPan/370/base -> origin/gh/XuehaiPan/370/base 2025-12-04T08:57:03.4708138Z * [new branch] gh/XuehaiPan/370/head -> origin/gh/XuehaiPan/370/head 2025-12-04T08:57:03.4709710Z * [new branch] gh/XuehaiPan/370/orig -> origin/gh/XuehaiPan/370/orig 2025-12-04T08:57:03.4711891Z * [new branch] gh/XuehaiPan/390/base -> origin/gh/XuehaiPan/390/base 2025-12-04T08:57:03.4713484Z * [new branch] gh/XuehaiPan/390/head -> origin/gh/XuehaiPan/390/head 2025-12-04T08:57:03.4715069Z * [new branch] gh/XuehaiPan/390/orig -> origin/gh/XuehaiPan/390/orig 2025-12-04T08:57:03.4717149Z * [new branch] gh/XuehaiPan/391/base -> origin/gh/XuehaiPan/391/base 2025-12-04T08:57:03.4719054Z * [new branch] gh/XuehaiPan/391/head -> origin/gh/XuehaiPan/391/head 2025-12-04T08:57:03.4720409Z * [new branch] gh/XuehaiPan/391/orig -> origin/gh/XuehaiPan/391/orig 2025-12-04T08:57:03.4722815Z * [new branch] gh/XuehaiPan/392/base -> origin/gh/XuehaiPan/392/base 2025-12-04T08:57:03.4724238Z * [new branch] gh/XuehaiPan/392/head -> origin/gh/XuehaiPan/392/head 2025-12-04T08:57:03.4725751Z * [new branch] gh/XuehaiPan/392/orig -> origin/gh/XuehaiPan/392/orig 2025-12-04T08:57:03.4728688Z * [new branch] gh/XuehaiPan/394/base -> origin/gh/XuehaiPan/394/base 2025-12-04T08:57:03.4729867Z * [new branch] gh/XuehaiPan/394/head -> origin/gh/XuehaiPan/394/head 2025-12-04T08:57:03.4731423Z * [new branch] gh/XuehaiPan/394/orig -> origin/gh/XuehaiPan/394/orig 2025-12-04T08:57:03.4734378Z * [new branch] gh/XuehaiPan/397/base -> origin/gh/XuehaiPan/397/base 2025-12-04T08:57:03.4735615Z * [new branch] gh/XuehaiPan/397/head -> origin/gh/XuehaiPan/397/head 2025-12-04T08:57:03.4737225Z * [new branch] gh/XuehaiPan/397/orig -> origin/gh/XuehaiPan/397/orig 2025-12-04T08:57:03.4739712Z * [new branch] gh/XuehaiPan/398/base -> origin/gh/XuehaiPan/398/base 2025-12-04T08:57:03.4740939Z * [new branch] gh/XuehaiPan/398/head -> origin/gh/XuehaiPan/398/head 2025-12-04T08:57:03.4742599Z * [new branch] gh/XuehaiPan/398/orig -> origin/gh/XuehaiPan/398/orig 2025-12-04T08:57:03.4745192Z * [new branch] gh/XuehaiPan/399/base -> origin/gh/XuehaiPan/399/base 2025-12-04T08:57:03.4746452Z * [new branch] gh/XuehaiPan/399/head -> origin/gh/XuehaiPan/399/head 2025-12-04T08:57:03.4748020Z * [new branch] gh/XuehaiPan/399/orig -> origin/gh/XuehaiPan/399/orig 2025-12-04T08:57:03.4750491Z * [new branch] gh/XuehaiPan/400/base -> origin/gh/XuehaiPan/400/base 2025-12-04T08:57:03.4751756Z * [new branch] gh/XuehaiPan/400/head -> origin/gh/XuehaiPan/400/head 2025-12-04T08:57:03.4753290Z * [new branch] gh/XuehaiPan/400/orig -> origin/gh/XuehaiPan/400/orig 2025-12-04T08:57:03.4756158Z * [new branch] gh/ZhiweiYan-96/39/base -> origin/gh/ZhiweiYan-96/39/base 2025-12-04T08:57:03.4757360Z * [new branch] gh/ZhiweiYan-96/39/head -> origin/gh/ZhiweiYan-96/39/head 2025-12-04T08:57:03.4759045Z * [new branch] gh/ZhiweiYan-96/39/orig -> origin/gh/ZhiweiYan-96/39/orig 2025-12-04T08:57:03.4761388Z * [new branch] gh/ZhiweiYan-96/44/base -> origin/gh/ZhiweiYan-96/44/base 2025-12-04T08:57:03.4762902Z * [new branch] gh/ZhiweiYan-96/44/head -> origin/gh/ZhiweiYan-96/44/head 2025-12-04T08:57:03.4765004Z * [new branch] gh/ZhiweiYan-96/45/base -> origin/gh/ZhiweiYan-96/45/base 2025-12-04T08:57:03.4766423Z * [new branch] gh/ZhiweiYan-96/45/head -> origin/gh/ZhiweiYan-96/45/head 2025-12-04T08:57:03.4768944Z * [new branch] gh/ZhiweiYan-96/49/base -> origin/gh/ZhiweiYan-96/49/base 2025-12-04T08:57:03.4770381Z * [new branch] gh/ZhiweiYan-96/49/head -> origin/gh/ZhiweiYan-96/49/head 2025-12-04T08:57:03.4772624Z * [new branch] gh/ZhiweiYan-96/62/base -> origin/gh/ZhiweiYan-96/62/base 2025-12-04T08:57:03.4773929Z * [new branch] gh/ZhiweiYan-96/62/head -> origin/gh/ZhiweiYan-96/62/head 2025-12-04T08:57:03.4776510Z * [new branch] gh/ZhiweiYan-96/66/base -> origin/gh/ZhiweiYan-96/66/base 2025-12-04T08:57:03.4777713Z * [new branch] gh/ZhiweiYan-96/66/head -> origin/gh/ZhiweiYan-96/66/head 2025-12-04T08:57:03.4780067Z * [new branch] gh/ZhiweiYan-96/67/base -> origin/gh/ZhiweiYan-96/67/base 2025-12-04T08:57:03.4781243Z * [new branch] gh/ZhiweiYan-96/67/head -> origin/gh/ZhiweiYan-96/67/head 2025-12-04T08:57:03.4783867Z * [new branch] gh/ZhiweiYan-96/68/base -> origin/gh/ZhiweiYan-96/68/base 2025-12-04T08:57:03.4785273Z * [new branch] gh/ZhiweiYan-96/68/head -> origin/gh/ZhiweiYan-96/68/head 2025-12-04T08:57:03.4786834Z * [new branch] gh/ZhiweiYan-96/68/orig -> origin/gh/ZhiweiYan-96/68/orig 2025-12-04T08:57:03.4789734Z * [new branch] gh/aakhundov/1/base -> origin/gh/aakhundov/1/base 2025-12-04T08:57:03.4791002Z * [new branch] gh/aakhundov/1/head -> origin/gh/aakhundov/1/head 2025-12-04T08:57:03.4793880Z * [new branch] gh/aakhundov/2/base -> origin/gh/aakhundov/2/base 2025-12-04T08:57:03.4795089Z * [new branch] gh/aakhundov/2/head -> origin/gh/aakhundov/2/head 2025-12-04T08:57:03.4797499Z * [new branch] gh/aditew01/openblas -> origin/gh/aditew01/openblas 2025-12-04T08:57:03.4798805Z * [new branch] gh/aditew01/sbgemm -> origin/gh/aditew01/sbgemm 2025-12-04T08:57:03.4800459Z * [new branch] gh/aditew01/vecbf16 -> origin/gh/aditew01/vecbf16 2025-12-04T08:57:03.4803290Z * [new branch] gh/albanD/4/base -> origin/gh/albanD/4/base 2025-12-04T08:57:03.4804695Z * [new branch] gh/albanD/4/head -> origin/gh/albanD/4/head 2025-12-04T08:57:03.4808468Z * [new branch] gh/albanD/4/orig -> origin/gh/albanD/4/orig 2025-12-04T08:57:03.4810774Z * [new branch] gh/alexbrauckmann/paddedtensor_faketensor_init -> origin/gh/alexbrauckmann/paddedtensor_faketensor_init 2025-12-04T08:57:03.4813423Z * [new branch] gh/alexsamardzic/12/base -> origin/gh/alexsamardzic/12/base 2025-12-04T08:57:03.4814775Z * [new branch] gh/alexsamardzic/12/head -> origin/gh/alexsamardzic/12/head 2025-12-04T08:57:03.4816298Z * [new branch] gh/alexsamardzic/12/orig -> origin/gh/alexsamardzic/12/orig 2025-12-04T08:57:03.4818748Z * [new branch] gh/alexsamardzic/14/base -> origin/gh/alexsamardzic/14/base 2025-12-04T08:57:03.4820022Z * [new branch] gh/alexsamardzic/14/head -> origin/gh/alexsamardzic/14/head 2025-12-04T08:57:03.4821601Z * [new branch] gh/alexsamardzic/14/orig -> origin/gh/alexsamardzic/14/orig 2025-12-04T08:57:03.4824050Z * [new branch] gh/alexsamardzic/15/base -> origin/gh/alexsamardzic/15/base 2025-12-04T08:57:03.4825212Z * [new branch] gh/alexsamardzic/15/head -> origin/gh/alexsamardzic/15/head 2025-12-04T08:57:03.4826878Z * [new branch] gh/alexsamardzic/15/orig -> origin/gh/alexsamardzic/15/orig 2025-12-04T08:57:03.4829956Z * [new branch] gh/amjames/18/base -> origin/gh/amjames/18/base 2025-12-04T08:57:03.4831040Z * [new branch] gh/amjames/18/head -> origin/gh/amjames/18/head 2025-12-04T08:57:03.4832615Z * [new branch] gh/amjames/18/orig -> origin/gh/amjames/18/orig 2025-12-04T08:57:03.4835805Z * [new branch] gh/andrewor14/35/base -> origin/gh/andrewor14/35/base 2025-12-04T08:57:03.4837131Z * [new branch] gh/andrewor14/35/head -> origin/gh/andrewor14/35/head 2025-12-04T08:57:03.4838681Z * [new branch] gh/andrewor14/35/orig -> origin/gh/andrewor14/35/orig 2025-12-04T08:57:03.4841345Z * [new branch] gh/andrewor14/50/base -> origin/gh/andrewor14/50/base 2025-12-04T08:57:03.4842719Z * [new branch] gh/andrewor14/50/head -> origin/gh/andrewor14/50/head 2025-12-04T08:57:03.4844432Z * [new branch] gh/andrewor14/50/orig -> origin/gh/andrewor14/50/orig 2025-12-04T08:57:03.4847281Z * [new branch] gh/andyanwang/30/base -> origin/gh/andyanwang/30/base 2025-12-04T08:57:03.4848754Z * [new branch] gh/andyanwang/30/orig -> origin/gh/andyanwang/30/orig 2025-12-04T08:57:03.4851249Z * [new branch] gh/andyanwang/31/base -> origin/gh/andyanwang/31/base 2025-12-04T08:57:03.4852678Z * [new branch] gh/andyanwang/31/orig -> origin/gh/andyanwang/31/orig 2025-12-04T08:57:03.4855326Z * [new branch] gh/andyanwang/39/base -> origin/gh/andyanwang/39/base 2025-12-04T08:57:03.4856561Z * [new branch] gh/andyanwang/39/head -> origin/gh/andyanwang/39/head 2025-12-04T08:57:03.4858113Z * [new branch] gh/andyanwang/39/orig -> origin/gh/andyanwang/39/orig 2025-12-04T08:57:03.4860823Z * [new branch] gh/andyanwang/42/base -> origin/gh/andyanwang/42/base 2025-12-04T08:57:03.4862045Z * [new branch] gh/andyanwang/42/head -> origin/gh/andyanwang/42/head 2025-12-04T08:57:03.4863665Z * [new branch] gh/andyanwang/42/orig -> origin/gh/andyanwang/42/orig 2025-12-04T08:57:03.4866199Z * [new branch] gh/andyanwang/45/base -> origin/gh/andyanwang/45/base 2025-12-04T08:57:03.4867486Z * [new branch] gh/andyanwang/45/head -> origin/gh/andyanwang/45/head 2025-12-04T08:57:03.4869060Z * [new branch] gh/andyanwang/45/orig -> origin/gh/andyanwang/45/orig 2025-12-04T08:57:03.4871903Z * [new branch] gh/angelayi/107/base -> origin/gh/angelayi/107/base 2025-12-04T08:57:03.4873171Z * [new branch] gh/angelayi/107/head -> origin/gh/angelayi/107/head 2025-12-04T08:57:03.4875574Z * [new branch] gh/angelayi/114/base -> origin/gh/angelayi/114/base 2025-12-04T08:57:03.4876920Z * [new branch] gh/angelayi/114/head -> origin/gh/angelayi/114/head 2025-12-04T08:57:03.4879209Z * [new branch] gh/angelayi/114/orig -> origin/gh/angelayi/114/orig 2025-12-04T08:57:03.4881314Z * [new branch] gh/angelayi/116/base -> origin/gh/angelayi/116/base 2025-12-04T08:57:03.4882588Z * [new branch] gh/angelayi/116/head -> origin/gh/angelayi/116/head 2025-12-04T08:57:03.4884276Z * [new branch] gh/angelayi/116/orig -> origin/gh/angelayi/116/orig 2025-12-04T08:57:03.4886824Z * [new branch] gh/angelayi/122/base -> origin/gh/angelayi/122/base 2025-12-04T08:57:03.4888431Z * [new branch] gh/angelayi/122/head -> origin/gh/angelayi/122/head 2025-12-04T08:57:03.4889974Z * [new branch] gh/angelayi/122/orig -> origin/gh/angelayi/122/orig 2025-12-04T08:57:03.4892553Z * [new branch] gh/angelayi/124/base -> origin/gh/angelayi/124/base 2025-12-04T08:57:03.4893912Z * [new branch] gh/angelayi/124/head -> origin/gh/angelayi/124/head 2025-12-04T08:57:03.4895680Z * [new branch] gh/angelayi/124/orig -> origin/gh/angelayi/124/orig 2025-12-04T08:57:03.4897982Z * [new branch] gh/angelayi/128/base -> origin/gh/angelayi/128/base 2025-12-04T08:57:03.4899322Z * [new branch] gh/angelayi/128/head -> origin/gh/angelayi/128/head 2025-12-04T08:57:03.4900866Z * [new branch] gh/angelayi/128/orig -> origin/gh/angelayi/128/orig 2025-12-04T08:57:03.4903374Z * [new branch] gh/angelayi/131/base -> origin/gh/angelayi/131/base 2025-12-04T08:57:03.4904805Z * [new branch] gh/angelayi/131/head -> origin/gh/angelayi/131/head 2025-12-04T08:57:03.4906518Z * [new branch] gh/angelayi/131/orig -> origin/gh/angelayi/131/orig 2025-12-04T08:57:03.4909216Z * [new branch] gh/angelayi/132/base -> origin/gh/angelayi/132/base 2025-12-04T08:57:03.4910809Z * [new branch] gh/angelayi/132/head -> origin/gh/angelayi/132/head 2025-12-04T08:57:03.4912649Z * [new branch] gh/angelayi/132/orig -> origin/gh/angelayi/132/orig 2025-12-04T08:57:03.4914799Z * [new branch] gh/angelayi/133/base -> origin/gh/angelayi/133/base 2025-12-04T08:57:03.4916208Z * [new branch] gh/angelayi/133/head -> origin/gh/angelayi/133/head 2025-12-04T08:57:03.4917800Z * [new branch] gh/angelayi/133/orig -> origin/gh/angelayi/133/orig 2025-12-04T08:57:03.4920478Z * [new branch] gh/angelayi/134/base -> origin/gh/angelayi/134/base 2025-12-04T08:57:03.4921977Z * [new branch] gh/angelayi/134/head -> origin/gh/angelayi/134/head 2025-12-04T08:57:03.4923519Z * [new branch] gh/angelayi/134/orig -> origin/gh/angelayi/134/orig 2025-12-04T08:57:03.4925997Z * [new branch] gh/angelayi/135/base -> origin/gh/angelayi/135/base 2025-12-04T08:57:03.4927427Z * [new branch] gh/angelayi/135/head -> origin/gh/angelayi/135/head 2025-12-04T08:57:03.4929044Z * [new branch] gh/angelayi/135/orig -> origin/gh/angelayi/135/orig 2025-12-04T08:57:03.4931406Z * [new branch] gh/angelayi/136/base -> origin/gh/angelayi/136/base 2025-12-04T08:57:03.4932737Z * [new branch] gh/angelayi/136/head -> origin/gh/angelayi/136/head 2025-12-04T08:57:03.4934623Z * [new branch] gh/angelayi/136/orig -> origin/gh/angelayi/136/orig 2025-12-04T08:57:03.4937037Z * [new branch] gh/angelayi/137/base -> origin/gh/angelayi/137/base 2025-12-04T08:57:03.4938160Z * [new branch] gh/angelayi/137/head -> origin/gh/angelayi/137/head 2025-12-04T08:57:03.4939910Z * [new branch] gh/angelayi/137/orig -> origin/gh/angelayi/137/orig 2025-12-04T08:57:03.4942188Z * [new branch] gh/angelayi/138/base -> origin/gh/angelayi/138/base 2025-12-04T08:57:03.4943655Z * [new branch] gh/angelayi/138/head -> origin/gh/angelayi/138/head 2025-12-04T08:57:03.4945300Z * [new branch] gh/angelayi/138/orig -> origin/gh/angelayi/138/orig 2025-12-04T08:57:03.4947643Z * [new branch] gh/angelayi/139/base -> origin/gh/angelayi/139/base 2025-12-04T08:57:03.4949024Z * [new branch] gh/angelayi/139/head -> origin/gh/angelayi/139/head 2025-12-04T08:57:03.4950609Z * [new branch] gh/angelayi/139/orig -> origin/gh/angelayi/139/orig 2025-12-04T08:57:03.4953243Z * [new branch] gh/angelayi/140/base -> origin/gh/angelayi/140/base 2025-12-04T08:57:03.4954835Z * [new branch] gh/angelayi/140/head -> origin/gh/angelayi/140/head 2025-12-04T08:57:03.4956626Z * [new branch] gh/angelayi/140/orig -> origin/gh/angelayi/140/orig 2025-12-04T08:57:03.4959456Z * [new branch] gh/angelayi/141/base -> origin/gh/angelayi/141/base 2025-12-04T08:57:03.4961015Z * [new branch] gh/angelayi/141/head -> origin/gh/angelayi/141/head 2025-12-04T08:57:03.4962462Z * [new branch] gh/angelayi/141/orig -> origin/gh/angelayi/141/orig 2025-12-04T08:57:03.4964952Z * [new branch] gh/angelayi/142/base -> origin/gh/angelayi/142/base 2025-12-04T08:57:03.4966218Z * [new branch] gh/angelayi/142/head -> origin/gh/angelayi/142/head 2025-12-04T08:57:03.4967788Z * [new branch] gh/angelayi/142/orig -> origin/gh/angelayi/142/orig 2025-12-04T08:57:03.4970312Z * [new branch] gh/angelayi/143/base -> origin/gh/angelayi/143/base 2025-12-04T08:57:03.4971527Z * [new branch] gh/angelayi/143/head -> origin/gh/angelayi/143/head 2025-12-04T08:57:03.4973107Z * [new branch] gh/angelayi/143/orig -> origin/gh/angelayi/143/orig 2025-12-04T08:57:03.4975694Z * [new branch] gh/angelayi/144/base -> origin/gh/angelayi/144/base 2025-12-04T08:57:03.4977072Z * [new branch] gh/angelayi/144/head -> origin/gh/angelayi/144/head 2025-12-04T08:57:03.4978701Z * [new branch] gh/angelayi/144/orig -> origin/gh/angelayi/144/orig 2025-12-04T08:57:03.4981810Z * [new branch] gh/anijain2305/753/base -> origin/gh/anijain2305/753/base 2025-12-04T08:57:03.4983084Z * [new branch] gh/anijain2305/753/head -> origin/gh/anijain2305/753/head 2025-12-04T08:57:03.4984687Z * [new branch] gh/anijain2305/753/orig -> origin/gh/anijain2305/753/orig 2025-12-04T08:57:03.4987210Z * [new branch] gh/anijain2305/810/base -> origin/gh/anijain2305/810/base 2025-12-04T08:57:03.4988401Z * [new branch] gh/anijain2305/810/head -> origin/gh/anijain2305/810/head 2025-12-04T08:57:03.4990006Z * [new branch] gh/anijain2305/810/orig -> origin/gh/anijain2305/810/orig 2025-12-04T08:57:03.4992411Z * [new branch] gh/anijain2305/854/base -> origin/gh/anijain2305/854/base 2025-12-04T08:57:03.4993755Z * [new branch] gh/anijain2305/854/head -> origin/gh/anijain2305/854/head 2025-12-04T08:57:03.4995305Z * [new branch] gh/anijain2305/854/orig -> origin/gh/anijain2305/854/orig 2025-12-04T08:57:03.4997855Z * [new branch] gh/anijain2305/864/base -> origin/gh/anijain2305/864/base 2025-12-04T08:57:03.4999107Z * [new branch] gh/anijain2305/864/head -> origin/gh/anijain2305/864/head 2025-12-04T08:57:03.5000654Z * [new branch] gh/anijain2305/864/orig -> origin/gh/anijain2305/864/orig 2025-12-04T08:57:03.5003261Z * [new branch] gh/anijain2305/870/base -> origin/gh/anijain2305/870/base 2025-12-04T08:57:03.5004450Z * [new branch] gh/anijain2305/870/head -> origin/gh/anijain2305/870/head 2025-12-04T08:57:03.5006407Z * [new branch] gh/anijain2305/870/orig -> origin/gh/anijain2305/870/orig 2025-12-04T08:57:03.5009032Z * [new branch] gh/anijain2305/873/base -> origin/gh/anijain2305/873/base 2025-12-04T08:57:03.5010117Z * [new branch] gh/anijain2305/873/head -> origin/gh/anijain2305/873/head 2025-12-04T08:57:03.5011660Z * [new branch] gh/anijain2305/873/orig -> origin/gh/anijain2305/873/orig 2025-12-04T08:57:03.5014140Z * [new branch] gh/anijain2305/894/base -> origin/gh/anijain2305/894/base 2025-12-04T08:57:03.5015335Z * [new branch] gh/anijain2305/894/head -> origin/gh/anijain2305/894/head 2025-12-04T08:57:03.5016952Z * [new branch] gh/anijain2305/894/orig -> origin/gh/anijain2305/894/orig 2025-12-04T08:57:03.5019473Z * [new branch] gh/anijain2305/895/base -> origin/gh/anijain2305/895/base 2025-12-04T08:57:03.5020761Z * [new branch] gh/anijain2305/895/head -> origin/gh/anijain2305/895/head 2025-12-04T08:57:03.5022356Z * [new branch] gh/anijain2305/895/orig -> origin/gh/anijain2305/895/orig 2025-12-04T08:57:03.5025173Z * [new branch] gh/anijain2305/910/base -> origin/gh/anijain2305/910/base 2025-12-04T08:57:03.5026126Z * [new branch] gh/anijain2305/910/head -> origin/gh/anijain2305/910/head 2025-12-04T08:57:03.5027774Z * [new branch] gh/anijain2305/910/orig -> origin/gh/anijain2305/910/orig 2025-12-04T08:57:03.5030376Z * [new branch] gh/anijain2305/919/base -> origin/gh/anijain2305/919/base 2025-12-04T08:57:03.5031583Z * [new branch] gh/anijain2305/919/head -> origin/gh/anijain2305/919/head 2025-12-04T08:57:03.5033247Z * [new branch] gh/anijain2305/919/orig -> origin/gh/anijain2305/919/orig 2025-12-04T08:57:03.5035673Z * [new branch] gh/anijain2305/922/base -> origin/gh/anijain2305/922/base 2025-12-04T08:57:03.5036960Z * [new branch] gh/anijain2305/922/head -> origin/gh/anijain2305/922/head 2025-12-04T08:57:03.5038583Z * [new branch] gh/anijain2305/922/orig -> origin/gh/anijain2305/922/orig 2025-12-04T08:57:03.5041056Z * [new branch] gh/anijain2305/932/base -> origin/gh/anijain2305/932/base 2025-12-04T08:57:03.5042383Z * [new branch] gh/anijain2305/932/head -> origin/gh/anijain2305/932/head 2025-12-04T08:57:03.5044354Z * [new branch] gh/anijain2305/932/orig -> origin/gh/anijain2305/932/orig 2025-12-04T08:57:03.5046601Z * [new branch] gh/anijain2305/940/base -> origin/gh/anijain2305/940/base 2025-12-04T08:57:03.5047851Z * [new branch] gh/anijain2305/940/head -> origin/gh/anijain2305/940/head 2025-12-04T08:57:03.5049496Z * [new branch] gh/anijain2305/940/orig -> origin/gh/anijain2305/940/orig 2025-12-04T08:57:03.5052069Z * [new branch] gh/anijain2305/941/base -> origin/gh/anijain2305/941/base 2025-12-04T08:57:03.5053320Z * [new branch] gh/anijain2305/941/head -> origin/gh/anijain2305/941/head 2025-12-04T08:57:03.5054923Z * [new branch] gh/anijain2305/941/orig -> origin/gh/anijain2305/941/orig 2025-12-04T08:57:03.5057038Z * [new branch] gh/anijain2305/942/base -> origin/gh/anijain2305/942/base 2025-12-04T08:57:03.5058521Z * [new branch] gh/anijain2305/942/head -> origin/gh/anijain2305/942/head 2025-12-04T08:57:03.5060159Z * [new branch] gh/anijain2305/942/orig -> origin/gh/anijain2305/942/orig 2025-12-04T08:57:03.5062277Z * [new branch] gh/anijain2305/943/base -> origin/gh/anijain2305/943/base 2025-12-04T08:57:03.5063925Z * [new branch] gh/anijain2305/943/head -> origin/gh/anijain2305/943/head 2025-12-04T08:57:03.5065486Z * [new branch] gh/anijain2305/943/orig -> origin/gh/anijain2305/943/orig 2025-12-04T08:57:03.5068126Z * [new branch] gh/anijain2305/944/base -> origin/gh/anijain2305/944/base 2025-12-04T08:57:03.5069704Z * [new branch] gh/anijain2305/944/head -> origin/gh/anijain2305/944/head 2025-12-04T08:57:03.5071232Z * [new branch] gh/anijain2305/944/orig -> origin/gh/anijain2305/944/orig 2025-12-04T08:57:03.5073873Z * [new branch] gh/anijain2305/945/base -> origin/gh/anijain2305/945/base 2025-12-04T08:57:03.5075431Z * [new branch] gh/anijain2305/945/head -> origin/gh/anijain2305/945/head 2025-12-04T08:57:03.5077026Z * [new branch] gh/anijain2305/945/orig -> origin/gh/anijain2305/945/orig 2025-12-04T08:57:03.5079149Z * [new branch] gh/anijain2305/946/base -> origin/gh/anijain2305/946/base 2025-12-04T08:57:03.5080719Z * [new branch] gh/anijain2305/946/head -> origin/gh/anijain2305/946/head 2025-12-04T08:57:03.5082503Z * [new branch] gh/anijain2305/946/orig -> origin/gh/anijain2305/946/orig 2025-12-04T08:57:03.5084765Z * [new branch] gh/anijain2305/947/base -> origin/gh/anijain2305/947/base 2025-12-04T08:57:03.5086452Z * [new branch] gh/anijain2305/947/head -> origin/gh/anijain2305/947/head 2025-12-04T08:57:03.5088021Z * [new branch] gh/anijain2305/947/orig -> origin/gh/anijain2305/947/orig 2025-12-04T08:57:03.5090071Z * [new branch] gh/anijain2305/948/base -> origin/gh/anijain2305/948/base 2025-12-04T08:57:03.5091561Z * [new branch] gh/anijain2305/948/head -> origin/gh/anijain2305/948/head 2025-12-04T08:57:03.5093160Z * [new branch] gh/anijain2305/948/orig -> origin/gh/anijain2305/948/orig 2025-12-04T08:57:03.5095399Z * [new branch] gh/anijain2305/949/base -> origin/gh/anijain2305/949/base 2025-12-04T08:57:03.5097018Z * [new branch] gh/anijain2305/949/head -> origin/gh/anijain2305/949/head 2025-12-04T08:57:03.5098734Z * [new branch] gh/anijain2305/949/orig -> origin/gh/anijain2305/949/orig 2025-12-04T08:57:03.5100967Z * [new branch] gh/anijain2305/950/base -> origin/gh/anijain2305/950/base 2025-12-04T08:57:03.5102579Z * [new branch] gh/anijain2305/950/head -> origin/gh/anijain2305/950/head 2025-12-04T08:57:03.5104189Z * [new branch] gh/anijain2305/950/orig -> origin/gh/anijain2305/950/orig 2025-12-04T08:57:03.5106768Z * [new branch] gh/anijain2305/951/base -> origin/gh/anijain2305/951/base 2025-12-04T08:57:03.5108324Z * [new branch] gh/anijain2305/951/head -> origin/gh/anijain2305/951/head 2025-12-04T08:57:03.5109889Z * [new branch] gh/anijain2305/951/orig -> origin/gh/anijain2305/951/orig 2025-12-04T08:57:03.5112179Z * [new branch] gh/anijain2305/952/base -> origin/gh/anijain2305/952/base 2025-12-04T08:57:03.5113705Z * [new branch] gh/anijain2305/952/head -> origin/gh/anijain2305/952/head 2025-12-04T08:57:03.5115909Z * [new branch] gh/anijain2305/952/orig -> origin/gh/anijain2305/952/orig 2025-12-04T08:57:03.5118307Z * [new branch] gh/anijain2305/953/base -> origin/gh/anijain2305/953/base 2025-12-04T08:57:03.5119466Z * [new branch] gh/anijain2305/953/head -> origin/gh/anijain2305/953/head 2025-12-04T08:57:03.5121137Z * [new branch] gh/anijain2305/953/orig -> origin/gh/anijain2305/953/orig 2025-12-04T08:57:03.5123625Z * [new branch] gh/anijain2305/954/base -> origin/gh/anijain2305/954/base 2025-12-04T08:57:03.5125079Z * [new branch] gh/anijain2305/954/head -> origin/gh/anijain2305/954/head 2025-12-04T08:57:03.5126499Z * [new branch] gh/anijain2305/954/orig -> origin/gh/anijain2305/954/orig 2025-12-04T08:57:03.5128972Z * [new branch] gh/anijain2305/955/base -> origin/gh/anijain2305/955/base 2025-12-04T08:57:03.5130235Z * [new branch] gh/anijain2305/955/head -> origin/gh/anijain2305/955/head 2025-12-04T08:57:03.5131955Z * [new branch] gh/anijain2305/955/orig -> origin/gh/anijain2305/955/orig 2025-12-04T08:57:03.5134673Z * [new branch] gh/anijain2305/956/base -> origin/gh/anijain2305/956/base 2025-12-04T08:57:03.5135873Z * [new branch] gh/anijain2305/956/head -> origin/gh/anijain2305/956/head 2025-12-04T08:57:03.5137468Z * [new branch] gh/anijain2305/956/orig -> origin/gh/anijain2305/956/orig 2025-12-04T08:57:03.5140021Z * [new branch] gh/anijain2305/957/base -> origin/gh/anijain2305/957/base 2025-12-04T08:57:03.5141286Z * [new branch] gh/anijain2305/957/head -> origin/gh/anijain2305/957/head 2025-12-04T08:57:03.5142926Z * [new branch] gh/anijain2305/957/orig -> origin/gh/anijain2305/957/orig 2025-12-04T08:57:03.5145474Z * [new branch] gh/anijain2305/958/base -> origin/gh/anijain2305/958/base 2025-12-04T08:57:03.5146918Z * [new branch] gh/anijain2305/958/head -> origin/gh/anijain2305/958/head 2025-12-04T08:57:03.5148707Z * [new branch] gh/anijain2305/958/orig -> origin/gh/anijain2305/958/orig 2025-12-04T08:57:03.5150665Z * [new branch] gh/anijain2305/959/base -> origin/gh/anijain2305/959/base 2025-12-04T08:57:03.5152172Z * [new branch] gh/anijain2305/959/head -> origin/gh/anijain2305/959/head 2025-12-04T08:57:03.5153876Z * [new branch] gh/anijain2305/959/orig -> origin/gh/anijain2305/959/orig 2025-12-04T08:57:03.5156469Z * [new branch] gh/anijain2305/960/base -> origin/gh/anijain2305/960/base 2025-12-04T08:57:03.5157729Z * [new branch] gh/anijain2305/960/head -> origin/gh/anijain2305/960/head 2025-12-04T08:57:03.5159331Z * [new branch] gh/anijain2305/960/orig -> origin/gh/anijain2305/960/orig 2025-12-04T08:57:03.5161967Z * [new branch] gh/anijain2305/961/base -> origin/gh/anijain2305/961/base 2025-12-04T08:57:03.5163270Z * [new branch] gh/anijain2305/961/head -> origin/gh/anijain2305/961/head 2025-12-04T08:57:03.5164783Z * [new branch] gh/anijain2305/961/orig -> origin/gh/anijain2305/961/orig 2025-12-04T08:57:03.5167316Z * [new branch] gh/anijain2305/962/base -> origin/gh/anijain2305/962/base 2025-12-04T08:57:03.5168477Z * [new branch] gh/anijain2305/962/head -> origin/gh/anijain2305/962/head 2025-12-04T08:57:03.5170052Z * [new branch] gh/anijain2305/962/orig -> origin/gh/anijain2305/962/orig 2025-12-04T08:57:03.5172894Z * [new branch] gh/anijain2305/963/base -> origin/gh/anijain2305/963/base 2025-12-04T08:57:03.5174442Z * [new branch] gh/anijain2305/963/head -> origin/gh/anijain2305/963/head 2025-12-04T08:57:03.5176065Z * [new branch] gh/anijain2305/963/orig -> origin/gh/anijain2305/963/orig 2025-12-04T08:57:03.5178834Z * [new branch] gh/anijain2305/964/base -> origin/gh/anijain2305/964/base 2025-12-04T08:57:03.5180088Z * [new branch] gh/anijain2305/964/head -> origin/gh/anijain2305/964/head 2025-12-04T08:57:03.5181603Z * [new branch] gh/anijain2305/964/orig -> origin/gh/anijain2305/964/orig 2025-12-04T08:57:03.5184121Z * [new branch] gh/anijain2305/965/base -> origin/gh/anijain2305/965/base 2025-12-04T08:57:03.5185276Z * [new branch] gh/anijain2305/965/head -> origin/gh/anijain2305/965/head 2025-12-04T08:57:03.5187031Z * [new branch] gh/anijain2305/965/orig -> origin/gh/anijain2305/965/orig 2025-12-04T08:57:03.5189439Z * [new branch] gh/anijain2305/966/base -> origin/gh/anijain2305/966/base 2025-12-04T08:57:03.5190689Z * [new branch] gh/anijain2305/966/head -> origin/gh/anijain2305/966/head 2025-12-04T08:57:03.5192248Z * [new branch] gh/anijain2305/966/orig -> origin/gh/anijain2305/966/orig 2025-12-04T08:57:03.5194844Z * [new branch] gh/anijain2305/967/base -> origin/gh/anijain2305/967/base 2025-12-04T08:57:03.5196082Z * [new branch] gh/anijain2305/967/head -> origin/gh/anijain2305/967/head 2025-12-04T08:57:03.5197710Z * [new branch] gh/anijain2305/967/orig -> origin/gh/anijain2305/967/orig 2025-12-04T08:57:03.5200238Z * [new branch] gh/anijain2305/968/base -> origin/gh/anijain2305/968/base 2025-12-04T08:57:03.5201556Z * [new branch] gh/anijain2305/968/head -> origin/gh/anijain2305/968/head 2025-12-04T08:57:03.5203180Z * [new branch] gh/anijain2305/968/orig -> origin/gh/anijain2305/968/orig 2025-12-04T08:57:03.5207232Z * [new branch] gh/anijain2305/969/base -> origin/gh/anijain2305/969/base 2025-12-04T08:57:03.5208542Z * [new branch] gh/anijain2305/969/head -> origin/gh/anijain2305/969/head 2025-12-04T08:57:03.5210123Z * [new branch] gh/anijain2305/969/orig -> origin/gh/anijain2305/969/orig 2025-12-04T08:57:03.5212916Z * [new branch] gh/anijain2305/970/base -> origin/gh/anijain2305/970/base 2025-12-04T08:57:03.5214159Z * [new branch] gh/anijain2305/970/head -> origin/gh/anijain2305/970/head 2025-12-04T08:57:03.5216134Z * [new branch] gh/anijain2305/970/orig -> origin/gh/anijain2305/970/orig 2025-12-04T08:57:03.5218416Z * [new branch] gh/anjali411/216/base -> origin/gh/anjali411/216/base 2025-12-04T08:57:03.5219875Z * [new branch] gh/anjali411/216/head -> origin/gh/anjali411/216/head 2025-12-04T08:57:03.5221674Z * [new branch] gh/anjali411/216/orig -> origin/gh/anjali411/216/orig 2025-12-04T08:57:03.5224309Z * [new branch] gh/anshul-si/1/base -> origin/gh/anshul-si/1/base 2025-12-04T08:57:03.5225928Z * [new branch] gh/anshul-si/1/head -> origin/gh/anshul-si/1/head 2025-12-04T08:57:03.5227884Z * [new branch] gh/anshul-si/2/base -> origin/gh/anshul-si/2/base 2025-12-04T08:57:03.5229473Z * [new branch] gh/anshul-si/2/head -> origin/gh/anshul-si/2/head 2025-12-04T08:57:03.5231481Z * [new branch] gh/anshul-si/3/base -> origin/gh/anshul-si/3/base 2025-12-04T08:57:03.5233059Z * [new branch] gh/anshul-si/3/head -> origin/gh/anshul-si/3/head 2025-12-04T08:57:03.5235105Z * [new branch] gh/anshul-si/4/base -> origin/gh/anshul-si/4/base 2025-12-04T08:57:03.5236669Z * [new branch] gh/anshul-si/4/head -> origin/gh/anshul-si/4/head 2025-12-04T08:57:03.5238628Z * [new branch] gh/anshul-si/5/base -> origin/gh/anshul-si/5/base 2025-12-04T08:57:03.5240641Z * [new branch] gh/anshul-si/5/head -> origin/gh/anshul-si/5/head 2025-12-04T08:57:03.5242672Z * [new branch] gh/anshul-si/53/base -> origin/gh/anshul-si/53/base 2025-12-04T08:57:03.5244228Z * [new branch] gh/anshul-si/53/head -> origin/gh/anshul-si/53/head 2025-12-04T08:57:03.5246672Z * [new branch] gh/anshul-si/58/base -> origin/gh/anshul-si/58/base 2025-12-04T08:57:03.5248033Z * [new branch] gh/anshul-si/58/head -> origin/gh/anshul-si/58/head 2025-12-04T08:57:03.5250330Z * [new branch] gh/anshul-si/66/base -> origin/gh/anshul-si/66/base 2025-12-04T08:57:03.5251566Z * [new branch] gh/anshul-si/66/head -> origin/gh/anshul-si/66/head 2025-12-04T08:57:03.5253106Z * [new branch] gh/anshul-si/66/orig -> origin/gh/anshul-si/66/orig 2025-12-04T08:57:03.5255520Z * [new branch] gh/anshul-si/67/base -> origin/gh/anshul-si/67/base 2025-12-04T08:57:03.5256740Z * [new branch] gh/anshul-si/67/head -> origin/gh/anshul-si/67/head 2025-12-04T08:57:03.5258318Z * [new branch] gh/anshul-si/67/orig -> origin/gh/anshul-si/67/orig 2025-12-04T08:57:03.5261180Z * [new branch] gh/anshul-si/68/base -> origin/gh/anshul-si/68/base 2025-12-04T08:57:03.5262436Z * [new branch] gh/anshul-si/68/head -> origin/gh/anshul-si/68/head 2025-12-04T08:57:03.5264015Z * [new branch] gh/anshul-si/68/orig -> origin/gh/anshul-si/68/orig 2025-12-04T08:57:03.5266659Z * [new branch] gh/anshul-si/69/base -> origin/gh/anshul-si/69/base 2025-12-04T08:57:03.5267770Z * [new branch] gh/anshul-si/69/head -> origin/gh/anshul-si/69/head 2025-12-04T08:57:03.5269458Z * [new branch] gh/anshul-si/69/orig -> origin/gh/anshul-si/69/orig 2025-12-04T08:57:03.5271852Z * [new branch] gh/anshul-si/70/base -> origin/gh/anshul-si/70/base 2025-12-04T08:57:03.5273132Z * [new branch] gh/anshul-si/70/head -> origin/gh/anshul-si/70/head 2025-12-04T08:57:03.5274754Z * [new branch] gh/anshul-si/70/orig -> origin/gh/anshul-si/70/orig 2025-12-04T08:57:03.5277289Z * [new branch] gh/anshul-si/71/base -> origin/gh/anshul-si/71/base 2025-12-04T08:57:03.5278473Z * [new branch] gh/anshul-si/71/head -> origin/gh/anshul-si/71/head 2025-12-04T08:57:03.5280055Z * [new branch] gh/anshul-si/71/orig -> origin/gh/anshul-si/71/orig 2025-12-04T08:57:03.5282712Z * [new branch] gh/anshul-si/72/base -> origin/gh/anshul-si/72/base 2025-12-04T08:57:03.5284036Z * [new branch] gh/anshul-si/72/head -> origin/gh/anshul-si/72/head 2025-12-04T08:57:03.5285634Z * [new branch] gh/anshul-si/72/orig -> origin/gh/anshul-si/72/orig 2025-12-04T08:57:03.5288246Z * [new branch] gh/anshul-si/73/base -> origin/gh/anshul-si/73/base 2025-12-04T08:57:03.5289589Z * [new branch] gh/anshul-si/73/head -> origin/gh/anshul-si/73/head 2025-12-04T08:57:03.5291129Z * [new branch] gh/anshul-si/73/orig -> origin/gh/anshul-si/73/orig 2025-12-04T08:57:03.5294261Z * [new branch] gh/aorenste/132/base -> origin/gh/aorenste/132/base 2025-12-04T08:57:03.5295566Z * [new branch] gh/aorenste/132/head -> origin/gh/aorenste/132/head 2025-12-04T08:57:03.5298160Z * [new branch] gh/aorenste/134/base -> origin/gh/aorenste/134/base 2025-12-04T08:57:03.5299527Z * [new branch] gh/aorenste/134/head -> origin/gh/aorenste/134/head 2025-12-04T08:57:03.5301236Z * [new branch] gh/aorenste/134/orig -> origin/gh/aorenste/134/orig 2025-12-04T08:57:03.5303866Z * [new branch] gh/aorenste/139/base -> origin/gh/aorenste/139/base 2025-12-04T08:57:03.5305334Z * [new branch] gh/aorenste/139/head -> origin/gh/aorenste/139/head 2025-12-04T08:57:03.5307035Z * [new branch] gh/aorenste/139/orig -> origin/gh/aorenste/139/orig 2025-12-04T08:57:03.5309452Z * [new branch] gh/aorenste/141/base -> origin/gh/aorenste/141/base 2025-12-04T08:57:03.5310749Z * [new branch] gh/aorenste/141/head -> origin/gh/aorenste/141/head 2025-12-04T08:57:03.5313445Z * [new branch] gh/aorenste/145/base -> origin/gh/aorenste/145/base 2025-12-04T08:57:03.5314668Z * [new branch] gh/aorenste/145/head -> origin/gh/aorenste/145/head 2025-12-04T08:57:03.5316465Z * [new branch] gh/aorenste/145/orig -> origin/gh/aorenste/145/orig 2025-12-04T08:57:03.5318789Z * [new branch] gh/aorenste/146/base -> origin/gh/aorenste/146/base 2025-12-04T08:57:03.5320335Z * [new branch] gh/aorenste/146/head -> origin/gh/aorenste/146/head 2025-12-04T08:57:03.5321955Z * [new branch] gh/aorenste/146/orig -> origin/gh/aorenste/146/orig 2025-12-04T08:57:03.5324142Z * [new branch] gh/aorenste/147/base -> origin/gh/aorenste/147/base 2025-12-04T08:57:03.5325922Z * [new branch] gh/aorenste/147/head -> origin/gh/aorenste/147/head 2025-12-04T08:57:03.5327436Z * [new branch] gh/aorenste/147/orig -> origin/gh/aorenste/147/orig 2025-12-04T08:57:03.5329549Z * [new branch] gh/aorenste/148/base -> origin/gh/aorenste/148/base 2025-12-04T08:57:03.5331180Z * [new branch] gh/aorenste/148/head -> origin/gh/aorenste/148/head 2025-12-04T08:57:03.5332777Z * [new branch] gh/aorenste/148/orig -> origin/gh/aorenste/148/orig 2025-12-04T08:57:03.5335085Z * [new branch] gh/aorenste/149/base -> origin/gh/aorenste/149/base 2025-12-04T08:57:03.5336610Z * [new branch] gh/aorenste/149/head -> origin/gh/aorenste/149/head 2025-12-04T08:57:03.5338188Z * [new branch] gh/aorenste/149/orig -> origin/gh/aorenste/149/orig 2025-12-04T08:57:03.5340294Z * [new branch] gh/aorenste/150/base -> origin/gh/aorenste/150/base 2025-12-04T08:57:03.5342040Z * [new branch] gh/aorenste/150/head -> origin/gh/aorenste/150/head 2025-12-04T08:57:03.5343498Z * [new branch] gh/aorenste/150/orig -> origin/gh/aorenste/150/orig 2025-12-04T08:57:03.5345579Z * [new branch] gh/aorenste/151/base -> origin/gh/aorenste/151/base 2025-12-04T08:57:03.5347173Z * [new branch] gh/aorenste/151/head -> origin/gh/aorenste/151/head 2025-12-04T08:57:03.5348779Z * [new branch] gh/aorenste/151/orig -> origin/gh/aorenste/151/orig 2025-12-04T08:57:03.5350915Z * [new branch] gh/aorenste/152/base -> origin/gh/aorenste/152/base 2025-12-04T08:57:03.5352475Z * [new branch] gh/aorenste/152/head -> origin/gh/aorenste/152/head 2025-12-04T08:57:03.5354059Z * [new branch] gh/aorenste/152/orig -> origin/gh/aorenste/152/orig 2025-12-04T08:57:03.5356035Z * [new branch] gh/aorenste/153/base -> origin/gh/aorenste/153/base 2025-12-04T08:57:03.5357639Z * [new branch] gh/aorenste/153/head -> origin/gh/aorenste/153/head 2025-12-04T08:57:03.5359177Z * [new branch] gh/aorenste/153/orig -> origin/gh/aorenste/153/orig 2025-12-04T08:57:03.5361138Z * [new branch] gh/aorenste/154/base -> origin/gh/aorenste/154/base 2025-12-04T08:57:03.5362741Z * [new branch] gh/aorenste/154/head -> origin/gh/aorenste/154/head 2025-12-04T08:57:03.5364572Z * [new branch] gh/aorenste/154/orig -> origin/gh/aorenste/154/orig 2025-12-04T08:57:03.5366349Z * [new branch] gh/aorenste/155/base -> origin/gh/aorenste/155/base 2025-12-04T08:57:03.5368008Z * [new branch] gh/aorenste/155/head -> origin/gh/aorenste/155/head 2025-12-04T08:57:03.5369590Z * [new branch] gh/aorenste/155/orig -> origin/gh/aorenste/155/orig 2025-12-04T08:57:03.5371702Z * [new branch] gh/aorenste/156/base -> origin/gh/aorenste/156/base 2025-12-04T08:57:03.5373155Z * [new branch] gh/aorenste/156/head -> origin/gh/aorenste/156/head 2025-12-04T08:57:03.5374688Z * [new branch] gh/aorenste/156/orig -> origin/gh/aorenste/156/orig 2025-12-04T08:57:03.5377241Z * [new branch] gh/aorenste/157/base -> origin/gh/aorenste/157/base 2025-12-04T08:57:03.5378921Z * [new branch] gh/aorenste/157/head -> origin/gh/aorenste/157/head 2025-12-04T08:57:03.5380394Z * [new branch] gh/aorenste/157/orig -> origin/gh/aorenste/157/orig 2025-12-04T08:57:03.5382537Z * [new branch] gh/aorenste/158/base -> origin/gh/aorenste/158/base 2025-12-04T08:57:03.5384241Z * [new branch] gh/aorenste/158/head -> origin/gh/aorenste/158/head 2025-12-04T08:57:03.5385768Z * [new branch] gh/aorenste/158/orig -> origin/gh/aorenste/158/orig 2025-12-04T08:57:03.5387817Z * [new branch] gh/aorenste/159/base -> origin/gh/aorenste/159/base 2025-12-04T08:57:03.5389517Z * [new branch] gh/aorenste/159/head -> origin/gh/aorenste/159/head 2025-12-04T08:57:03.5391073Z * [new branch] gh/aorenste/159/orig -> origin/gh/aorenste/159/orig 2025-12-04T08:57:03.5393716Z * [new branch] gh/avikchaudhuri/1/base -> origin/gh/avikchaudhuri/1/base 2025-12-04T08:57:03.5395245Z * [new branch] gh/avikchaudhuri/1/head -> origin/gh/avikchaudhuri/1/head 2025-12-04T08:57:03.5397264Z * [new branch] gh/avikchaudhuri/2/base -> origin/gh/avikchaudhuri/2/base 2025-12-04T08:57:03.5398837Z * [new branch] gh/avikchaudhuri/2/head -> origin/gh/avikchaudhuri/2/head 2025-12-04T08:57:03.5400319Z * [new branch] gh/avikchaudhuri/2/orig -> origin/gh/avikchaudhuri/2/orig 2025-12-04T08:57:03.5403198Z * [new branch] gh/bdhirsh/666/base -> origin/gh/bdhirsh/666/base 2025-12-04T08:57:03.5405066Z * [new branch] gh/bdhirsh/666/head -> origin/gh/bdhirsh/666/head 2025-12-04T08:57:03.5406817Z * [new branch] gh/bdhirsh/666/orig -> origin/gh/bdhirsh/666/orig 2025-12-04T08:57:03.5408935Z * [new branch] gh/bdhirsh/668/base -> origin/gh/bdhirsh/668/base 2025-12-04T08:57:03.5410547Z * [new branch] gh/bdhirsh/668/head -> origin/gh/bdhirsh/668/head 2025-12-04T08:57:03.5412071Z * [new branch] gh/bdhirsh/668/orig -> origin/gh/bdhirsh/668/orig 2025-12-04T08:57:03.5414596Z * [new branch] gh/bdhirsh/669/base -> origin/gh/bdhirsh/669/base 2025-12-04T08:57:03.5416058Z * [new branch] gh/bdhirsh/669/head -> origin/gh/bdhirsh/669/head 2025-12-04T08:57:03.5417575Z * [new branch] gh/bdhirsh/669/orig -> origin/gh/bdhirsh/669/orig 2025-12-04T08:57:03.5419896Z * [new branch] gh/bdhirsh/670/base -> origin/gh/bdhirsh/670/base 2025-12-04T08:57:03.5421585Z * [new branch] gh/bdhirsh/670/head -> origin/gh/bdhirsh/670/head 2025-12-04T08:57:03.5423230Z * [new branch] gh/bdhirsh/670/orig -> origin/gh/bdhirsh/670/orig 2025-12-04T08:57:03.5425387Z * [new branch] gh/bdhirsh/672/base -> origin/gh/bdhirsh/672/base 2025-12-04T08:57:03.5426784Z * [new branch] gh/bdhirsh/672/head -> origin/gh/bdhirsh/672/head 2025-12-04T08:57:03.5428481Z * [new branch] gh/bdhirsh/672/orig -> origin/gh/bdhirsh/672/orig 2025-12-04T08:57:03.5430903Z * [new branch] gh/bdhirsh/675/base -> origin/gh/bdhirsh/675/base 2025-12-04T08:57:03.5432580Z * [new branch] gh/bdhirsh/675/head -> origin/gh/bdhirsh/675/head 2025-12-04T08:57:03.5434302Z * [new branch] gh/bdhirsh/675/orig -> origin/gh/bdhirsh/675/orig 2025-12-04T08:57:03.5436382Z * [new branch] gh/bdhirsh/676/base -> origin/gh/bdhirsh/676/base 2025-12-04T08:57:03.5438055Z * [new branch] gh/bdhirsh/676/head -> origin/gh/bdhirsh/676/head 2025-12-04T08:57:03.5439637Z * [new branch] gh/bdhirsh/676/orig -> origin/gh/bdhirsh/676/orig 2025-12-04T08:57:03.5441747Z * [new branch] gh/bdhirsh/677/base -> origin/gh/bdhirsh/677/base 2025-12-04T08:57:03.5443604Z * [new branch] gh/bdhirsh/677/head -> origin/gh/bdhirsh/677/head 2025-12-04T08:57:03.5445154Z * [new branch] gh/bdhirsh/677/orig -> origin/gh/bdhirsh/677/orig 2025-12-04T08:57:03.5447392Z * [new branch] gh/bdhirsh/678/base -> origin/gh/bdhirsh/678/base 2025-12-04T08:57:03.5449006Z * [new branch] gh/bdhirsh/678/head -> origin/gh/bdhirsh/678/head 2025-12-04T08:57:03.5450690Z * [new branch] gh/bdhirsh/678/orig -> origin/gh/bdhirsh/678/orig 2025-12-04T08:57:03.5452953Z * [new branch] gh/bdhirsh/679/base -> origin/gh/bdhirsh/679/base 2025-12-04T08:57:03.5454827Z * [new branch] gh/bdhirsh/679/head -> origin/gh/bdhirsh/679/head 2025-12-04T08:57:03.5456332Z * [new branch] gh/bdhirsh/679/orig -> origin/gh/bdhirsh/679/orig 2025-12-04T08:57:03.5458537Z * [new branch] gh/bdhirsh/680/base -> origin/gh/bdhirsh/680/base 2025-12-04T08:57:03.5460187Z * [new branch] gh/bdhirsh/680/head -> origin/gh/bdhirsh/680/head 2025-12-04T08:57:03.5461758Z * [new branch] gh/bdhirsh/680/orig -> origin/gh/bdhirsh/680/orig 2025-12-04T08:57:03.5463770Z * [new branch] gh/bdhirsh/681/base -> origin/gh/bdhirsh/681/base 2025-12-04T08:57:03.5465432Z * [new branch] gh/bdhirsh/681/head -> origin/gh/bdhirsh/681/head 2025-12-04T08:57:03.5466977Z * [new branch] gh/bdhirsh/681/orig -> origin/gh/bdhirsh/681/orig 2025-12-04T08:57:03.5469735Z * [new branch] gh/benjaminglass1/101/base -> origin/gh/benjaminglass1/101/base 2025-12-04T08:57:03.5471063Z * [new branch] gh/benjaminglass1/101/head -> origin/gh/benjaminglass1/101/head 2025-12-04T08:57:03.5472856Z * [new branch] gh/benjaminglass1/101/orig -> origin/gh/benjaminglass1/101/orig 2025-12-04T08:57:03.5475431Z * [new branch] gh/benjaminglass1/102/base -> origin/gh/benjaminglass1/102/base 2025-12-04T08:57:03.5477134Z * [new branch] gh/benjaminglass1/102/head -> origin/gh/benjaminglass1/102/head 2025-12-04T08:57:03.5478755Z * [new branch] gh/benjaminglass1/102/orig -> origin/gh/benjaminglass1/102/orig 2025-12-04T08:57:03.5480885Z * [new branch] gh/benjaminglass1/106/base -> origin/gh/benjaminglass1/106/base 2025-12-04T08:57:03.5482431Z * [new branch] gh/benjaminglass1/106/head -> origin/gh/benjaminglass1/106/head 2025-12-04T08:57:03.5484245Z * [new branch] gh/benjaminglass1/106/orig -> origin/gh/benjaminglass1/106/orig 2025-12-04T08:57:03.5486506Z * [new branch] gh/benjaminglass1/107/base -> origin/gh/benjaminglass1/107/base 2025-12-04T08:57:03.5488054Z * [new branch] gh/benjaminglass1/107/head -> origin/gh/benjaminglass1/107/head 2025-12-04T08:57:03.5489664Z * [new branch] gh/benjaminglass1/107/orig -> origin/gh/benjaminglass1/107/orig 2025-12-04T08:57:03.5491775Z * [new branch] gh/benjaminglass1/108/base -> origin/gh/benjaminglass1/108/base 2025-12-04T08:57:03.5493447Z * [new branch] gh/benjaminglass1/108/head -> origin/gh/benjaminglass1/108/head 2025-12-04T08:57:03.5495021Z * [new branch] gh/benjaminglass1/108/orig -> origin/gh/benjaminglass1/108/orig 2025-12-04T08:57:03.5497178Z * [new branch] gh/benjaminglass1/109/base -> origin/gh/benjaminglass1/109/base 2025-12-04T08:57:03.5498738Z * [new branch] gh/benjaminglass1/109/head -> origin/gh/benjaminglass1/109/head 2025-12-04T08:57:03.5500348Z * [new branch] gh/benjaminglass1/109/orig -> origin/gh/benjaminglass1/109/orig 2025-12-04T08:57:03.5502418Z * [new branch] gh/benjaminglass1/97/base -> origin/gh/benjaminglass1/97/base 2025-12-04T08:57:03.5504047Z * [new branch] gh/benjaminglass1/97/head -> origin/gh/benjaminglass1/97/head 2025-12-04T08:57:03.5506077Z * [new branch] gh/benjaminglass1/97/orig -> origin/gh/benjaminglass1/97/orig 2025-12-04T08:57:03.5508493Z * [new branch] gh/bobrenjc93/570/base -> origin/gh/bobrenjc93/570/base 2025-12-04T08:57:03.5510096Z * [new branch] gh/bobrenjc93/570/head -> origin/gh/bobrenjc93/570/head 2025-12-04T08:57:03.5511618Z * [new branch] gh/bobrenjc93/570/orig -> origin/gh/bobrenjc93/570/orig 2025-12-04T08:57:03.5513613Z * [new branch] gh/bobrenjc93/604/base -> origin/gh/bobrenjc93/604/base 2025-12-04T08:57:03.5515103Z * [new branch] gh/bobrenjc93/604/head -> origin/gh/bobrenjc93/604/head 2025-12-04T08:57:03.5516734Z * [new branch] gh/bobrenjc93/604/orig -> origin/gh/bobrenjc93/604/orig 2025-12-04T08:57:03.5518922Z * [new branch] gh/bobrenjc93/638/base -> origin/gh/bobrenjc93/638/base 2025-12-04T08:57:03.5520510Z * [new branch] gh/bobrenjc93/638/head -> origin/gh/bobrenjc93/638/head 2025-12-04T08:57:03.5522070Z * [new branch] gh/bobrenjc93/638/orig -> origin/gh/bobrenjc93/638/orig 2025-12-04T08:57:03.5524293Z * [new branch] gh/bobrenjc93/653/base -> origin/gh/bobrenjc93/653/base 2025-12-04T08:57:03.5525791Z * [new branch] gh/bobrenjc93/653/head -> origin/gh/bobrenjc93/653/head 2025-12-04T08:57:03.5527364Z * [new branch] gh/bobrenjc93/653/orig -> origin/gh/bobrenjc93/653/orig 2025-12-04T08:57:03.5529555Z * [new branch] gh/bobrenjc93/654/base -> origin/gh/bobrenjc93/654/base 2025-12-04T08:57:03.5531267Z * [new branch] gh/bobrenjc93/654/head -> origin/gh/bobrenjc93/654/head 2025-12-04T08:57:03.5532646Z * [new branch] gh/bobrenjc93/654/orig -> origin/gh/bobrenjc93/654/orig 2025-12-04T08:57:03.5535309Z * [new branch] gh/bobrenjc93/657/base -> origin/gh/bobrenjc93/657/base 2025-12-04T08:57:03.5536435Z * [new branch] gh/bobrenjc93/657/head -> origin/gh/bobrenjc93/657/head 2025-12-04T08:57:03.5538249Z * [new branch] gh/bobrenjc93/657/orig -> origin/gh/bobrenjc93/657/orig 2025-12-04T08:57:03.5540370Z * [new branch] gh/bobrenjc93/672/base -> origin/gh/bobrenjc93/672/base 2025-12-04T08:57:03.5541924Z * [new branch] gh/bobrenjc93/672/head -> origin/gh/bobrenjc93/672/head 2025-12-04T08:57:03.5543536Z * [new branch] gh/bobrenjc93/672/orig -> origin/gh/bobrenjc93/672/orig 2025-12-04T08:57:03.5545554Z * [new branch] gh/bobrenjc93/679/base -> origin/gh/bobrenjc93/679/base 2025-12-04T08:57:03.5547394Z * [new branch] gh/bobrenjc93/679/head -> origin/gh/bobrenjc93/679/head 2025-12-04T08:57:03.5549014Z * [new branch] gh/bobrenjc93/679/orig -> origin/gh/bobrenjc93/679/orig 2025-12-04T08:57:03.5551154Z * [new branch] gh/bobrenjc93/680/base -> origin/gh/bobrenjc93/680/base 2025-12-04T08:57:03.5552725Z * [new branch] gh/bobrenjc93/680/head -> origin/gh/bobrenjc93/680/head 2025-12-04T08:57:03.5554319Z * [new branch] gh/bobrenjc93/680/orig -> origin/gh/bobrenjc93/680/orig 2025-12-04T08:57:03.5556297Z * [new branch] gh/bobrenjc93/681/base -> origin/gh/bobrenjc93/681/base 2025-12-04T08:57:03.5557874Z * [new branch] gh/bobrenjc93/681/head -> origin/gh/bobrenjc93/681/head 2025-12-04T08:57:03.5559571Z * [new branch] gh/bobrenjc93/681/orig -> origin/gh/bobrenjc93/681/orig 2025-12-04T08:57:03.5561556Z * [new branch] gh/bobrenjc93/682/base -> origin/gh/bobrenjc93/682/base 2025-12-04T08:57:03.5563107Z * [new branch] gh/bobrenjc93/682/head -> origin/gh/bobrenjc93/682/head 2025-12-04T08:57:03.5564613Z * [new branch] gh/bobrenjc93/682/orig -> origin/gh/bobrenjc93/682/orig 2025-12-04T08:57:03.5566693Z * [new branch] gh/bobrenjc93/683/base -> origin/gh/bobrenjc93/683/base 2025-12-04T08:57:03.5568305Z * [new branch] gh/bobrenjc93/683/head -> origin/gh/bobrenjc93/683/head 2025-12-04T08:57:03.5569868Z * [new branch] gh/bobrenjc93/683/orig -> origin/gh/bobrenjc93/683/orig 2025-12-04T08:57:03.5572025Z * [new branch] gh/bobrenjc93/684/base -> origin/gh/bobrenjc93/684/base 2025-12-04T08:57:03.5573883Z * [new branch] gh/bobrenjc93/684/head -> origin/gh/bobrenjc93/684/head 2025-12-04T08:57:03.5575945Z * [new branch] gh/bobrenjc93/684/orig -> origin/gh/bobrenjc93/684/orig 2025-12-04T08:57:03.5577645Z * [new branch] gh/bobrenjc93/685/base -> origin/gh/bobrenjc93/685/base 2025-12-04T08:57:03.5579347Z * [new branch] gh/bobrenjc93/685/head -> origin/gh/bobrenjc93/685/head 2025-12-04T08:57:03.5581239Z * [new branch] gh/bobrenjc93/685/orig -> origin/gh/bobrenjc93/685/orig 2025-12-04T08:57:03.5583592Z * [new branch] gh/bobrenjc93/686/base -> origin/gh/bobrenjc93/686/base 2025-12-04T08:57:03.5585119Z * [new branch] gh/bobrenjc93/686/head -> origin/gh/bobrenjc93/686/head 2025-12-04T08:57:03.5586752Z * [new branch] gh/bobrenjc93/686/orig -> origin/gh/bobrenjc93/686/orig 2025-12-04T08:57:03.5589799Z * [new branch] gh/bobrenjc93/687/base -> origin/gh/bobrenjc93/687/base 2025-12-04T08:57:03.5590496Z * [new branch] gh/bobrenjc93/687/head -> origin/gh/bobrenjc93/687/head 2025-12-04T08:57:03.5592154Z * [new branch] gh/bobrenjc93/687/orig -> origin/gh/bobrenjc93/687/orig 2025-12-04T08:57:03.5594684Z * [new branch] gh/bobrenjc93/688/base -> origin/gh/bobrenjc93/688/base 2025-12-04T08:57:03.5596270Z * [new branch] gh/bobrenjc93/688/head -> origin/gh/bobrenjc93/688/head 2025-12-04T08:57:03.5597806Z * [new branch] gh/bobrenjc93/688/orig -> origin/gh/bobrenjc93/688/orig 2025-12-04T08:57:03.5599891Z * [new branch] gh/bobrenjc93/689/base -> origin/gh/bobrenjc93/689/base 2025-12-04T08:57:03.5601548Z * [new branch] gh/bobrenjc93/689/head -> origin/gh/bobrenjc93/689/head 2025-12-04T08:57:03.5603362Z * [new branch] gh/bobrenjc93/689/orig -> origin/gh/bobrenjc93/689/orig 2025-12-04T08:57:03.5605760Z * [new branch] gh/bobrenjc93/690/base -> origin/gh/bobrenjc93/690/base 2025-12-04T08:57:03.5607283Z * [new branch] gh/bobrenjc93/690/head -> origin/gh/bobrenjc93/690/head 2025-12-04T08:57:03.5608865Z * [new branch] gh/bobrenjc93/690/orig -> origin/gh/bobrenjc93/690/orig 2025-12-04T08:57:03.5611669Z * [new branch] gh/bobrenjc93/691/base -> origin/gh/bobrenjc93/691/base 2025-12-04T08:57:03.5613674Z * [new branch] gh/bobrenjc93/691/head -> origin/gh/bobrenjc93/691/head 2025-12-04T08:57:03.5615755Z * [new branch] gh/bobrenjc93/691/orig -> origin/gh/bobrenjc93/691/orig 2025-12-04T08:57:03.5618441Z * [new branch] gh/bobrenjc93/692/base -> origin/gh/bobrenjc93/692/base 2025-12-04T08:57:03.5619819Z * [new branch] gh/bobrenjc93/692/head -> origin/gh/bobrenjc93/692/head 2025-12-04T08:57:03.5621377Z * [new branch] gh/bobrenjc93/692/orig -> origin/gh/bobrenjc93/692/orig 2025-12-04T08:57:03.5623718Z * [new branch] gh/bobrenjc93/693/base -> origin/gh/bobrenjc93/693/base 2025-12-04T08:57:03.5625043Z * [new branch] gh/bobrenjc93/693/head -> origin/gh/bobrenjc93/693/head 2025-12-04T08:57:03.5626717Z * [new branch] gh/bobrenjc93/693/orig -> origin/gh/bobrenjc93/693/orig 2025-12-04T08:57:03.5629266Z * [new branch] gh/bobrenjc93/694/base -> origin/gh/bobrenjc93/694/base 2025-12-04T08:57:03.5630588Z * [new branch] gh/bobrenjc93/694/head -> origin/gh/bobrenjc93/694/head 2025-12-04T08:57:03.5632120Z * [new branch] gh/bobrenjc93/694/orig -> origin/gh/bobrenjc93/694/orig 2025-12-04T08:57:03.5634223Z * [new branch] gh/bobrenjc93/695/base -> origin/gh/bobrenjc93/695/base 2025-12-04T08:57:03.5635760Z * [new branch] gh/bobrenjc93/695/head -> origin/gh/bobrenjc93/695/head 2025-12-04T08:57:03.5637298Z * [new branch] gh/bobrenjc93/695/orig -> origin/gh/bobrenjc93/695/orig 2025-12-04T08:57:03.5639911Z * [new branch] gh/c00w/23/base -> origin/gh/c00w/23/base 2025-12-04T08:57:03.5641474Z * [new branch] gh/c00w/23/head -> origin/gh/c00w/23/head 2025-12-04T08:57:03.5643616Z * [new branch] gh/c00w/53/base -> origin/gh/c00w/53/base 2025-12-04T08:57:03.5645109Z * [new branch] gh/c00w/53/head -> origin/gh/c00w/53/head 2025-12-04T08:57:03.5646727Z * [new branch] gh/c00w/53/orig -> origin/gh/c00w/53/orig 2025-12-04T08:57:03.5648664Z * [new branch] gh/c00w/54/base -> origin/gh/c00w/54/base 2025-12-04T08:57:03.5650237Z * [new branch] gh/c00w/54/head -> origin/gh/c00w/54/head 2025-12-04T08:57:03.5651945Z * [new branch] gh/c00w/54/orig -> origin/gh/c00w/54/orig 2025-12-04T08:57:03.5654337Z * [new branch] gh/c00w/56/base -> origin/gh/c00w/56/base 2025-12-04T08:57:03.5655827Z * [new branch] gh/c00w/56/head -> origin/gh/c00w/56/head 2025-12-04T08:57:03.5657389Z * [new branch] gh/c00w/56/orig -> origin/gh/c00w/56/orig 2025-12-04T08:57:03.5659628Z * [new branch] gh/c00w/57/base -> origin/gh/c00w/57/base 2025-12-04T08:57:03.5661152Z * [new branch] gh/c00w/57/head -> origin/gh/c00w/57/head 2025-12-04T08:57:03.5662930Z * [new branch] gh/c00w/57/orig -> origin/gh/c00w/57/orig 2025-12-04T08:57:03.5665305Z * [new branch] gh/c00w/58/base -> origin/gh/c00w/58/base 2025-12-04T08:57:03.5666903Z * [new branch] gh/c00w/58/head -> origin/gh/c00w/58/head 2025-12-04T08:57:03.5668543Z * [new branch] gh/c00w/58/orig -> origin/gh/c00w/58/orig 2025-12-04T08:57:03.5671264Z * [new branch] gh/clee2000/1/base -> origin/gh/clee2000/1/base 2025-12-04T08:57:03.5672921Z * [new branch] gh/clee2000/1/head -> origin/gh/clee2000/1/head 2025-12-04T08:57:03.5674736Z * [new branch] gh/clee2000/1/orig -> origin/gh/clee2000/1/orig 2025-12-04T08:57:03.5677477Z * [new branch] gh/coconutruben/1/base -> origin/gh/coconutruben/1/base 2025-12-04T08:57:03.5690418Z * [new branch] gh/coconutruben/1/head -> origin/gh/coconutruben/1/head 2025-12-04T08:57:03.5690998Z * [new branch] gh/coconutruben/55/base -> origin/gh/coconutruben/55/base 2025-12-04T08:57:03.5691529Z * [new branch] gh/coconutruben/55/head -> origin/gh/coconutruben/55/head 2025-12-04T08:57:03.5692060Z * [new branch] gh/coconutruben/55/orig -> origin/gh/coconutruben/55/orig 2025-12-04T08:57:03.5692563Z * [new branch] gh/coconutruben/57/base -> origin/gh/coconutruben/57/base 2025-12-04T08:57:03.5693056Z * [new branch] gh/coconutruben/57/head -> origin/gh/coconutruben/57/head 2025-12-04T08:57:03.5693611Z * [new branch] gh/coconutruben/57/orig -> origin/gh/coconutruben/57/orig 2025-12-04T08:57:03.5694288Z * [new branch] gh/coconutruben/70/base -> origin/gh/coconutruben/70/base 2025-12-04T08:57:03.5694788Z * [new branch] gh/coconutruben/70/head -> origin/gh/coconutruben/70/head 2025-12-04T08:57:03.5695691Z * [new branch] gh/coconutruben/70/orig -> origin/gh/coconutruben/70/orig 2025-12-04T08:57:03.5697848Z * [new branch] gh/coconutruben/71/base -> origin/gh/coconutruben/71/base 2025-12-04T08:57:03.5699390Z * [new branch] gh/coconutruben/71/head -> origin/gh/coconutruben/71/head 2025-12-04T08:57:03.5701101Z * [new branch] gh/coconutruben/71/orig -> origin/gh/coconutruben/71/orig 2025-12-04T08:57:03.5703108Z * [new branch] gh/coconutruben/72/base -> origin/gh/coconutruben/72/base 2025-12-04T08:57:03.5705276Z * [new branch] gh/coconutruben/72/head -> origin/gh/coconutruben/72/head 2025-12-04T08:57:03.5709404Z * [new branch] gh/coconutruben/72/orig -> origin/gh/coconutruben/72/orig 2025-12-04T08:57:03.5711326Z * [new branch] gh/coconutruben/73/base -> origin/gh/coconutruben/73/base 2025-12-04T08:57:03.5712991Z * [new branch] gh/coconutruben/73/head -> origin/gh/coconutruben/73/head 2025-12-04T08:57:03.5714467Z * [new branch] gh/coconutruben/73/orig -> origin/gh/coconutruben/73/orig 2025-12-04T08:57:03.5716724Z * [new branch] gh/coconutruben/74/base -> origin/gh/coconutruben/74/base 2025-12-04T08:57:03.5718333Z * [new branch] gh/coconutruben/74/head -> origin/gh/coconutruben/74/head 2025-12-04T08:57:03.5719972Z * [new branch] gh/coconutruben/74/orig -> origin/gh/coconutruben/74/orig 2025-12-04T08:57:03.5722254Z * [new branch] gh/coconutruben/79/base -> origin/gh/coconutruben/79/base 2025-12-04T08:57:03.5723935Z * [new branch] gh/coconutruben/79/head -> origin/gh/coconutruben/79/head 2025-12-04T08:57:03.5725721Z * [new branch] gh/coconutruben/79/orig -> origin/gh/coconutruben/79/orig 2025-12-04T08:57:03.5727970Z * [new branch] gh/coconutruben/80/base -> origin/gh/coconutruben/80/base 2025-12-04T08:57:03.5729118Z * [new branch] gh/coconutruben/80/head -> origin/gh/coconutruben/80/head 2025-12-04T08:57:03.5730836Z * [new branch] gh/coconutruben/80/orig -> origin/gh/coconutruben/80/orig 2025-12-04T08:57:03.5733048Z * [new branch] gh/coconutruben/82/base -> origin/gh/coconutruben/82/base 2025-12-04T08:57:03.5734914Z * [new branch] gh/coconutruben/82/head -> origin/gh/coconutruben/82/head 2025-12-04T08:57:03.5736381Z * [new branch] gh/coconutruben/82/orig -> origin/gh/coconutruben/82/orig 2025-12-04T08:57:03.5738602Z * [new branch] gh/coconutruben/83/base -> origin/gh/coconutruben/83/base 2025-12-04T08:57:03.5740180Z * [new branch] gh/coconutruben/83/head -> origin/gh/coconutruben/83/head 2025-12-04T08:57:03.5741687Z * [new branch] gh/coconutruben/83/orig -> origin/gh/coconutruben/83/orig 2025-12-04T08:57:03.5743897Z * [new branch] gh/coconutruben/84/base -> origin/gh/coconutruben/84/base 2025-12-04T08:57:03.5745505Z * [new branch] gh/coconutruben/84/head -> origin/gh/coconutruben/84/head 2025-12-04T08:57:03.5747083Z * [new branch] gh/coconutruben/84/orig -> origin/gh/coconutruben/84/orig 2025-12-04T08:57:03.5749184Z * [new branch] gh/coconutruben/85/base -> origin/gh/coconutruben/85/base 2025-12-04T08:57:03.5750811Z * [new branch] gh/coconutruben/85/head -> origin/gh/coconutruben/85/head 2025-12-04T08:57:03.5752420Z * [new branch] gh/coconutruben/85/orig -> origin/gh/coconutruben/85/orig 2025-12-04T08:57:03.5754595Z * [new branch] gh/coconutruben/86/base -> origin/gh/coconutruben/86/base 2025-12-04T08:57:03.5756287Z * [new branch] gh/coconutruben/86/head -> origin/gh/coconutruben/86/head 2025-12-04T08:57:03.5757835Z * [new branch] gh/coconutruben/86/orig -> origin/gh/coconutruben/86/orig 2025-12-04T08:57:03.5760437Z * [new branch] gh/colinchan15/1/base -> origin/gh/colinchan15/1/base 2025-12-04T08:57:03.5762068Z * [new branch] gh/colinchan15/1/head -> origin/gh/colinchan15/1/head 2025-12-04T08:57:03.5764061Z * [new branch] gh/colinchan15/2/base -> origin/gh/colinchan15/2/base 2025-12-04T08:57:03.5765647Z * [new branch] gh/colinchan15/2/head -> origin/gh/colinchan15/2/head 2025-12-04T08:57:03.5767623Z * [new branch] gh/colinchan15/3/base -> origin/gh/colinchan15/3/base 2025-12-04T08:57:03.5769095Z * [new branch] gh/colinchan15/3/head -> origin/gh/colinchan15/3/head 2025-12-04T08:57:03.5771079Z * [new branch] gh/colinchan15/6/base -> origin/gh/colinchan15/6/base 2025-12-04T08:57:03.5772646Z * [new branch] gh/colinchan15/6/head -> origin/gh/colinchan15/6/head 2025-12-04T08:57:03.5775424Z * [new branch] gh/d4l3k/1/base -> origin/gh/d4l3k/1/base 2025-12-04T08:57:03.5776962Z * [new branch] gh/d4l3k/1/head -> origin/gh/d4l3k/1/head 2025-12-04T08:57:03.5779097Z * [new branch] gh/d4l3k/2/base -> origin/gh/d4l3k/2/base 2025-12-04T08:57:03.5780671Z * [new branch] gh/d4l3k/2/head -> origin/gh/d4l3k/2/head 2025-12-04T08:57:03.5782255Z * [new branch] gh/d4l3k/2/orig -> origin/gh/d4l3k/2/orig 2025-12-04T08:57:03.5784635Z * [new branch] gh/d4l3k/3/base -> origin/gh/d4l3k/3/base 2025-12-04T08:57:03.5786239Z * [new branch] gh/d4l3k/3/head -> origin/gh/d4l3k/3/head 2025-12-04T08:57:03.5787825Z * [new branch] gh/d4l3k/3/orig -> origin/gh/d4l3k/3/orig 2025-12-04T08:57:03.5789957Z * [new branch] gh/d4l3k/4/base -> origin/gh/d4l3k/4/base 2025-12-04T08:57:03.5791532Z * [new branch] gh/d4l3k/4/head -> origin/gh/d4l3k/4/head 2025-12-04T08:57:03.5793037Z * [new branch] gh/d4l3k/4/orig -> origin/gh/d4l3k/4/orig 2025-12-04T08:57:03.5795125Z * [new branch] gh/d4l3k/5/base -> origin/gh/d4l3k/5/base 2025-12-04T08:57:03.5796717Z * [new branch] gh/d4l3k/5/orig -> origin/gh/d4l3k/5/orig 2025-12-04T08:57:03.5799885Z * [new branch] gh/davidberard98/392/base -> origin/gh/davidberard98/392/base 2025-12-04T08:57:03.5801503Z * [new branch] gh/davidberard98/392/head -> origin/gh/davidberard98/392/head 2025-12-04T08:57:03.5803146Z * [new branch] gh/davidberard98/392/orig -> origin/gh/davidberard98/392/orig 2025-12-04T08:57:03.5805837Z * [new branch] gh/davidberard98/399/base -> origin/gh/davidberard98/399/base 2025-12-04T08:57:03.5807510Z * [new branch] gh/davidberard98/399/head -> origin/gh/davidberard98/399/head 2025-12-04T08:57:03.5809132Z * [new branch] gh/davidberard98/399/orig -> origin/gh/davidberard98/399/orig 2025-12-04T08:57:03.5811593Z * [new branch] gh/desertfire/605/base -> origin/gh/desertfire/605/base 2025-12-04T08:57:03.5813337Z * [new branch] gh/desertfire/605/head -> origin/gh/desertfire/605/head 2025-12-04T08:57:03.5815313Z * [new branch] gh/desertfire/605/orig -> origin/gh/desertfire/605/orig 2025-12-04T08:57:03.5817469Z * [new branch] gh/desertfire/606/base -> origin/gh/desertfire/606/base 2025-12-04T08:57:03.5818977Z * [new branch] gh/desertfire/606/head -> origin/gh/desertfire/606/head 2025-12-04T08:57:03.5820823Z * [new branch] gh/desertfire/606/orig -> origin/gh/desertfire/606/orig 2025-12-04T08:57:03.5822902Z * [new branch] gh/desertfire/607/base -> origin/gh/desertfire/607/base 2025-12-04T08:57:03.5824511Z * [new branch] gh/desertfire/607/head -> origin/gh/desertfire/607/head 2025-12-04T08:57:03.5826041Z * [new branch] gh/desertfire/607/orig -> origin/gh/desertfire/607/orig 2025-12-04T08:57:03.5828144Z * [new branch] gh/desertfire/608/base -> origin/gh/desertfire/608/base 2025-12-04T08:57:03.5829700Z * [new branch] gh/desertfire/608/head -> origin/gh/desertfire/608/head 2025-12-04T08:57:03.5831302Z * [new branch] gh/desertfire/608/orig -> origin/gh/desertfire/608/orig 2025-12-04T08:57:03.5833408Z * [new branch] gh/desertfire/609/base -> origin/gh/desertfire/609/base 2025-12-04T08:57:03.5834926Z * [new branch] gh/desertfire/609/head -> origin/gh/desertfire/609/head 2025-12-04T08:57:03.5836570Z * [new branch] gh/desertfire/609/orig -> origin/gh/desertfire/609/orig 2025-12-04T08:57:03.5838888Z * [new branch] gh/desertfire/610/base -> origin/gh/desertfire/610/base 2025-12-04T08:57:03.5840520Z * [new branch] gh/desertfire/610/head -> origin/gh/desertfire/610/head 2025-12-04T08:57:03.5842105Z * [new branch] gh/desertfire/610/orig -> origin/gh/desertfire/610/orig 2025-12-04T08:57:03.5844269Z * [new branch] gh/desertfire/611/base -> origin/gh/desertfire/611/base 2025-12-04T08:57:03.5845875Z * [new branch] gh/desertfire/611/head -> origin/gh/desertfire/611/head 2025-12-04T08:57:03.5847494Z * [new branch] gh/desertfire/611/orig -> origin/gh/desertfire/611/orig 2025-12-04T08:57:03.5849719Z * [new branch] gh/desertfire/612/base -> origin/gh/desertfire/612/base 2025-12-04T08:57:03.5851258Z * [new branch] gh/desertfire/612/head -> origin/gh/desertfire/612/head 2025-12-04T08:57:03.5852843Z * [new branch] gh/desertfire/612/orig -> origin/gh/desertfire/612/orig 2025-12-04T08:57:03.5855293Z * [new branch] gh/desertfire/613/base -> origin/gh/desertfire/613/base 2025-12-04T08:57:03.5856943Z * [new branch] gh/desertfire/613/head -> origin/gh/desertfire/613/head 2025-12-04T08:57:03.5858494Z * [new branch] gh/desertfire/613/orig -> origin/gh/desertfire/613/orig 2025-12-04T08:57:03.5860845Z * [new branch] gh/desertfire/614/base -> origin/gh/desertfire/614/base 2025-12-04T08:57:03.5862508Z * [new branch] gh/desertfire/614/head -> origin/gh/desertfire/614/head 2025-12-04T08:57:03.5864186Z * [new branch] gh/desertfire/614/orig -> origin/gh/desertfire/614/orig 2025-12-04T08:57:03.5866339Z * [new branch] gh/desertfire/615/base -> origin/gh/desertfire/615/base 2025-12-04T08:57:03.5868094Z * [new branch] gh/desertfire/615/head -> origin/gh/desertfire/615/head 2025-12-04T08:57:03.5869658Z * [new branch] gh/desertfire/615/orig -> origin/gh/desertfire/615/orig 2025-12-04T08:57:03.5871631Z * [new branch] gh/desertfire/616/base -> origin/gh/desertfire/616/base 2025-12-04T08:57:03.5873293Z * [new branch] gh/desertfire/616/head -> origin/gh/desertfire/616/head 2025-12-04T08:57:03.5874747Z * [new branch] gh/desertfire/616/orig -> origin/gh/desertfire/616/orig 2025-12-04T08:57:03.5876728Z * [new branch] gh/desertfire/617/base -> origin/gh/desertfire/617/base 2025-12-04T08:57:03.5878364Z * [new branch] gh/desertfire/617/head -> origin/gh/desertfire/617/head 2025-12-04T08:57:03.5879901Z * [new branch] gh/desertfire/617/orig -> origin/gh/desertfire/617/orig 2025-12-04T08:57:03.5882518Z * [new branch] gh/dharakk/1/base -> origin/gh/dharakk/1/base 2025-12-04T08:57:03.5884081Z * [new branch] gh/dharakk/1/head -> origin/gh/dharakk/1/head 2025-12-04T08:57:03.5886661Z * [new branch] gh/drisspg/170/base -> origin/gh/drisspg/170/base 2025-12-04T08:57:03.5888208Z * [new branch] gh/drisspg/170/head -> origin/gh/drisspg/170/head 2025-12-04T08:57:03.5889836Z * [new branch] gh/drisspg/170/orig -> origin/gh/drisspg/170/orig 2025-12-04T08:57:03.5891902Z * [new branch] gh/drisspg/182/base -> origin/gh/drisspg/182/base 2025-12-04T08:57:03.5893630Z * [new branch] gh/drisspg/182/head -> origin/gh/drisspg/182/head 2025-12-04T08:57:03.5895665Z * [new branch] gh/drisspg/183/base -> origin/gh/drisspg/183/base 2025-12-04T08:57:03.5897138Z * [new branch] gh/drisspg/183/head -> origin/gh/drisspg/183/head 2025-12-04T08:57:03.5899147Z * [new branch] gh/drisspg/184/base -> origin/gh/drisspg/184/base 2025-12-04T08:57:03.5900692Z * [new branch] gh/drisspg/184/head -> origin/gh/drisspg/184/head 2025-12-04T08:57:03.5902823Z * [new branch] gh/drisspg/185/base -> origin/gh/drisspg/185/base 2025-12-04T08:57:03.5904805Z * [new branch] gh/drisspg/185/head -> origin/gh/drisspg/185/head 2025-12-04T08:57:03.5909226Z * [new branch] gh/drisspg/194/base -> origin/gh/drisspg/194/base 2025-12-04T08:57:03.5910814Z * [new branch] gh/drisspg/194/head -> origin/gh/drisspg/194/head 2025-12-04T08:57:03.5912356Z * [new branch] gh/drisspg/194/orig -> origin/gh/drisspg/194/orig 2025-12-04T08:57:03.5914457Z * [new branch] gh/drisspg/200/base -> origin/gh/drisspg/200/base 2025-12-04T08:57:03.5915992Z * [new branch] gh/drisspg/200/head -> origin/gh/drisspg/200/head 2025-12-04T08:57:03.5917520Z * [new branch] gh/drisspg/200/orig -> origin/gh/drisspg/200/orig 2025-12-04T08:57:03.5919645Z * [new branch] gh/drisspg/218/base -> origin/gh/drisspg/218/base 2025-12-04T08:57:03.5921334Z * [new branch] gh/drisspg/218/head -> origin/gh/drisspg/218/head 2025-12-04T08:57:03.5922857Z * [new branch] gh/drisspg/218/orig -> origin/gh/drisspg/218/orig 2025-12-04T08:57:03.5924864Z * [new branch] gh/drisspg/219/base -> origin/gh/drisspg/219/base 2025-12-04T08:57:03.5926407Z * [new branch] gh/drisspg/219/head -> origin/gh/drisspg/219/head 2025-12-04T08:57:03.5928029Z * [new branch] gh/drisspg/219/orig -> origin/gh/drisspg/219/orig 2025-12-04T08:57:03.5930096Z * [new branch] gh/drisspg/220/base -> origin/gh/drisspg/220/base 2025-12-04T08:57:03.5931677Z * [new branch] gh/drisspg/220/head -> origin/gh/drisspg/220/head 2025-12-04T08:57:03.5933322Z * [new branch] gh/drisspg/220/orig -> origin/gh/drisspg/220/orig 2025-12-04T08:57:03.5935583Z * [new branch] gh/drisspg/221/base -> origin/gh/drisspg/221/base 2025-12-04T08:57:03.5937109Z * [new branch] gh/drisspg/221/head -> origin/gh/drisspg/221/head 2025-12-04T08:57:03.5938724Z * [new branch] gh/drisspg/221/orig -> origin/gh/drisspg/221/orig 2025-12-04T08:57:03.5940794Z * [new branch] gh/drisspg/222/base -> origin/gh/drisspg/222/base 2025-12-04T08:57:03.5942392Z * [new branch] gh/drisspg/222/head -> origin/gh/drisspg/222/head 2025-12-04T08:57:03.5944086Z * [new branch] gh/drisspg/222/orig -> origin/gh/drisspg/222/orig 2025-12-04T08:57:03.5946182Z * [new branch] gh/drisspg/223/base -> origin/gh/drisspg/223/base 2025-12-04T08:57:03.5947738Z * [new branch] gh/drisspg/223/head -> origin/gh/drisspg/223/head 2025-12-04T08:57:03.5949349Z * [new branch] gh/drisspg/223/orig -> origin/gh/drisspg/223/orig 2025-12-04T08:57:03.5951531Z * [new branch] gh/drisspg/224/base -> origin/gh/drisspg/224/base 2025-12-04T08:57:03.5953092Z * [new branch] gh/drisspg/224/head -> origin/gh/drisspg/224/head 2025-12-04T08:57:03.5954603Z * [new branch] gh/drisspg/224/orig -> origin/gh/drisspg/224/orig 2025-12-04T08:57:03.5956689Z * [new branch] gh/drisspg/225/base -> origin/gh/drisspg/225/base 2025-12-04T08:57:03.5958286Z * [new branch] gh/drisspg/225/head -> origin/gh/drisspg/225/head 2025-12-04T08:57:03.5959843Z * [new branch] gh/drisspg/225/orig -> origin/gh/drisspg/225/orig 2025-12-04T08:57:03.5961877Z * [new branch] gh/drisspg/226/base -> origin/gh/drisspg/226/base 2025-12-04T08:57:03.5963456Z * [new branch] gh/drisspg/226/head -> origin/gh/drisspg/226/head 2025-12-04T08:57:03.5965133Z * [new branch] gh/drisspg/226/orig -> origin/gh/drisspg/226/orig 2025-12-04T08:57:03.5967668Z * [new branch] gh/drisspg/227/base -> origin/gh/drisspg/227/base 2025-12-04T08:57:03.5969249Z * [new branch] gh/drisspg/227/head -> origin/gh/drisspg/227/head 2025-12-04T08:57:03.5970786Z * [new branch] gh/drisspg/227/orig -> origin/gh/drisspg/227/orig 2025-12-04T08:57:03.5972969Z * [new branch] gh/drisspg/228/base -> origin/gh/drisspg/228/base 2025-12-04T08:57:03.5974660Z * [new branch] gh/drisspg/228/head -> origin/gh/drisspg/228/head 2025-12-04T08:57:03.5976726Z * [new branch] gh/drisspg/228/orig -> origin/gh/drisspg/228/orig 2025-12-04T08:57:03.5978884Z * [new branch] gh/drisspg/229/base -> origin/gh/drisspg/229/base 2025-12-04T08:57:03.5980503Z * [new branch] gh/drisspg/229/head -> origin/gh/drisspg/229/head 2025-12-04T08:57:03.5982062Z * [new branch] gh/drisspg/229/orig -> origin/gh/drisspg/229/orig 2025-12-04T08:57:03.5984400Z * [new branch] gh/drisspg/230/base -> origin/gh/drisspg/230/base 2025-12-04T08:57:03.5985966Z * [new branch] gh/drisspg/230/head -> origin/gh/drisspg/230/head 2025-12-04T08:57:03.5987557Z * [new branch] gh/drisspg/230/orig -> origin/gh/drisspg/230/orig 2025-12-04T08:57:03.5990169Z * [new branch] gh/dsjohns2/1/base -> origin/gh/dsjohns2/1/base 2025-12-04T08:57:03.5991742Z * [new branch] gh/dsjohns2/1/head -> origin/gh/dsjohns2/1/head 2025-12-04T08:57:03.5994274Z * [new branch] gh/dzmitry-huba/1/base -> origin/gh/dzmitry-huba/1/base 2025-12-04T08:57:03.5995912Z * [new branch] gh/dzmitry-huba/1/head -> origin/gh/dzmitry-huba/1/head 2025-12-04T08:57:03.5998004Z * [new branch] gh/dzmitry-huba/12/base -> origin/gh/dzmitry-huba/12/base 2025-12-04T08:57:03.5999684Z * [new branch] gh/dzmitry-huba/12/head -> origin/gh/dzmitry-huba/12/head 2025-12-04T08:57:03.6001351Z * [new branch] gh/dzmitry-huba/12/orig -> origin/gh/dzmitry-huba/12/orig 2025-12-04T08:57:03.6003586Z * [new branch] gh/dzmitry-huba/13/base -> origin/gh/dzmitry-huba/13/base 2025-12-04T08:57:03.6005240Z * [new branch] gh/dzmitry-huba/13/head -> origin/gh/dzmitry-huba/13/head 2025-12-04T08:57:03.6007192Z * [new branch] gh/dzmitry-huba/13/orig -> origin/gh/dzmitry-huba/13/orig 2025-12-04T08:57:03.6009299Z * [new branch] gh/dzmitry-huba/14/base -> origin/gh/dzmitry-huba/14/base 2025-12-04T08:57:03.6010891Z * [new branch] gh/dzmitry-huba/14/head -> origin/gh/dzmitry-huba/14/head 2025-12-04T08:57:03.6012529Z * [new branch] gh/dzmitry-huba/14/orig -> origin/gh/dzmitry-huba/14/orig 2025-12-04T08:57:03.6014902Z * [new branch] gh/dzmitry-huba/15/base -> origin/gh/dzmitry-huba/15/base 2025-12-04T08:57:03.6016540Z * [new branch] gh/dzmitry-huba/15/head -> origin/gh/dzmitry-huba/15/head 2025-12-04T08:57:03.6017992Z * [new branch] gh/dzmitry-huba/15/orig -> origin/gh/dzmitry-huba/15/orig 2025-12-04T08:57:03.6020246Z * [new branch] gh/dzmitry-huba/16/base -> origin/gh/dzmitry-huba/16/base 2025-12-04T08:57:03.6021969Z * [new branch] gh/dzmitry-huba/16/head -> origin/gh/dzmitry-huba/16/head 2025-12-04T08:57:03.6023612Z * [new branch] gh/dzmitry-huba/16/orig -> origin/gh/dzmitry-huba/16/orig 2025-12-04T08:57:03.6025790Z * [new branch] gh/dzmitry-huba/17/base -> origin/gh/dzmitry-huba/17/base 2025-12-04T08:57:03.6027399Z * [new branch] gh/dzmitry-huba/17/head -> origin/gh/dzmitry-huba/17/head 2025-12-04T08:57:03.6029072Z * [new branch] gh/dzmitry-huba/17/orig -> origin/gh/dzmitry-huba/17/orig 2025-12-04T08:57:03.6031057Z * [new branch] gh/dzmitry-huba/2/base -> origin/gh/dzmitry-huba/2/base 2025-12-04T08:57:03.6032611Z * [new branch] gh/dzmitry-huba/2/head -> origin/gh/dzmitry-huba/2/head 2025-12-04T08:57:03.6034565Z * [new branch] gh/dzmitry-huba/3/base -> origin/gh/dzmitry-huba/3/base 2025-12-04T08:57:03.6036061Z * [new branch] gh/dzmitry-huba/3/head -> origin/gh/dzmitry-huba/3/head 2025-12-04T08:57:03.6038749Z * [new branch] gh/eellison/808/base -> origin/gh/eellison/808/base 2025-12-04T08:57:03.6040332Z * [new branch] gh/eellison/808/head -> origin/gh/eellison/808/head 2025-12-04T08:57:03.6041919Z * [new branch] gh/eellison/808/orig -> origin/gh/eellison/808/orig 2025-12-04T08:57:03.6044224Z * [new branch] gh/eellison/822/base -> origin/gh/eellison/822/base 2025-12-04T08:57:03.6045738Z * [new branch] gh/eellison/822/head -> origin/gh/eellison/822/head 2025-12-04T08:57:03.6047820Z * [new branch] gh/eellison/822/orig -> origin/gh/eellison/822/orig 2025-12-04T08:57:03.6049783Z * [new branch] gh/eellison/823/base -> origin/gh/eellison/823/base 2025-12-04T08:57:03.6051331Z * [new branch] gh/eellison/823/head -> origin/gh/eellison/823/head 2025-12-04T08:57:03.6052929Z * [new branch] gh/eellison/823/orig -> origin/gh/eellison/823/orig 2025-12-04T08:57:03.6055175Z * [new branch] gh/eellison/862/base -> origin/gh/eellison/862/base 2025-12-04T08:57:03.6056663Z * [new branch] gh/eellison/862/head -> origin/gh/eellison/862/head 2025-12-04T08:57:03.6058269Z * [new branch] gh/eellison/862/orig -> origin/gh/eellison/862/orig 2025-12-04T08:57:03.6060357Z * [new branch] gh/eellison/863/base -> origin/gh/eellison/863/base 2025-12-04T08:57:03.6061918Z * [new branch] gh/eellison/863/head -> origin/gh/eellison/863/head 2025-12-04T08:57:03.6063466Z * [new branch] gh/eellison/863/orig -> origin/gh/eellison/863/orig 2025-12-04T08:57:03.6065580Z * [new branch] gh/eellison/864/base -> origin/gh/eellison/864/base 2025-12-04T08:57:03.6067189Z * [new branch] gh/eellison/864/head -> origin/gh/eellison/864/head 2025-12-04T08:57:03.6068812Z * [new branch] gh/eellison/864/orig -> origin/gh/eellison/864/orig 2025-12-04T08:57:03.6070958Z * [new branch] gh/eellison/865/base -> origin/gh/eellison/865/base 2025-12-04T08:57:03.6072689Z * [new branch] gh/eellison/865/head -> origin/gh/eellison/865/head 2025-12-04T08:57:03.6074142Z * [new branch] gh/eellison/865/orig -> origin/gh/eellison/865/orig 2025-12-04T08:57:03.6076189Z * [new branch] gh/eellison/866/base -> origin/gh/eellison/866/base 2025-12-04T08:57:03.6077933Z * [new branch] gh/eellison/866/head -> origin/gh/eellison/866/head 2025-12-04T08:57:03.6079346Z * [new branch] gh/eellison/866/orig -> origin/gh/eellison/866/orig 2025-12-04T08:57:03.6081553Z * [new branch] gh/eellison/867/base -> origin/gh/eellison/867/base 2025-12-04T08:57:03.6083091Z * [new branch] gh/eellison/867/head -> origin/gh/eellison/867/head 2025-12-04T08:57:03.6084638Z * [new branch] gh/eellison/867/orig -> origin/gh/eellison/867/orig 2025-12-04T08:57:03.6086926Z * [new branch] gh/eellison/868/base -> origin/gh/eellison/868/base 2025-12-04T08:57:03.6088742Z * [new branch] gh/eellison/868/head -> origin/gh/eellison/868/head 2025-12-04T08:57:03.6090312Z * [new branch] gh/eellison/868/orig -> origin/gh/eellison/868/orig 2025-12-04T08:57:03.6093047Z * [new branch] gh/eellison/869/base -> origin/gh/eellison/869/base 2025-12-04T08:57:03.6094807Z * [new branch] gh/eellison/869/head -> origin/gh/eellison/869/head 2025-12-04T08:57:03.6096352Z * [new branch] gh/eellison/869/orig -> origin/gh/eellison/869/orig 2025-12-04T08:57:03.6098478Z * [new branch] gh/eellison/870/base -> origin/gh/eellison/870/base 2025-12-04T08:57:03.6100042Z * [new branch] gh/eellison/870/head -> origin/gh/eellison/870/head 2025-12-04T08:57:03.6101575Z * [new branch] gh/eellison/870/orig -> origin/gh/eellison/870/orig 2025-12-04T08:57:03.6103708Z * [new branch] gh/eellison/871/base -> origin/gh/eellison/871/base 2025-12-04T08:57:03.6105406Z * [new branch] gh/eellison/871/head -> origin/gh/eellison/871/head 2025-12-04T08:57:03.6107067Z * [new branch] gh/eellison/871/orig -> origin/gh/eellison/871/orig 2025-12-04T08:57:03.6109606Z * [new branch] gh/eellison/872/base -> origin/gh/eellison/872/base 2025-12-04T08:57:03.6110943Z * [new branch] gh/eellison/872/head -> origin/gh/eellison/872/head 2025-12-04T08:57:03.6112426Z * [new branch] gh/eellison/872/orig -> origin/gh/eellison/872/orig 2025-12-04T08:57:03.6115040Z * [new branch] gh/eellison/873/base -> origin/gh/eellison/873/base 2025-12-04T08:57:03.6116563Z * [new branch] gh/eellison/873/head -> origin/gh/eellison/873/head 2025-12-04T08:57:03.6118201Z * [new branch] gh/eellison/873/orig -> origin/gh/eellison/873/orig 2025-12-04T08:57:03.6120354Z * [new branch] gh/eellison/874/base -> origin/gh/eellison/874/base 2025-12-04T08:57:03.6121956Z * [new branch] gh/eellison/874/head -> origin/gh/eellison/874/head 2025-12-04T08:57:03.6123588Z * [new branch] gh/eellison/874/orig -> origin/gh/eellison/874/orig 2025-12-04T08:57:03.6126232Z * [new branch] gh/eellison/875/base -> origin/gh/eellison/875/base 2025-12-04T08:57:03.6127870Z * [new branch] gh/eellison/875/head -> origin/gh/eellison/875/head 2025-12-04T08:57:03.6129445Z * [new branch] gh/eellison/875/orig -> origin/gh/eellison/875/orig 2025-12-04T08:57:03.6131691Z * [new branch] gh/eellison/876/base -> origin/gh/eellison/876/base 2025-12-04T08:57:03.6133853Z * [new branch] gh/eellison/876/head -> origin/gh/eellison/876/head 2025-12-04T08:57:03.6135531Z * [new branch] gh/eellison/876/orig -> origin/gh/eellison/876/orig 2025-12-04T08:57:03.6137805Z * [new branch] gh/eellison/877/base -> origin/gh/eellison/877/base 2025-12-04T08:57:03.6139336Z * [new branch] gh/eellison/877/head -> origin/gh/eellison/877/head 2025-12-04T08:57:03.6140905Z * [new branch] gh/eellison/877/orig -> origin/gh/eellison/877/orig 2025-12-04T08:57:03.6143131Z * [new branch] gh/eellison/878/base -> origin/gh/eellison/878/base 2025-12-04T08:57:03.6144707Z * [new branch] gh/eellison/878/head -> origin/gh/eellison/878/head 2025-12-04T08:57:03.6146286Z * [new branch] gh/eellison/878/orig -> origin/gh/eellison/878/orig 2025-12-04T08:57:03.6148474Z * [new branch] gh/eellison/879/base -> origin/gh/eellison/879/base 2025-12-04T08:57:03.6150084Z * [new branch] gh/eellison/879/head -> origin/gh/eellison/879/head 2025-12-04T08:57:03.6151628Z * [new branch] gh/eellison/879/orig -> origin/gh/eellison/879/orig 2025-12-04T08:57:03.6153880Z * [new branch] gh/eellison/880/base -> origin/gh/eellison/880/base 2025-12-04T08:57:03.6155612Z * [new branch] gh/eellison/880/head -> origin/gh/eellison/880/head 2025-12-04T08:57:03.6157198Z * [new branch] gh/eellison/880/orig -> origin/gh/eellison/880/orig 2025-12-04T08:57:03.6159478Z * [new branch] gh/eellison/881/base -> origin/gh/eellison/881/base 2025-12-04T08:57:03.6161045Z * [new branch] gh/eellison/881/head -> origin/gh/eellison/881/head 2025-12-04T08:57:03.6162598Z * [new branch] gh/eellison/881/orig -> origin/gh/eellison/881/orig 2025-12-04T08:57:03.6164786Z * [new branch] gh/eellison/882/base -> origin/gh/eellison/882/base 2025-12-04T08:57:03.6166360Z * [new branch] gh/eellison/882/head -> origin/gh/eellison/882/head 2025-12-04T08:57:03.6168039Z * [new branch] gh/eellison/882/orig -> origin/gh/eellison/882/orig 2025-12-04T08:57:03.6170162Z * [new branch] gh/eellison/883/base -> origin/gh/eellison/883/base 2025-12-04T08:57:03.6171737Z * [new branch] gh/eellison/883/head -> origin/gh/eellison/883/head 2025-12-04T08:57:03.6173345Z * [new branch] gh/eellison/883/orig -> origin/gh/eellison/883/orig 2025-12-04T08:57:03.6175598Z * [new branch] gh/eellison/884/base -> origin/gh/eellison/884/base 2025-12-04T08:57:03.6177001Z * [new branch] gh/eellison/884/head -> origin/gh/eellison/884/head 2025-12-04T08:57:03.6178583Z * [new branch] gh/eellison/884/orig -> origin/gh/eellison/884/orig 2025-12-04T08:57:03.6181249Z * [new branch] gh/etaf/147/base -> origin/gh/etaf/147/base 2025-12-04T08:57:03.6182829Z * [new branch] gh/etaf/147/head -> origin/gh/etaf/147/head 2025-12-04T08:57:03.6185051Z * [new branch] gh/etaf/154/base -> origin/gh/etaf/154/base 2025-12-04T08:57:03.6186646Z * [new branch] gh/etaf/154/head -> origin/gh/etaf/154/head 2025-12-04T08:57:03.6188214Z * [new branch] gh/etaf/154/orig -> origin/gh/etaf/154/orig 2025-12-04T08:57:03.6190321Z * [new branch] gh/etaf/156/base -> origin/gh/etaf/156/base 2025-12-04T08:57:03.6191852Z * [new branch] gh/etaf/156/head -> origin/gh/etaf/156/head 2025-12-04T08:57:03.6193454Z * [new branch] gh/etaf/156/orig -> origin/gh/etaf/156/orig 2025-12-04T08:57:03.6195691Z * [new branch] gh/etaf/157/base -> origin/gh/etaf/157/base 2025-12-04T08:57:03.6197301Z * [new branch] gh/etaf/157/head -> origin/gh/etaf/157/head 2025-12-04T08:57:03.6198899Z * [new branch] gh/etaf/157/orig -> origin/gh/etaf/157/orig 2025-12-04T08:57:03.6201029Z * [new branch] gh/etaf/158/base -> origin/gh/etaf/158/base 2025-12-04T08:57:03.6202685Z * [new branch] gh/etaf/158/head -> origin/gh/etaf/158/head 2025-12-04T08:57:03.6204340Z * [new branch] gh/etaf/158/orig -> origin/gh/etaf/158/orig 2025-12-04T08:57:03.6206828Z * [new branch] gh/etaf/159/base -> origin/gh/etaf/159/base 2025-12-04T08:57:03.6208394Z * [new branch] gh/etaf/159/head -> origin/gh/etaf/159/head 2025-12-04T08:57:03.6209973Z * [new branch] gh/etaf/159/orig -> origin/gh/etaf/159/orig 2025-12-04T08:57:03.6212170Z * [new branch] gh/etaf/160/base -> origin/gh/etaf/160/base 2025-12-04T08:57:03.6214743Z * [new branch] gh/etaf/160/head -> origin/gh/etaf/160/head 2025-12-04T08:57:03.6216339Z * [new branch] gh/etaf/160/orig -> origin/gh/etaf/160/orig 2025-12-04T08:57:03.6218496Z * [new branch] gh/etaf/161/base -> origin/gh/etaf/161/base 2025-12-04T08:57:03.6220146Z * [new branch] gh/etaf/161/head -> origin/gh/etaf/161/head 2025-12-04T08:57:03.6221757Z * [new branch] gh/etaf/161/orig -> origin/gh/etaf/161/orig 2025-12-04T08:57:03.6224228Z * [new branch] gh/etaf/166/base -> origin/gh/etaf/166/base 2025-12-04T08:57:03.6226070Z * [new branch] gh/etaf/166/head -> origin/gh/etaf/166/head 2025-12-04T08:57:03.6227634Z * [new branch] gh/etaf/166/orig -> origin/gh/etaf/166/orig 2025-12-04T08:57:03.6229702Z * [new branch] gh/etaf/167/base -> origin/gh/etaf/167/base 2025-12-04T08:57:03.6231265Z * [new branch] gh/etaf/167/head -> origin/gh/etaf/167/head 2025-12-04T08:57:03.6232856Z * [new branch] gh/etaf/167/orig -> origin/gh/etaf/167/orig 2025-12-04T08:57:03.6235278Z * [new branch] gh/etaf/168/base -> origin/gh/etaf/168/base 2025-12-04T08:57:03.6236885Z * [new branch] gh/etaf/168/head -> origin/gh/etaf/168/head 2025-12-04T08:57:03.6238515Z * [new branch] gh/etaf/168/orig -> origin/gh/etaf/168/orig 2025-12-04T08:57:03.6240677Z * [new branch] gh/etaf/172/base -> origin/gh/etaf/172/base 2025-12-04T08:57:03.6242313Z * [new branch] gh/etaf/172/head -> origin/gh/etaf/172/head 2025-12-04T08:57:03.6244225Z * [new branch] gh/etaf/172/orig -> origin/gh/etaf/172/orig 2025-12-04T08:57:03.6246418Z * [new branch] gh/etaf/173/base -> origin/gh/etaf/173/base 2025-12-04T08:57:03.6248117Z * [new branch] gh/etaf/173/head -> origin/gh/etaf/173/head 2025-12-04T08:57:03.6249627Z * [new branch] gh/etaf/173/orig -> origin/gh/etaf/173/orig 2025-12-04T08:57:03.6251746Z * [new branch] gh/etaf/174/base -> origin/gh/etaf/174/base 2025-12-04T08:57:03.6253388Z * [new branch] gh/etaf/174/head -> origin/gh/etaf/174/head 2025-12-04T08:57:03.6255538Z * [new branch] gh/etaf/175/base -> origin/gh/etaf/175/base 2025-12-04T08:57:03.6257110Z * [new branch] gh/etaf/175/head -> origin/gh/etaf/175/head 2025-12-04T08:57:03.6258607Z * [new branch] gh/etaf/175/orig -> origin/gh/etaf/175/orig 2025-12-04T08:57:03.6261277Z * [new branch] gh/etaf/176/base -> origin/gh/etaf/176/base 2025-12-04T08:57:03.6262903Z * [new branch] gh/etaf/176/head -> origin/gh/etaf/176/head 2025-12-04T08:57:03.6264543Z * [new branch] gh/etaf/176/orig -> origin/gh/etaf/176/orig 2025-12-04T08:57:03.6267276Z * [new branch] gh/etaf/177/base -> origin/gh/etaf/177/base 2025-12-04T08:57:03.6268978Z * [new branch] gh/etaf/177/head -> origin/gh/etaf/177/head 2025-12-04T08:57:03.6270551Z * [new branch] gh/etaf/177/orig -> origin/gh/etaf/177/orig 2025-12-04T08:57:03.6272823Z * [new branch] gh/etaf/178/base -> origin/gh/etaf/178/base 2025-12-04T08:57:03.6274492Z * [new branch] gh/etaf/178/head -> origin/gh/etaf/178/head 2025-12-04T08:57:03.6276134Z * [new branch] gh/etaf/178/orig -> origin/gh/etaf/178/orig 2025-12-04T08:57:03.6278393Z * [new branch] gh/etaf/179/base -> origin/gh/etaf/179/base 2025-12-04T08:57:03.6280007Z * [new branch] gh/etaf/179/head -> origin/gh/etaf/179/head 2025-12-04T08:57:03.6281551Z * [new branch] gh/etaf/179/orig -> origin/gh/etaf/179/orig 2025-12-04T08:57:03.6283607Z * [new branch] gh/etaf/180/base -> origin/gh/etaf/180/base 2025-12-04T08:57:03.6285176Z * [new branch] gh/etaf/180/head -> origin/gh/etaf/180/head 2025-12-04T08:57:03.6286735Z * [new branch] gh/etaf/180/orig -> origin/gh/etaf/180/orig 2025-12-04T08:57:03.6289377Z * [new branch] gh/exclamaforte/1/base -> origin/gh/exclamaforte/1/base 2025-12-04T08:57:03.6290979Z * [new branch] gh/exclamaforte/1/head -> origin/gh/exclamaforte/1/head 2025-12-04T08:57:03.6293163Z * [new branch] gh/exclamaforte/2/base -> origin/gh/exclamaforte/2/base 2025-12-04T08:57:03.6294732Z * [new branch] gh/exclamaforte/2/head -> origin/gh/exclamaforte/2/head 2025-12-04T08:57:03.6296839Z * [new branch] gh/exclamaforte/3/base -> origin/gh/exclamaforte/3/base 2025-12-04T08:57:03.6298427Z * [new branch] gh/exclamaforte/3/head -> origin/gh/exclamaforte/3/head 2025-12-04T08:57:03.6300527Z * [new branch] gh/exclamaforte/4/base -> origin/gh/exclamaforte/4/base 2025-12-04T08:57:03.6302065Z * [new branch] gh/exclamaforte/4/head -> origin/gh/exclamaforte/4/head 2025-12-04T08:57:03.6304906Z * [new branch] gh/ezyang/2374/base -> origin/gh/ezyang/2374/base 2025-12-04T08:57:03.6308902Z * [new branch] gh/ezyang/2374/head -> origin/gh/ezyang/2374/head 2025-12-04T08:57:03.6310510Z * [new branch] gh/ezyang/2374/orig -> origin/gh/ezyang/2374/orig 2025-12-04T08:57:03.6312745Z * [new branch] gh/ezyang/2973/base -> origin/gh/ezyang/2973/base 2025-12-04T08:57:03.6314195Z * [new branch] gh/ezyang/2973/head -> origin/gh/ezyang/2973/head 2025-12-04T08:57:03.6315778Z * [new branch] gh/ezyang/2973/orig -> origin/gh/ezyang/2973/orig 2025-12-04T08:57:03.6317869Z * [new branch] gh/ezyang/2974/base -> origin/gh/ezyang/2974/base 2025-12-04T08:57:03.6319461Z * [new branch] gh/ezyang/2974/head -> origin/gh/ezyang/2974/head 2025-12-04T08:57:03.6321009Z * [new branch] gh/ezyang/2974/orig -> origin/gh/ezyang/2974/orig 2025-12-04T08:57:03.6323089Z * [new branch] gh/ezyang/3131/base -> origin/gh/ezyang/3131/base 2025-12-04T08:57:03.6324632Z * [new branch] gh/ezyang/3131/head -> origin/gh/ezyang/3131/head 2025-12-04T08:57:03.6326207Z * [new branch] gh/ezyang/3131/orig -> origin/gh/ezyang/3131/orig 2025-12-04T08:57:03.6328305Z * [new branch] gh/ezyang/3139/base -> origin/gh/ezyang/3139/base 2025-12-04T08:57:03.6329858Z * [new branch] gh/ezyang/3139/head -> origin/gh/ezyang/3139/head 2025-12-04T08:57:03.6331489Z * [new branch] gh/ezyang/3139/orig -> origin/gh/ezyang/3139/orig 2025-12-04T08:57:03.6333651Z * [new branch] gh/ezyang/3140/base -> origin/gh/ezyang/3140/base 2025-12-04T08:57:03.6335173Z * [new branch] gh/ezyang/3140/head -> origin/gh/ezyang/3140/head 2025-12-04T08:57:03.6336736Z * [new branch] gh/ezyang/3140/orig -> origin/gh/ezyang/3140/orig 2025-12-04T08:57:03.6339371Z * [new branch] gh/ezyang/3143/base -> origin/gh/ezyang/3143/base 2025-12-04T08:57:03.6340903Z * [new branch] gh/ezyang/3143/head -> origin/gh/ezyang/3143/head 2025-12-04T08:57:03.6342441Z * [new branch] gh/ezyang/3143/orig -> origin/gh/ezyang/3143/orig 2025-12-04T08:57:03.6344608Z * [new branch] gh/ezyang/3144/base -> origin/gh/ezyang/3144/base 2025-12-04T08:57:03.6346194Z * [new branch] gh/ezyang/3144/head -> origin/gh/ezyang/3144/head 2025-12-04T08:57:03.6347810Z * [new branch] gh/ezyang/3144/orig -> origin/gh/ezyang/3144/orig 2025-12-04T08:57:03.6349907Z * [new branch] gh/ezyang/3167/base -> origin/gh/ezyang/3167/base 2025-12-04T08:57:03.6351522Z * [new branch] gh/ezyang/3167/head -> origin/gh/ezyang/3167/head 2025-12-04T08:57:03.6353146Z * [new branch] gh/ezyang/3167/orig -> origin/gh/ezyang/3167/orig 2025-12-04T08:57:03.6355180Z * [new branch] gh/ezyang/3173/base -> origin/gh/ezyang/3173/base 2025-12-04T08:57:03.6356799Z * [new branch] gh/ezyang/3173/head -> origin/gh/ezyang/3173/head 2025-12-04T08:57:03.6358419Z * [new branch] gh/ezyang/3173/orig -> origin/gh/ezyang/3173/orig 2025-12-04T08:57:03.6360482Z * [new branch] gh/ezyang/3175/base -> origin/gh/ezyang/3175/base 2025-12-04T08:57:03.6362041Z * [new branch] gh/ezyang/3175/head -> origin/gh/ezyang/3175/head 2025-12-04T08:57:03.6363676Z * [new branch] gh/ezyang/3175/orig -> origin/gh/ezyang/3175/orig 2025-12-04T08:57:03.6365756Z * [new branch] gh/ezyang/3182/base -> origin/gh/ezyang/3182/base 2025-12-04T08:57:03.6367286Z * [new branch] gh/ezyang/3182/head -> origin/gh/ezyang/3182/head 2025-12-04T08:57:03.6368860Z * [new branch] gh/ezyang/3182/orig -> origin/gh/ezyang/3182/orig 2025-12-04T08:57:03.6370934Z * [new branch] gh/ezyang/3185/base -> origin/gh/ezyang/3185/base 2025-12-04T08:57:03.6372549Z * [new branch] gh/ezyang/3185/head -> origin/gh/ezyang/3185/head 2025-12-04T08:57:03.6374291Z * [new branch] gh/ezyang/3185/orig -> origin/gh/ezyang/3185/orig 2025-12-04T08:57:03.6376523Z * [new branch] gh/ezyang/3189/base -> origin/gh/ezyang/3189/base 2025-12-04T08:57:03.6377934Z * [new branch] gh/ezyang/3189/head -> origin/gh/ezyang/3189/head 2025-12-04T08:57:03.6379491Z * [new branch] gh/ezyang/3189/orig -> origin/gh/ezyang/3189/orig 2025-12-04T08:57:03.6381556Z * [new branch] gh/ezyang/3191/base -> origin/gh/ezyang/3191/base 2025-12-04T08:57:03.6383147Z * [new branch] gh/ezyang/3191/head -> origin/gh/ezyang/3191/head 2025-12-04T08:57:03.6385022Z * [new branch] gh/ezyang/3191/orig -> origin/gh/ezyang/3191/orig 2025-12-04T08:57:03.6387638Z * [new branch] gh/ezyang/3192/base -> origin/gh/ezyang/3192/base 2025-12-04T08:57:03.6389213Z * [new branch] gh/ezyang/3192/head -> origin/gh/ezyang/3192/head 2025-12-04T08:57:03.6390817Z * [new branch] gh/ezyang/3192/orig -> origin/gh/ezyang/3192/orig 2025-12-04T08:57:03.6392949Z * [new branch] gh/ezyang/3193/base -> origin/gh/ezyang/3193/base 2025-12-04T08:57:03.6394601Z * [new branch] gh/ezyang/3193/head -> origin/gh/ezyang/3193/head 2025-12-04T08:57:03.6396342Z * [new branch] gh/ezyang/3193/orig -> origin/gh/ezyang/3193/orig 2025-12-04T08:57:03.6398730Z * [new branch] gh/ezyang/3194/base -> origin/gh/ezyang/3194/base 2025-12-04T08:57:03.6400118Z * [new branch] gh/ezyang/3194/head -> origin/gh/ezyang/3194/head 2025-12-04T08:57:03.6401618Z * [new branch] gh/ezyang/3194/orig -> origin/gh/ezyang/3194/orig 2025-12-04T08:57:03.6403751Z * [new branch] gh/ezyang/3195/base -> origin/gh/ezyang/3195/base 2025-12-04T08:57:03.6405524Z * [new branch] gh/ezyang/3195/head -> origin/gh/ezyang/3195/head 2025-12-04T08:57:03.6407228Z * [new branch] gh/ezyang/3195/orig -> origin/gh/ezyang/3195/orig 2025-12-04T08:57:03.6409381Z * [new branch] gh/ezyang/3196/base -> origin/gh/ezyang/3196/base 2025-12-04T08:57:03.6411031Z * [new branch] gh/ezyang/3196/head -> origin/gh/ezyang/3196/head 2025-12-04T08:57:03.6412606Z * [new branch] gh/ezyang/3196/orig -> origin/gh/ezyang/3196/orig 2025-12-04T08:57:03.6414985Z * [new branch] gh/ezyang/3197/base -> origin/gh/ezyang/3197/base 2025-12-04T08:57:03.6416485Z * [new branch] gh/ezyang/3197/head -> origin/gh/ezyang/3197/head 2025-12-04T08:57:03.6418161Z * [new branch] gh/ezyang/3197/orig -> origin/gh/ezyang/3197/orig 2025-12-04T08:57:03.6420317Z * [new branch] gh/ezyang/3198/base -> origin/gh/ezyang/3198/base 2025-12-04T08:57:03.6421809Z * [new branch] gh/ezyang/3198/head -> origin/gh/ezyang/3198/head 2025-12-04T08:57:03.6423476Z * [new branch] gh/ezyang/3198/orig -> origin/gh/ezyang/3198/orig 2025-12-04T08:57:03.6425771Z * [new branch] gh/ezyang/3199/base -> origin/gh/ezyang/3199/base 2025-12-04T08:57:03.6427326Z * [new branch] gh/ezyang/3199/head -> origin/gh/ezyang/3199/head 2025-12-04T08:57:03.6428900Z * [new branch] gh/ezyang/3199/orig -> origin/gh/ezyang/3199/orig 2025-12-04T08:57:03.6431052Z * [new branch] gh/ezyang/3200/base -> origin/gh/ezyang/3200/base 2025-12-04T08:57:03.6432665Z * [new branch] gh/ezyang/3200/head -> origin/gh/ezyang/3200/head 2025-12-04T08:57:03.6434211Z * [new branch] gh/ezyang/3200/orig -> origin/gh/ezyang/3200/orig 2025-12-04T08:57:03.6436853Z * [new branch] gh/ezyang/3201/base -> origin/gh/ezyang/3201/base 2025-12-04T08:57:03.6438587Z * [new branch] gh/ezyang/3201/head -> origin/gh/ezyang/3201/head 2025-12-04T08:57:03.6440333Z * [new branch] gh/ezyang/3201/orig -> origin/gh/ezyang/3201/orig 2025-12-04T08:57:03.6442363Z * [new branch] gh/ezyang/3202/base -> origin/gh/ezyang/3202/base 2025-12-04T08:57:03.6443772Z * [new branch] gh/ezyang/3202/head -> origin/gh/ezyang/3202/head 2025-12-04T08:57:03.6445263Z * [new branch] gh/ezyang/3202/orig -> origin/gh/ezyang/3202/orig 2025-12-04T08:57:03.6447431Z * [new branch] gh/ezyang/3203/base -> origin/gh/ezyang/3203/base 2025-12-04T08:57:03.6448970Z * [new branch] gh/ezyang/3203/head -> origin/gh/ezyang/3203/head 2025-12-04T08:57:03.6450684Z * [new branch] gh/ezyang/3203/orig -> origin/gh/ezyang/3203/orig 2025-12-04T08:57:03.6452820Z * [new branch] gh/ezyang/3204/base -> origin/gh/ezyang/3204/base 2025-12-04T08:57:03.6454545Z * [new branch] gh/ezyang/3204/head -> origin/gh/ezyang/3204/head 2025-12-04T08:57:03.6456087Z * [new branch] gh/ezyang/3204/orig -> origin/gh/ezyang/3204/orig 2025-12-04T08:57:03.6458683Z * [new branch] gh/ezyang/3205/base -> origin/gh/ezyang/3205/base 2025-12-04T08:57:03.6460224Z * [new branch] gh/ezyang/3205/head -> origin/gh/ezyang/3205/head 2025-12-04T08:57:03.6461805Z * [new branch] gh/ezyang/3205/orig -> origin/gh/ezyang/3205/orig 2025-12-04T08:57:03.6463976Z * [new branch] gh/ezyang/3206/base -> origin/gh/ezyang/3206/base 2025-12-04T08:57:03.6465649Z * [new branch] gh/ezyang/3206/head -> origin/gh/ezyang/3206/head 2025-12-04T08:57:03.6467248Z * [new branch] gh/ezyang/3206/orig -> origin/gh/ezyang/3206/orig 2025-12-04T08:57:03.6469352Z * [new branch] gh/ezyang/3207/base -> origin/gh/ezyang/3207/base 2025-12-04T08:57:03.6470991Z * [new branch] gh/ezyang/3207/head -> origin/gh/ezyang/3207/head 2025-12-04T08:57:03.6472570Z * [new branch] gh/ezyang/3207/orig -> origin/gh/ezyang/3207/orig 2025-12-04T08:57:03.6474702Z * [new branch] gh/ezyang/3208/base -> origin/gh/ezyang/3208/base 2025-12-04T08:57:03.6476717Z * [new branch] gh/ezyang/3208/head -> origin/gh/ezyang/3208/head 2025-12-04T08:57:03.6478276Z * [new branch] gh/ezyang/3208/orig -> origin/gh/ezyang/3208/orig 2025-12-04T08:57:03.6480415Z * [new branch] gh/ezyang/3209/base -> origin/gh/ezyang/3209/base 2025-12-04T08:57:03.6481965Z * [new branch] gh/ezyang/3209/head -> origin/gh/ezyang/3209/head 2025-12-04T08:57:03.6483662Z * [new branch] gh/ezyang/3209/orig -> origin/gh/ezyang/3209/orig 2025-12-04T08:57:03.6486270Z * [new branch] gh/fadara01/3/base -> origin/gh/fadara01/3/base 2025-12-04T08:57:03.6487797Z * [new branch] gh/fadara01/3/head -> origin/gh/fadara01/3/head 2025-12-04T08:57:03.6489409Z * [new branch] gh/fadara01/3/orig -> origin/gh/fadara01/3/orig 2025-12-04T08:57:03.6491460Z * [new branch] gh/fadara01/5/base -> origin/gh/fadara01/5/base 2025-12-04T08:57:03.6493002Z * [new branch] gh/fadara01/5/head -> origin/gh/fadara01/5/head 2025-12-04T08:57:03.6494636Z * [new branch] gh/fadara01/5/orig -> origin/gh/fadara01/5/orig 2025-12-04T08:57:03.6497195Z * [new branch] gh/fadara01/6/base -> origin/gh/fadara01/6/base 2025-12-04T08:57:03.6498733Z * [new branch] gh/fadara01/6/head -> origin/gh/fadara01/6/head 2025-12-04T08:57:03.6500283Z * [new branch] gh/fadara01/6/orig -> origin/gh/fadara01/6/orig 2025-12-04T08:57:03.6502378Z * [new branch] gh/fadara01/7/base -> origin/gh/fadara01/7/base 2025-12-04T08:57:03.6504143Z * [new branch] gh/fadara01/7/head -> origin/gh/fadara01/7/head 2025-12-04T08:57:03.6506046Z * [new branch] gh/fadara01/7/orig -> origin/gh/fadara01/7/orig 2025-12-04T08:57:03.6508187Z * [new branch] gh/fadara01/8/base -> origin/gh/fadara01/8/base 2025-12-04T08:57:03.6509743Z * [new branch] gh/fadara01/8/head -> origin/gh/fadara01/8/head 2025-12-04T08:57:03.6511288Z * [new branch] gh/fadara01/8/orig -> origin/gh/fadara01/8/orig 2025-12-04T08:57:03.6513392Z * [new branch] gh/fadara01/9/base -> origin/gh/fadara01/9/base 2025-12-04T08:57:03.6515045Z * [new branch] gh/fadara01/9/head -> origin/gh/fadara01/9/head 2025-12-04T08:57:03.6516652Z * [new branch] gh/fadara01/9/orig -> origin/gh/fadara01/9/orig 2025-12-04T08:57:03.6519162Z * [new branch] gh/fduwjj/182/base -> origin/gh/fduwjj/182/base 2025-12-04T08:57:03.6520780Z * [new branch] gh/fduwjj/182/head -> origin/gh/fduwjj/182/head 2025-12-04T08:57:03.6522281Z * [new branch] gh/fduwjj/182/orig -> origin/gh/fduwjj/182/orig 2025-12-04T08:57:03.6524523Z * [new branch] gh/fduwjj/211/base -> origin/gh/fduwjj/211/base 2025-12-04T08:57:03.6526129Z * [new branch] gh/fduwjj/211/head -> origin/gh/fduwjj/211/head 2025-12-04T08:57:03.6527733Z * [new branch] gh/fduwjj/211/orig -> origin/gh/fduwjj/211/orig 2025-12-04T08:57:03.6529809Z * [new branch] gh/fduwjj/212/base -> origin/gh/fduwjj/212/base 2025-12-04T08:57:03.6531373Z * [new branch] gh/fduwjj/212/head -> origin/gh/fduwjj/212/head 2025-12-04T08:57:03.6532959Z * [new branch] gh/fduwjj/212/orig -> origin/gh/fduwjj/212/orig 2025-12-04T08:57:03.6535105Z * [new branch] gh/fduwjj/213/base -> origin/gh/fduwjj/213/base 2025-12-04T08:57:03.6536630Z * [new branch] gh/fduwjj/213/head -> origin/gh/fduwjj/213/head 2025-12-04T08:57:03.6538159Z * [new branch] gh/fduwjj/213/orig -> origin/gh/fduwjj/213/orig 2025-12-04T08:57:03.6540388Z * [new branch] gh/fduwjj/226/base -> origin/gh/fduwjj/226/base 2025-12-04T08:57:03.6541897Z * [new branch] gh/fduwjj/226/head -> origin/gh/fduwjj/226/head 2025-12-04T08:57:03.6543442Z * [new branch] gh/fduwjj/226/orig -> origin/gh/fduwjj/226/orig 2025-12-04T08:57:03.6545853Z * [new branch] gh/fduwjj/229/base -> origin/gh/fduwjj/229/base 2025-12-04T08:57:03.6547348Z * [new branch] gh/fduwjj/229/head -> origin/gh/fduwjj/229/head 2025-12-04T08:57:03.6548876Z * [new branch] gh/fduwjj/229/orig -> origin/gh/fduwjj/229/orig 2025-12-04T08:57:03.6550981Z * [new branch] gh/fduwjj/233/base -> origin/gh/fduwjj/233/base 2025-12-04T08:57:03.6552608Z * [new branch] gh/fduwjj/233/head -> origin/gh/fduwjj/233/head 2025-12-04T08:57:03.6554192Z * [new branch] gh/fduwjj/233/orig -> origin/gh/fduwjj/233/orig 2025-12-04T08:57:03.6556352Z * [new branch] gh/fduwjj/234/base -> origin/gh/fduwjj/234/base 2025-12-04T08:57:03.6558009Z * [new branch] gh/fduwjj/234/head -> origin/gh/fduwjj/234/head 2025-12-04T08:57:03.6559523Z * [new branch] gh/fduwjj/234/orig -> origin/gh/fduwjj/234/orig 2025-12-04T08:57:03.6561622Z * [new branch] gh/fduwjj/235/base -> origin/gh/fduwjj/235/base 2025-12-04T08:57:03.6563236Z * [new branch] gh/fduwjj/235/head -> origin/gh/fduwjj/235/head 2025-12-04T08:57:03.6564743Z * [new branch] gh/fduwjj/235/orig -> origin/gh/fduwjj/235/orig 2025-12-04T08:57:03.6566774Z * [new branch] gh/fduwjj/236/base -> origin/gh/fduwjj/236/base 2025-12-04T08:57:03.6568521Z * [new branch] gh/fduwjj/236/head -> origin/gh/fduwjj/236/head 2025-12-04T08:57:03.6570060Z * [new branch] gh/fduwjj/236/orig -> origin/gh/fduwjj/236/orig 2025-12-04T08:57:03.6571925Z * [new branch] gh/fduwjj/237/base -> origin/gh/fduwjj/237/base 2025-12-04T08:57:03.6573597Z * [new branch] gh/fduwjj/237/head -> origin/gh/fduwjj/237/head 2025-12-04T08:57:03.6575152Z * [new branch] gh/fduwjj/237/orig -> origin/gh/fduwjj/237/orig 2025-12-04T08:57:03.6577231Z * [new branch] gh/fduwjj/238/base -> origin/gh/fduwjj/238/base 2025-12-04T08:57:03.6578906Z * [new branch] gh/fduwjj/238/head -> origin/gh/fduwjj/238/head 2025-12-04T08:57:03.6580513Z * [new branch] gh/fduwjj/238/orig -> origin/gh/fduwjj/238/orig 2025-12-04T08:57:03.6582657Z * [new branch] gh/fduwjj/239/base -> origin/gh/fduwjj/239/base 2025-12-04T08:57:03.6584363Z * [new branch] gh/fduwjj/239/head -> origin/gh/fduwjj/239/head 2025-12-04T08:57:03.6586094Z * [new branch] gh/fduwjj/239/orig -> origin/gh/fduwjj/239/orig 2025-12-04T08:57:03.6588619Z * [new branch] gh/fegin/332/base -> origin/gh/fegin/332/base 2025-12-04T08:57:03.6590208Z * [new branch] gh/fegin/332/head -> origin/gh/fegin/332/head 2025-12-04T08:57:03.6591745Z * [new branch] gh/fegin/332/orig -> origin/gh/fegin/332/orig 2025-12-04T08:57:03.6593948Z * [new branch] gh/fegin/333/base -> origin/gh/fegin/333/base 2025-12-04T08:57:03.6595571Z * [new branch] gh/fegin/333/head -> origin/gh/fegin/333/head 2025-12-04T08:57:03.6597055Z * [new branch] gh/fegin/333/orig -> origin/gh/fegin/333/orig 2025-12-04T08:57:03.6599157Z * [new branch] gh/fegin/334/base -> origin/gh/fegin/334/base 2025-12-04T08:57:03.6600759Z * [new branch] gh/fegin/334/head -> origin/gh/fegin/334/head 2025-12-04T08:57:03.6602382Z * [new branch] gh/fegin/334/orig -> origin/gh/fegin/334/orig 2025-12-04T08:57:03.6604417Z * [new branch] gh/fegin/335/base -> origin/gh/fegin/335/base 2025-12-04T08:57:03.6606398Z * [new branch] gh/fegin/335/head -> origin/gh/fegin/335/head 2025-12-04T08:57:03.6607982Z * [new branch] gh/fegin/335/orig -> origin/gh/fegin/335/orig 2025-12-04T08:57:03.6610531Z * [new branch] gh/fffrog/160/base -> origin/gh/fffrog/160/base 2025-12-04T08:57:03.6612266Z * [new branch] gh/fffrog/160/head -> origin/gh/fffrog/160/head 2025-12-04T08:57:03.6614585Z * [new branch] gh/fffrog/177/base -> origin/gh/fffrog/177/base 2025-12-04T08:57:03.6616160Z * [new branch] gh/fffrog/177/head -> origin/gh/fffrog/177/head 2025-12-04T08:57:03.6617717Z * [new branch] gh/fffrog/177/orig -> origin/gh/fffrog/177/orig 2025-12-04T08:57:03.6619801Z * [new branch] gh/fffrog/178/base -> origin/gh/fffrog/178/base 2025-12-04T08:57:03.6621421Z * [new branch] gh/fffrog/178/head -> origin/gh/fffrog/178/head 2025-12-04T08:57:03.6622972Z * [new branch] gh/fffrog/178/orig -> origin/gh/fffrog/178/orig 2025-12-04T08:57:03.6624903Z * [new branch] gh/fffrog/181/base -> origin/gh/fffrog/181/base 2025-12-04T08:57:03.6626429Z * [new branch] gh/fffrog/181/head -> origin/gh/fffrog/181/head 2025-12-04T08:57:03.6628011Z * [new branch] gh/fffrog/181/orig -> origin/gh/fffrog/181/orig 2025-12-04T08:57:03.6630147Z * [new branch] gh/fffrog/183/base -> origin/gh/fffrog/183/base 2025-12-04T08:57:03.6631708Z * [new branch] gh/fffrog/183/head -> origin/gh/fffrog/183/head 2025-12-04T08:57:03.6633377Z * [new branch] gh/fffrog/183/orig -> origin/gh/fffrog/183/orig 2025-12-04T08:57:03.6635919Z * [new branch] gh/fxdawnn/10/base -> origin/gh/fxdawnn/10/base 2025-12-04T08:57:03.6637406Z * [new branch] gh/fxdawnn/10/head -> origin/gh/fxdawnn/10/head 2025-12-04T08:57:03.6638978Z * [new branch] gh/fxdawnn/10/orig -> origin/gh/fxdawnn/10/orig 2025-12-04T08:57:03.6641335Z * [new branch] gh/fxdawnn/11/base -> origin/gh/fxdawnn/11/base 2025-12-04T08:57:03.6642760Z * [new branch] gh/fxdawnn/11/head -> origin/gh/fxdawnn/11/head 2025-12-04T08:57:03.6644401Z * [new branch] gh/fxdawnn/11/orig -> origin/gh/fxdawnn/11/orig 2025-12-04T08:57:03.6646490Z * [new branch] gh/fxdawnn/12/base -> origin/gh/fxdawnn/12/base 2025-12-04T08:57:03.6648266Z * [new branch] gh/fxdawnn/12/head -> origin/gh/fxdawnn/12/head 2025-12-04T08:57:03.6649732Z * [new branch] gh/fxdawnn/12/orig -> origin/gh/fxdawnn/12/orig 2025-12-04T08:57:03.6651803Z * [new branch] gh/fxdawnn/13/base -> origin/gh/fxdawnn/13/base 2025-12-04T08:57:03.6653474Z * [new branch] gh/fxdawnn/13/head -> origin/gh/fxdawnn/13/head 2025-12-04T08:57:03.6655275Z * [new branch] gh/fxdawnn/13/orig -> origin/gh/fxdawnn/13/orig 2025-12-04T08:57:03.6657445Z * [new branch] gh/fxdawnn/14/base -> origin/gh/fxdawnn/14/base 2025-12-04T08:57:03.6658939Z * [new branch] gh/fxdawnn/14/head -> origin/gh/fxdawnn/14/head 2025-12-04T08:57:03.6660456Z * [new branch] gh/fxdawnn/14/orig -> origin/gh/fxdawnn/14/orig 2025-12-04T08:57:03.6662521Z * [new branch] gh/fxdawnn/15/base -> origin/gh/fxdawnn/15/base 2025-12-04T08:57:03.6664094Z * [new branch] gh/fxdawnn/15/head -> origin/gh/fxdawnn/15/head 2025-12-04T08:57:03.6665609Z * [new branch] gh/fxdawnn/15/orig -> origin/gh/fxdawnn/15/orig 2025-12-04T08:57:03.6667699Z * [new branch] gh/fxdawnn/6/base -> origin/gh/fxdawnn/6/base 2025-12-04T08:57:03.6669258Z * [new branch] gh/fxdawnn/6/head -> origin/gh/fxdawnn/6/head 2025-12-04T08:57:03.6670905Z * [new branch] gh/fxdawnn/6/orig -> origin/gh/fxdawnn/6/orig 2025-12-04T08:57:03.6673038Z * [new branch] gh/fxdawnn/7/base -> origin/gh/fxdawnn/7/base 2025-12-04T08:57:03.6674758Z * [new branch] gh/fxdawnn/7/head -> origin/gh/fxdawnn/7/head 2025-12-04T08:57:03.6676239Z * [new branch] gh/fxdawnn/7/orig -> origin/gh/fxdawnn/7/orig 2025-12-04T08:57:03.6678344Z * [new branch] gh/fxdawnn/9/base -> origin/gh/fxdawnn/9/base 2025-12-04T08:57:03.6679905Z * [new branch] gh/fxdawnn/9/head -> origin/gh/fxdawnn/9/head 2025-12-04T08:57:03.6681478Z * [new branch] gh/fxdawnn/9/orig -> origin/gh/fxdawnn/9/orig 2025-12-04T08:57:03.6684077Z * [new branch] gh/galv/1/base -> origin/gh/galv/1/base 2025-12-04T08:57:03.6685974Z * [new branch] gh/galv/1/head -> origin/gh/galv/1/head 2025-12-04T08:57:03.6687154Z * [new branch] gh/galv/1/orig -> origin/gh/galv/1/orig 2025-12-04T08:57:03.6689366Z * [new branch] gh/galv/2/base -> origin/gh/galv/2/base 2025-12-04T08:57:03.6691097Z * [new branch] gh/galv/2/head -> origin/gh/galv/2/head 2025-12-04T08:57:03.6692722Z * [new branch] gh/galv/2/orig -> origin/gh/galv/2/orig 2025-12-04T08:57:03.6694929Z * [new branch] gh/galv/3/base -> origin/gh/galv/3/base 2025-12-04T08:57:03.6696443Z * [new branch] gh/galv/3/head -> origin/gh/galv/3/head 2025-12-04T08:57:03.6698240Z * [new branch] gh/galv/3/orig -> origin/gh/galv/3/orig 2025-12-04T08:57:03.6700713Z * [new branch] gh/guangyey/134/base -> origin/gh/guangyey/134/base 2025-12-04T08:57:03.6702288Z * [new branch] gh/guangyey/134/head -> origin/gh/guangyey/134/head 2025-12-04T08:57:03.6703888Z * [new branch] gh/guangyey/134/orig -> origin/gh/guangyey/134/orig 2025-12-04T08:57:03.6707100Z * [new branch] gh/guangyey/163/base -> origin/gh/guangyey/163/base 2025-12-04T08:57:03.6708654Z * [new branch] gh/guangyey/163/head -> origin/gh/guangyey/163/head 2025-12-04T08:57:03.6710229Z * [new branch] gh/guangyey/163/orig -> origin/gh/guangyey/163/orig 2025-12-04T08:57:03.6712401Z * [new branch] gh/guangyey/168/base -> origin/gh/guangyey/168/base 2025-12-04T08:57:03.6713953Z * [new branch] gh/guangyey/168/head -> origin/gh/guangyey/168/head 2025-12-04T08:57:03.6715505Z * [new branch] gh/guangyey/168/orig -> origin/gh/guangyey/168/orig 2025-12-04T08:57:03.6717546Z * [new branch] gh/guangyey/169/base -> origin/gh/guangyey/169/base 2025-12-04T08:57:03.6719160Z * [new branch] gh/guangyey/169/head -> origin/gh/guangyey/169/head 2025-12-04T08:57:03.6720714Z * [new branch] gh/guangyey/169/orig -> origin/gh/guangyey/169/orig 2025-12-04T08:57:03.6722795Z * [new branch] gh/guangyey/170/base -> origin/gh/guangyey/170/base 2025-12-04T08:57:03.6724401Z * [new branch] gh/guangyey/170/head -> origin/gh/guangyey/170/head 2025-12-04T08:57:03.6726013Z * [new branch] gh/guangyey/170/orig -> origin/gh/guangyey/170/orig 2025-12-04T08:57:03.6728028Z * [new branch] gh/guangyey/171/base -> origin/gh/guangyey/171/base 2025-12-04T08:57:03.6729570Z * [new branch] gh/guangyey/171/head -> origin/gh/guangyey/171/head 2025-12-04T08:57:03.6731144Z * [new branch] gh/guangyey/171/orig -> origin/gh/guangyey/171/orig 2025-12-04T08:57:03.6733388Z * [new branch] gh/guangyey/178/base -> origin/gh/guangyey/178/base 2025-12-04T08:57:03.6735028Z * [new branch] gh/guangyey/178/head -> origin/gh/guangyey/178/head 2025-12-04T08:57:03.6736688Z * [new branch] gh/guangyey/178/orig -> origin/gh/guangyey/178/orig 2025-12-04T08:57:03.6738959Z * [new branch] gh/guangyey/182/base -> origin/gh/guangyey/182/base 2025-12-04T08:57:03.6740513Z * [new branch] gh/guangyey/182/head -> origin/gh/guangyey/182/head 2025-12-04T08:57:03.6742085Z * [new branch] gh/guangyey/182/orig -> origin/gh/guangyey/182/orig 2025-12-04T08:57:03.6744084Z * [new branch] gh/guangyey/183/base -> origin/gh/guangyey/183/base 2025-12-04T08:57:03.6745776Z * [new branch] gh/guangyey/183/head -> origin/gh/guangyey/183/head 2025-12-04T08:57:03.6747352Z * [new branch] gh/guangyey/183/orig -> origin/gh/guangyey/183/orig 2025-12-04T08:57:03.6749532Z * [new branch] gh/guangyey/185/base -> origin/gh/guangyey/185/base 2025-12-04T08:57:03.6751055Z * [new branch] gh/guangyey/185/head -> origin/gh/guangyey/185/head 2025-12-04T08:57:03.6752601Z * [new branch] gh/guangyey/185/orig -> origin/gh/guangyey/185/orig 2025-12-04T08:57:03.6754672Z * [new branch] gh/guangyey/186/base -> origin/gh/guangyey/186/base 2025-12-04T08:57:03.6756310Z * [new branch] gh/guangyey/186/head -> origin/gh/guangyey/186/head 2025-12-04T08:57:03.6757809Z * [new branch] gh/guangyey/186/orig -> origin/gh/guangyey/186/orig 2025-12-04T08:57:03.6760427Z * [new branch] gh/guangyey/187/base -> origin/gh/guangyey/187/base 2025-12-04T08:57:03.6762191Z * [new branch] gh/guangyey/187/head -> origin/gh/guangyey/187/head 2025-12-04T08:57:03.6763816Z * [new branch] gh/guangyey/187/orig -> origin/gh/guangyey/187/orig 2025-12-04T08:57:03.6766082Z * [new branch] gh/guangyey/188/base -> origin/gh/guangyey/188/base 2025-12-04T08:57:03.6767760Z * [new branch] gh/guangyey/188/head -> origin/gh/guangyey/188/head 2025-12-04T08:57:03.6769323Z * [new branch] gh/guangyey/188/orig -> origin/gh/guangyey/188/orig 2025-12-04T08:57:03.6771444Z * [new branch] gh/guangyey/190/base -> origin/gh/guangyey/190/base 2025-12-04T08:57:03.6772988Z * [new branch] gh/guangyey/190/head -> origin/gh/guangyey/190/head 2025-12-04T08:57:03.6774591Z * [new branch] gh/guangyey/190/orig -> origin/gh/guangyey/190/orig 2025-12-04T08:57:03.6776694Z * [new branch] gh/guangyey/208/base -> origin/gh/guangyey/208/base 2025-12-04T08:57:03.6778286Z * [new branch] gh/guangyey/208/head -> origin/gh/guangyey/208/head 2025-12-04T08:57:03.6779844Z * [new branch] gh/guangyey/208/orig -> origin/gh/guangyey/208/orig 2025-12-04T08:57:03.6781925Z * [new branch] gh/guangyey/228/base -> origin/gh/guangyey/228/base 2025-12-04T08:57:03.6783486Z * [new branch] gh/guangyey/228/head -> origin/gh/guangyey/228/head 2025-12-04T08:57:03.6785063Z * [new branch] gh/guangyey/228/orig -> origin/gh/guangyey/228/orig 2025-12-04T08:57:03.6787687Z * [new branch] gh/guangyey/230/base -> origin/gh/guangyey/230/base 2025-12-04T08:57:03.6789235Z * [new branch] gh/guangyey/230/head -> origin/gh/guangyey/230/head 2025-12-04T08:57:03.6790785Z * [new branch] gh/guangyey/230/orig -> origin/gh/guangyey/230/orig 2025-12-04T08:57:03.6792953Z * [new branch] gh/guangyey/231/base -> origin/gh/guangyey/231/base 2025-12-04T08:57:03.6794533Z * [new branch] gh/guangyey/231/head -> origin/gh/guangyey/231/head 2025-12-04T08:57:03.6796271Z * [new branch] gh/guangyey/231/orig -> origin/gh/guangyey/231/orig 2025-12-04T08:57:03.6798517Z * [new branch] gh/guangyey/232/base -> origin/gh/guangyey/232/base 2025-12-04T08:57:03.6800102Z * [new branch] gh/guangyey/232/head -> origin/gh/guangyey/232/head 2025-12-04T08:57:03.6801664Z * [new branch] gh/guangyey/232/orig -> origin/gh/guangyey/232/orig 2025-12-04T08:57:03.6803837Z * [new branch] gh/guangyey/233/base -> origin/gh/guangyey/233/base 2025-12-04T08:57:03.6805925Z * [new branch] gh/guangyey/233/head -> origin/gh/guangyey/233/head 2025-12-04T08:57:03.6807447Z * [new branch] gh/guangyey/233/orig -> origin/gh/guangyey/233/orig 2025-12-04T08:57:03.6809610Z * [new branch] gh/guangyey/234/base -> origin/gh/guangyey/234/base 2025-12-04T08:57:03.6811170Z * [new branch] gh/guangyey/234/head -> origin/gh/guangyey/234/head 2025-12-04T08:57:03.6812736Z * [new branch] gh/guangyey/234/orig -> origin/gh/guangyey/234/orig 2025-12-04T08:57:03.6815016Z * [new branch] gh/guangyey/235/base -> origin/gh/guangyey/235/base 2025-12-04T08:57:03.6816579Z * [new branch] gh/guangyey/235/head -> origin/gh/guangyey/235/head 2025-12-04T08:57:03.6818178Z * [new branch] gh/guangyey/235/orig -> origin/gh/guangyey/235/orig 2025-12-04T08:57:03.6820384Z * [new branch] gh/guangyey/236/base -> origin/gh/guangyey/236/base 2025-12-04T08:57:03.6821955Z * [new branch] gh/guangyey/236/head -> origin/gh/guangyey/236/head 2025-12-04T08:57:03.6823509Z * [new branch] gh/guangyey/236/orig -> origin/gh/guangyey/236/orig 2025-12-04T08:57:03.6825922Z * [new branch] gh/guangyey/237/base -> origin/gh/guangyey/237/base 2025-12-04T08:57:03.6827369Z * [new branch] gh/guangyey/237/head -> origin/gh/guangyey/237/head 2025-12-04T08:57:03.6828914Z * [new branch] gh/guangyey/237/orig -> origin/gh/guangyey/237/orig 2025-12-04T08:57:03.6831115Z * [new branch] gh/guangyey/238/base -> origin/gh/guangyey/238/base 2025-12-04T08:57:03.6832634Z * [new branch] gh/guangyey/238/head -> origin/gh/guangyey/238/head 2025-12-04T08:57:03.6834827Z * [new branch] gh/guangyey/239/base -> origin/gh/guangyey/239/base 2025-12-04T08:57:03.6836358Z * [new branch] gh/guangyey/239/head -> origin/gh/guangyey/239/head 2025-12-04T08:57:03.6837978Z * [new branch] gh/guangyey/239/orig -> origin/gh/guangyey/239/orig 2025-12-04T08:57:03.6840150Z * [new branch] gh/guangyey/240/base -> origin/gh/guangyey/240/base 2025-12-04T08:57:03.6841773Z * [new branch] gh/guangyey/240/head -> origin/gh/guangyey/240/head 2025-12-04T08:57:03.6843522Z * [new branch] gh/guangyey/240/orig -> origin/gh/guangyey/240/orig 2025-12-04T08:57:03.6845507Z * [new branch] gh/guangyey/241/base -> origin/gh/guangyey/241/base 2025-12-04T08:57:03.6847070Z * [new branch] gh/guangyey/241/head -> origin/gh/guangyey/241/head 2025-12-04T08:57:03.6848627Z * [new branch] gh/guangyey/241/orig -> origin/gh/guangyey/241/orig 2025-12-04T08:57:03.6850822Z * [new branch] gh/guangyey/242/base -> origin/gh/guangyey/242/base 2025-12-04T08:57:03.6852373Z * [new branch] gh/guangyey/242/head -> origin/gh/guangyey/242/head 2025-12-04T08:57:03.6854034Z * [new branch] gh/guangyey/242/orig -> origin/gh/guangyey/242/orig 2025-12-04T08:57:03.6856206Z * [new branch] gh/guangyey/243/base -> origin/gh/guangyey/243/base 2025-12-04T08:57:03.6857766Z * [new branch] gh/guangyey/243/head -> origin/gh/guangyey/243/head 2025-12-04T08:57:03.6859336Z * [new branch] gh/guangyey/243/orig -> origin/gh/guangyey/243/orig 2025-12-04T08:57:03.6861671Z * [new branch] gh/guangyey/244/base -> origin/gh/guangyey/244/base 2025-12-04T08:57:03.6863299Z * [new branch] gh/guangyey/244/head -> origin/gh/guangyey/244/head 2025-12-04T08:57:03.6864886Z * [new branch] gh/guangyey/244/orig -> origin/gh/guangyey/244/orig 2025-12-04T08:57:03.6867036Z * [new branch] gh/guangyey/245/base -> origin/gh/guangyey/245/base 2025-12-04T08:57:03.6868611Z * [new branch] gh/guangyey/245/head -> origin/gh/guangyey/245/head 2025-12-04T08:57:03.6870165Z * [new branch] gh/guangyey/245/orig -> origin/gh/guangyey/245/orig 2025-12-04T08:57:03.6872375Z * [new branch] gh/guangyey/246/base -> origin/gh/guangyey/246/base 2025-12-04T08:57:03.6874043Z * [new branch] gh/guangyey/246/head -> origin/gh/guangyey/246/head 2025-12-04T08:57:03.6875557Z * [new branch] gh/guangyey/246/orig -> origin/gh/guangyey/246/orig 2025-12-04T08:57:03.6877728Z * [new branch] gh/guangyey/247/base -> origin/gh/guangyey/247/base 2025-12-04T08:57:03.6879327Z * [new branch] gh/guangyey/247/head -> origin/gh/guangyey/247/head 2025-12-04T08:57:03.6880937Z * [new branch] gh/guangyey/247/orig -> origin/gh/guangyey/247/orig 2025-12-04T08:57:03.6883106Z * [new branch] gh/guangyey/248/base -> origin/gh/guangyey/248/base 2025-12-04T08:57:03.6884768Z * [new branch] gh/guangyey/248/head -> origin/gh/guangyey/248/head 2025-12-04T08:57:03.6886402Z * [new branch] gh/guangyey/248/orig -> origin/gh/guangyey/248/orig 2025-12-04T08:57:03.6888722Z * [new branch] gh/guangyey/249/base -> origin/gh/guangyey/249/base 2025-12-04T08:57:03.6890152Z * [new branch] gh/guangyey/249/head -> origin/gh/guangyey/249/head 2025-12-04T08:57:03.6891784Z * [new branch] gh/guangyey/249/orig -> origin/gh/guangyey/249/orig 2025-12-04T08:57:03.6894122Z * [new branch] gh/guangyey/250/base -> origin/gh/guangyey/250/base 2025-12-04T08:57:03.6895641Z * [new branch] gh/guangyey/250/head -> origin/gh/guangyey/250/head 2025-12-04T08:57:03.6897232Z * [new branch] gh/guangyey/250/orig -> origin/gh/guangyey/250/orig 2025-12-04T08:57:03.6899377Z * [new branch] gh/guangyey/251/base -> origin/gh/guangyey/251/base 2025-12-04T08:57:03.6900927Z * [new branch] gh/guangyey/251/head -> origin/gh/guangyey/251/head 2025-12-04T08:57:03.6902488Z * [new branch] gh/guangyey/251/orig -> origin/gh/guangyey/251/orig 2025-12-04T08:57:03.6905005Z * [new branch] gh/guangyey/252/base -> origin/gh/guangyey/252/base 2025-12-04T08:57:03.6907051Z * [new branch] gh/guangyey/252/head -> origin/gh/guangyey/252/head 2025-12-04T08:57:03.6908522Z * [new branch] gh/guangyey/252/orig -> origin/gh/guangyey/252/orig 2025-12-04T08:57:03.6910616Z * [new branch] gh/guangyey/253/base -> origin/gh/guangyey/253/base 2025-12-04T08:57:03.6912313Z * [new branch] gh/guangyey/253/head -> origin/gh/guangyey/253/head 2025-12-04T08:57:03.6913976Z * [new branch] gh/guangyey/253/orig -> origin/gh/guangyey/253/orig 2025-12-04T08:57:03.6916194Z * [new branch] gh/guangyey/254/base -> origin/gh/guangyey/254/base 2025-12-04T08:57:03.6917863Z * [new branch] gh/guangyey/254/head -> origin/gh/guangyey/254/head 2025-12-04T08:57:03.6919393Z * [new branch] gh/guangyey/254/orig -> origin/gh/guangyey/254/orig 2025-12-04T08:57:03.6921500Z * [new branch] gh/guangyey/255/base -> origin/gh/guangyey/255/base 2025-12-04T08:57:03.6923115Z * [new branch] gh/guangyey/255/head -> origin/gh/guangyey/255/head 2025-12-04T08:57:03.6924580Z * [new branch] gh/guangyey/255/orig -> origin/gh/guangyey/255/orig 2025-12-04T08:57:03.6927351Z * [new branch] gh/guilhermeleobas/107/base -> origin/gh/guilhermeleobas/107/base 2025-12-04T08:57:03.6928921Z * [new branch] gh/guilhermeleobas/107/head -> origin/gh/guilhermeleobas/107/head 2025-12-04T08:57:03.6930492Z * [new branch] gh/guilhermeleobas/107/orig -> origin/gh/guilhermeleobas/107/orig 2025-12-04T08:57:03.6932732Z * [new branch] gh/guilhermeleobas/108/base -> origin/gh/guilhermeleobas/108/base 2025-12-04T08:57:03.6934592Z * [new branch] gh/guilhermeleobas/108/head -> origin/gh/guilhermeleobas/108/head 2025-12-04T08:57:03.6936083Z * [new branch] gh/guilhermeleobas/108/orig -> origin/gh/guilhermeleobas/108/orig 2025-12-04T08:57:03.6938248Z * [new branch] gh/guilhermeleobas/150/base -> origin/gh/guilhermeleobas/150/base 2025-12-04T08:57:03.6939574Z * [new branch] gh/guilhermeleobas/150/head -> origin/gh/guilhermeleobas/150/head 2025-12-04T08:57:03.6942109Z * [new branch] gh/guilhermeleobas/150/orig -> origin/gh/guilhermeleobas/150/orig 2025-12-04T08:57:03.6944268Z * [new branch] gh/guilhermeleobas/168/base -> origin/gh/guilhermeleobas/168/base 2025-12-04T08:57:03.6945873Z * [new branch] gh/guilhermeleobas/168/head -> origin/gh/guilhermeleobas/168/head 2025-12-04T08:57:03.6947400Z * [new branch] gh/guilhermeleobas/168/orig -> origin/gh/guilhermeleobas/168/orig 2025-12-04T08:57:03.6949696Z * [new branch] gh/guilhermeleobas/169/base -> origin/gh/guilhermeleobas/169/base 2025-12-04T08:57:03.6951310Z * [new branch] gh/guilhermeleobas/169/head -> origin/gh/guilhermeleobas/169/head 2025-12-04T08:57:03.6952857Z * [new branch] gh/guilhermeleobas/169/orig -> origin/gh/guilhermeleobas/169/orig 2025-12-04T08:57:03.6954770Z * [new branch] gh/guilhermeleobas/170/base -> origin/gh/guilhermeleobas/170/base 2025-12-04T08:57:03.6956365Z * [new branch] gh/guilhermeleobas/170/head -> origin/gh/guilhermeleobas/170/head 2025-12-04T08:57:03.6957978Z * [new branch] gh/guilhermeleobas/170/orig -> origin/gh/guilhermeleobas/170/orig 2025-12-04T08:57:03.6960083Z * [new branch] gh/guilhermeleobas/171/base -> origin/gh/guilhermeleobas/171/base 2025-12-04T08:57:03.6961669Z * [new branch] gh/guilhermeleobas/171/head -> origin/gh/guilhermeleobas/171/head 2025-12-04T08:57:03.6963223Z * [new branch] gh/guilhermeleobas/171/orig -> origin/gh/guilhermeleobas/171/orig 2025-12-04T08:57:03.6965168Z * [new branch] gh/guilhermeleobas/173/base -> origin/gh/guilhermeleobas/173/base 2025-12-04T08:57:03.6966826Z * [new branch] gh/guilhermeleobas/173/head -> origin/gh/guilhermeleobas/173/head 2025-12-04T08:57:03.6968379Z * [new branch] gh/guilhermeleobas/173/orig -> origin/gh/guilhermeleobas/173/orig 2025-12-04T08:57:03.6970526Z * [new branch] gh/guilhermeleobas/193/base -> origin/gh/guilhermeleobas/193/base 2025-12-04T08:57:03.6972115Z * [new branch] gh/guilhermeleobas/193/head -> origin/gh/guilhermeleobas/193/head 2025-12-04T08:57:03.6974247Z * [new branch] gh/guilhermeleobas/193/orig -> origin/gh/guilhermeleobas/193/orig 2025-12-04T08:57:03.6976436Z * [new branch] gh/guilhermeleobas/204/base -> origin/gh/guilhermeleobas/204/base 2025-12-04T08:57:03.6977982Z * [new branch] gh/guilhermeleobas/204/head -> origin/gh/guilhermeleobas/204/head 2025-12-04T08:57:03.6979544Z * [new branch] gh/guilhermeleobas/204/orig -> origin/gh/guilhermeleobas/204/orig 2025-12-04T08:57:03.6981649Z * [new branch] gh/guilhermeleobas/211/base -> origin/gh/guilhermeleobas/211/base 2025-12-04T08:57:03.6983574Z * [new branch] gh/guilhermeleobas/211/head -> origin/gh/guilhermeleobas/211/head 2025-12-04T08:57:03.6985263Z * [new branch] gh/guilhermeleobas/211/orig -> origin/gh/guilhermeleobas/211/orig 2025-12-04T08:57:03.6987351Z * [new branch] gh/guilhermeleobas/226/base -> origin/gh/guilhermeleobas/226/base 2025-12-04T08:57:03.6988907Z * [new branch] gh/guilhermeleobas/226/head -> origin/gh/guilhermeleobas/226/head 2025-12-04T08:57:03.6990428Z * [new branch] gh/guilhermeleobas/226/orig -> origin/gh/guilhermeleobas/226/orig 2025-12-04T08:57:03.6992625Z * [new branch] gh/guilhermeleobas/236/base -> origin/gh/guilhermeleobas/236/base 2025-12-04T08:57:03.6994199Z * [new branch] gh/guilhermeleobas/236/head -> origin/gh/guilhermeleobas/236/head 2025-12-04T08:57:03.6995754Z * [new branch] gh/guilhermeleobas/236/orig -> origin/gh/guilhermeleobas/236/orig 2025-12-04T08:57:03.6997841Z * [new branch] gh/guilhermeleobas/247/base -> origin/gh/guilhermeleobas/247/base 2025-12-04T08:57:03.6999389Z * [new branch] gh/guilhermeleobas/247/head -> origin/gh/guilhermeleobas/247/head 2025-12-04T08:57:03.7000964Z * [new branch] gh/guilhermeleobas/247/orig -> origin/gh/guilhermeleobas/247/orig 2025-12-04T08:57:03.7003054Z * [new branch] gh/guilhermeleobas/248/base -> origin/gh/guilhermeleobas/248/base 2025-12-04T08:57:03.7004759Z * [new branch] gh/guilhermeleobas/248/head -> origin/gh/guilhermeleobas/248/head 2025-12-04T08:57:03.7007040Z * [new branch] gh/guilhermeleobas/248/orig -> origin/gh/guilhermeleobas/248/orig 2025-12-04T08:57:03.7009654Z * [new branch] gh/guilhermeleobas/250/base -> origin/gh/guilhermeleobas/250/base 2025-12-04T08:57:03.7011351Z * [new branch] gh/guilhermeleobas/250/head -> origin/gh/guilhermeleobas/250/head 2025-12-04T08:57:03.7014233Z * [new branch] gh/guilhermeleobas/250/orig -> origin/gh/guilhermeleobas/250/orig 2025-12-04T08:57:03.7015840Z * [new branch] gh/guilhermeleobas/253/base -> origin/gh/guilhermeleobas/253/base 2025-12-04T08:57:03.7017098Z * [new branch] gh/guilhermeleobas/253/head -> origin/gh/guilhermeleobas/253/head 2025-12-04T08:57:03.7018726Z * [new branch] gh/guilhermeleobas/253/orig -> origin/gh/guilhermeleobas/253/orig 2025-12-04T08:57:03.7020917Z * [new branch] gh/guilhermeleobas/254/base -> origin/gh/guilhermeleobas/254/base 2025-12-04T08:57:03.7022518Z * [new branch] gh/guilhermeleobas/254/head -> origin/gh/guilhermeleobas/254/head 2025-12-04T08:57:03.7024040Z * [new branch] gh/guilhermeleobas/254/orig -> origin/gh/guilhermeleobas/254/orig 2025-12-04T08:57:03.7026183Z * [new branch] gh/guilhermeleobas/255/base -> origin/gh/guilhermeleobas/255/base 2025-12-04T08:57:03.7027763Z * [new branch] gh/guilhermeleobas/255/head -> origin/gh/guilhermeleobas/255/head 2025-12-04T08:57:03.7029347Z * [new branch] gh/guilhermeleobas/255/orig -> origin/gh/guilhermeleobas/255/orig 2025-12-04T08:57:03.7031579Z * [new branch] gh/guilhermeleobas/256/base -> origin/gh/guilhermeleobas/256/base 2025-12-04T08:57:03.7033307Z * [new branch] gh/guilhermeleobas/256/head -> origin/gh/guilhermeleobas/256/head 2025-12-04T08:57:03.7034675Z * [new branch] gh/guilhermeleobas/256/orig -> origin/gh/guilhermeleobas/256/orig 2025-12-04T08:57:03.7037107Z * [new branch] gh/guilhermeleobas/257/base -> origin/gh/guilhermeleobas/257/base 2025-12-04T08:57:03.7038478Z * [new branch] gh/guilhermeleobas/257/head -> origin/gh/guilhermeleobas/257/head 2025-12-04T08:57:03.7040157Z * [new branch] gh/guilhermeleobas/257/orig -> origin/gh/guilhermeleobas/257/orig 2025-12-04T08:57:03.7042294Z * [new branch] gh/guilhermeleobas/258/base -> origin/gh/guilhermeleobas/258/base 2025-12-04T08:57:03.7043891Z * [new branch] gh/guilhermeleobas/258/head -> origin/gh/guilhermeleobas/258/head 2025-12-04T08:57:03.7045512Z * [new branch] gh/guilhermeleobas/258/orig -> origin/gh/guilhermeleobas/258/orig 2025-12-04T08:57:03.7047715Z * [new branch] gh/guilhermeleobas/259/base -> origin/gh/guilhermeleobas/259/base 2025-12-04T08:57:03.7049253Z * [new branch] gh/guilhermeleobas/259/head -> origin/gh/guilhermeleobas/259/head 2025-12-04T08:57:03.7050799Z * [new branch] gh/guilhermeleobas/259/orig -> origin/gh/guilhermeleobas/259/orig 2025-12-04T08:57:03.7052951Z * [new branch] gh/guilhermeleobas/260/base -> origin/gh/guilhermeleobas/260/base 2025-12-04T08:57:03.7055189Z * [new branch] gh/guilhermeleobas/260/head -> origin/gh/guilhermeleobas/260/head 2025-12-04T08:57:03.7056757Z * [new branch] gh/guilhermeleobas/260/orig -> origin/gh/guilhermeleobas/260/orig 2025-12-04T08:57:03.7058983Z * [new branch] gh/guilhermeleobas/261/base -> origin/gh/guilhermeleobas/261/base 2025-12-04T08:57:03.7060555Z * [new branch] gh/guilhermeleobas/261/head -> origin/gh/guilhermeleobas/261/head 2025-12-04T08:57:03.7062065Z * [new branch] gh/guilhermeleobas/261/orig -> origin/gh/guilhermeleobas/261/orig 2025-12-04T08:57:03.7064212Z * [new branch] gh/guilhermeleobas/262/base -> origin/gh/guilhermeleobas/262/base 2025-12-04T08:57:03.7065981Z * [new branch] gh/guilhermeleobas/262/head -> origin/gh/guilhermeleobas/262/head 2025-12-04T08:57:03.7067488Z * [new branch] gh/guilhermeleobas/262/orig -> origin/gh/guilhermeleobas/262/orig 2025-12-04T08:57:03.7069644Z * [new branch] gh/guilhermeleobas/263/base -> origin/gh/guilhermeleobas/263/base 2025-12-04T08:57:03.7071306Z * [new branch] gh/guilhermeleobas/263/head -> origin/gh/guilhermeleobas/263/head 2025-12-04T08:57:03.7072744Z * [new branch] gh/guilhermeleobas/263/orig -> origin/gh/guilhermeleobas/263/orig 2025-12-04T08:57:03.7075299Z * [new branch] gh/guilhermeleobas/264/base -> origin/gh/guilhermeleobas/264/base 2025-12-04T08:57:03.7076458Z * [new branch] gh/guilhermeleobas/264/head -> origin/gh/guilhermeleobas/264/head 2025-12-04T08:57:03.7077967Z * [new branch] gh/guilhermeleobas/264/orig -> origin/gh/guilhermeleobas/264/orig 2025-12-04T08:57:03.7080239Z * [new branch] gh/guilhermeleobas/265/base -> origin/gh/guilhermeleobas/265/base 2025-12-04T08:57:03.7081822Z * [new branch] gh/guilhermeleobas/265/head -> origin/gh/guilhermeleobas/265/head 2025-12-04T08:57:03.7083485Z * [new branch] gh/guilhermeleobas/265/orig -> origin/gh/guilhermeleobas/265/orig 2025-12-04T08:57:03.7086116Z * [new branch] gh/guilhermeleobas/266/base -> origin/gh/guilhermeleobas/266/base 2025-12-04T08:57:03.7087779Z * [new branch] gh/guilhermeleobas/266/head -> origin/gh/guilhermeleobas/266/head 2025-12-04T08:57:03.7089355Z * [new branch] gh/guilhermeleobas/266/orig -> origin/gh/guilhermeleobas/266/orig 2025-12-04T08:57:03.7091590Z * [new branch] gh/guilhermeleobas/267/base -> origin/gh/guilhermeleobas/267/base 2025-12-04T08:57:03.7093128Z * [new branch] gh/guilhermeleobas/267/head -> origin/gh/guilhermeleobas/267/head 2025-12-04T08:57:03.7094796Z * [new branch] gh/guilhermeleobas/267/orig -> origin/gh/guilhermeleobas/267/orig 2025-12-04T08:57:03.7097318Z * [new branch] gh/hameerabbasi/1/base -> origin/gh/hameerabbasi/1/base 2025-12-04T08:57:03.7098897Z * [new branch] gh/hameerabbasi/1/head -> origin/gh/hameerabbasi/1/head 2025-12-04T08:57:03.7101001Z * [new branch] gh/hameerabbasi/2/base -> origin/gh/hameerabbasi/2/base 2025-12-04T08:57:03.7102573Z * [new branch] gh/hameerabbasi/2/head -> origin/gh/hameerabbasi/2/head 2025-12-04T08:57:03.7104163Z * [new branch] gh/hameerabbasi/2/orig -> origin/gh/hameerabbasi/2/orig 2025-12-04T08:57:03.7108179Z * [new branch] gh/hameerabbasi/3/base -> origin/gh/hameerabbasi/3/base 2025-12-04T08:57:03.7109743Z * [new branch] gh/hameerabbasi/3/head -> origin/gh/hameerabbasi/3/head 2025-12-04T08:57:03.7111441Z * [new branch] gh/hameerabbasi/3/orig -> origin/gh/hameerabbasi/3/orig 2025-12-04T08:57:03.7113409Z * [new branch] gh/hameerabbasi/4/base -> origin/gh/hameerabbasi/4/base 2025-12-04T08:57:03.7114895Z * [new branch] gh/hameerabbasi/4/head -> origin/gh/hameerabbasi/4/head 2025-12-04T08:57:03.7116372Z * [new branch] gh/hameerabbasi/4/orig -> origin/gh/hameerabbasi/4/orig 2025-12-04T08:57:03.7118909Z * [new branch] gh/huydhn/1/next -> origin/gh/huydhn/1/next 2025-12-04T08:57:03.7120848Z * [new branch] gh/huydhn/2/next -> origin/gh/huydhn/2/next 2025-12-04T08:57:03.7122989Z * [new branch] gh/huydhn/3/next -> origin/gh/huydhn/3/next 2025-12-04T08:57:03.7125160Z * [new branch] gh/huydhn/4/next -> origin/gh/huydhn/4/next 2025-12-04T08:57:03.7127234Z * [new branch] gh/huydhn/5/next -> origin/gh/huydhn/5/next 2025-12-04T08:57:03.7129251Z * [new branch] gh/huydhn/6/next -> origin/gh/huydhn/6/next 2025-12-04T08:57:03.7131738Z * [new branch] gh/int3/97/base -> origin/gh/int3/97/base 2025-12-04T08:57:03.7133396Z * [new branch] gh/int3/97/head -> origin/gh/int3/97/head 2025-12-04T08:57:03.7136026Z * [new branch] gh/isuruf/101/base -> origin/gh/isuruf/101/base 2025-12-04T08:57:03.7137715Z * [new branch] gh/isuruf/101/head -> origin/gh/isuruf/101/head 2025-12-04T08:57:03.7139786Z * [new branch] gh/isuruf/146/base -> origin/gh/isuruf/146/base 2025-12-04T08:57:03.7141410Z * [new branch] gh/isuruf/146/head -> origin/gh/isuruf/146/head 2025-12-04T08:57:03.7142911Z * [new branch] gh/isuruf/146/orig -> origin/gh/isuruf/146/orig 2025-12-04T08:57:03.7144993Z * [new branch] gh/isuruf/158/base -> origin/gh/isuruf/158/base 2025-12-04T08:57:03.7146588Z * [new branch] gh/isuruf/158/head -> origin/gh/isuruf/158/head 2025-12-04T08:57:03.7148716Z * [new branch] gh/isuruf/159/base -> origin/gh/isuruf/159/base 2025-12-04T08:57:03.7150226Z * [new branch] gh/isuruf/159/head -> origin/gh/isuruf/159/head 2025-12-04T08:57:03.7152317Z * [new branch] gh/isuruf/160/base -> origin/gh/isuruf/160/base 2025-12-04T08:57:03.7153883Z * [new branch] gh/isuruf/160/head -> origin/gh/isuruf/160/head 2025-12-04T08:57:03.7155452Z * [new branch] gh/isuruf/160/orig -> origin/gh/isuruf/160/orig 2025-12-04T08:57:03.7157552Z * [new branch] gh/isuruf/81/base -> origin/gh/isuruf/81/base 2025-12-04T08:57:03.7159174Z * [new branch] gh/isuruf/81/head -> origin/gh/isuruf/81/head 2025-12-04T08:57:03.7160705Z * [new branch] gh/isuruf/81/orig -> origin/gh/isuruf/81/orig 2025-12-04T08:57:03.7163215Z * [new branch] gh/jamesjwu/176/base -> origin/gh/jamesjwu/176/base 2025-12-04T08:57:03.7164734Z * [new branch] gh/jamesjwu/176/head -> origin/gh/jamesjwu/176/head 2025-12-04T08:57:03.7166296Z * [new branch] gh/jamesjwu/176/orig -> origin/gh/jamesjwu/176/orig 2025-12-04T08:57:03.7168491Z * [new branch] gh/jamesjwu/187/base -> origin/gh/jamesjwu/187/base 2025-12-04T08:57:03.7169897Z * [new branch] gh/jamesjwu/187/head -> origin/gh/jamesjwu/187/head 2025-12-04T08:57:03.7171425Z * [new branch] gh/jamesjwu/187/orig -> origin/gh/jamesjwu/187/orig 2025-12-04T08:57:03.7173902Z * [new branch] gh/jamesjwu/196/base -> origin/gh/jamesjwu/196/base 2025-12-04T08:57:03.7175450Z * [new branch] gh/jamesjwu/196/head -> origin/gh/jamesjwu/196/head 2025-12-04T08:57:03.7176932Z * [new branch] gh/jamesjwu/196/orig -> origin/gh/jamesjwu/196/orig 2025-12-04T08:57:03.7179059Z * [new branch] gh/jamesjwu/198/base -> origin/gh/jamesjwu/198/base 2025-12-04T08:57:03.7180631Z * [new branch] gh/jamesjwu/198/head -> origin/gh/jamesjwu/198/head 2025-12-04T08:57:03.7182198Z * [new branch] gh/jamesjwu/198/orig -> origin/gh/jamesjwu/198/orig 2025-12-04T08:57:03.7184279Z * [new branch] gh/jamesjwu/207/base -> origin/gh/jamesjwu/207/base 2025-12-04T08:57:03.7185964Z * [new branch] gh/jamesjwu/207/head -> origin/gh/jamesjwu/207/head 2025-12-04T08:57:03.7187540Z * [new branch] gh/jamesjwu/207/orig -> origin/gh/jamesjwu/207/orig 2025-12-04T08:57:03.7189686Z * [new branch] gh/jamesjwu/208/base -> origin/gh/jamesjwu/208/base 2025-12-04T08:57:03.7191216Z * [new branch] gh/jamesjwu/208/head -> origin/gh/jamesjwu/208/head 2025-12-04T08:57:03.7192783Z * [new branch] gh/jamesjwu/208/orig -> origin/gh/jamesjwu/208/orig 2025-12-04T08:57:03.7194932Z * [new branch] gh/jamesjwu/52/base -> origin/gh/jamesjwu/52/base 2025-12-04T08:57:03.7196454Z * [new branch] gh/jamesjwu/52/head -> origin/gh/jamesjwu/52/head 2025-12-04T08:57:03.7198866Z * [new branch] gh/jamesjwu/53/base -> origin/gh/jamesjwu/53/base 2025-12-04T08:57:03.7200480Z * [new branch] gh/jamesjwu/53/head -> origin/gh/jamesjwu/53/head 2025-12-04T08:57:03.7202528Z * [new branch] gh/jamesjwu/54/base -> origin/gh/jamesjwu/54/base 2025-12-04T08:57:03.7204060Z * [new branch] gh/jamesjwu/54/head -> origin/gh/jamesjwu/54/head 2025-12-04T08:57:03.7206525Z * [new branch] gh/jamesjwu/55/base -> origin/gh/jamesjwu/55/base 2025-12-04T08:57:03.7208046Z * [new branch] gh/jamesjwu/55/head -> origin/gh/jamesjwu/55/head 2025-12-04T08:57:03.7209996Z * [new branch] gh/jamesjwu/56/base -> origin/gh/jamesjwu/56/base 2025-12-04T08:57:03.7211489Z * [new branch] gh/jamesjwu/56/head -> origin/gh/jamesjwu/56/head 2025-12-04T08:57:03.7213500Z * [new branch] gh/jamesjwu/57/base -> origin/gh/jamesjwu/57/base 2025-12-04T08:57:03.7215122Z * [new branch] gh/jamesjwu/57/head -> origin/gh/jamesjwu/57/head 2025-12-04T08:57:03.7217014Z * [new branch] gh/jamesjwu/58/base -> origin/gh/jamesjwu/58/base 2025-12-04T08:57:03.7218536Z * [new branch] gh/jamesjwu/58/head -> origin/gh/jamesjwu/58/head 2025-12-04T08:57:03.7220637Z * [new branch] gh/jamesjwu/59/base -> origin/gh/jamesjwu/59/base 2025-12-04T08:57:03.7222190Z * [new branch] gh/jamesjwu/59/head -> origin/gh/jamesjwu/59/head 2025-12-04T08:57:03.7224127Z * [new branch] gh/jamesjwu/60/base -> origin/gh/jamesjwu/60/base 2025-12-04T08:57:03.7225698Z * [new branch] gh/jamesjwu/60/head -> origin/gh/jamesjwu/60/head 2025-12-04T08:57:03.7227659Z * [new branch] gh/jamesjwu/61/base -> origin/gh/jamesjwu/61/base 2025-12-04T08:57:03.7229182Z * [new branch] gh/jamesjwu/61/head -> origin/gh/jamesjwu/61/head 2025-12-04T08:57:03.7231111Z * [new branch] gh/jamesjwu/62/base -> origin/gh/jamesjwu/62/base 2025-12-04T08:57:03.7232629Z * [new branch] gh/jamesjwu/62/head -> origin/gh/jamesjwu/62/head 2025-12-04T08:57:03.7234563Z * [new branch] gh/jamesjwu/63/base -> origin/gh/jamesjwu/63/base 2025-12-04T08:57:03.7236156Z * [new branch] gh/jamesjwu/63/head -> origin/gh/jamesjwu/63/head 2025-12-04T08:57:03.7238868Z * [new branch] gh/jamesjwu/64/base -> origin/gh/jamesjwu/64/base 2025-12-04T08:57:03.7240412Z * [new branch] gh/jamesjwu/64/head -> origin/gh/jamesjwu/64/head 2025-12-04T08:57:03.7242502Z * [new branch] gh/jamesjwu/65/base -> origin/gh/jamesjwu/65/base 2025-12-04T08:57:03.7244014Z * [new branch] gh/jamesjwu/65/head -> origin/gh/jamesjwu/65/head 2025-12-04T08:57:03.7247046Z * [new branch] gh/janeyx99/165/base -> origin/gh/janeyx99/165/base 2025-12-04T08:57:03.7248656Z * [new branch] gh/janeyx99/165/head -> origin/gh/janeyx99/165/head 2025-12-04T08:57:03.7250501Z * [new branch] gh/janeyx99/165/orig -> origin/gh/janeyx99/165/orig 2025-12-04T08:57:03.7252407Z * [new branch] gh/janeyx99/201/base -> origin/gh/janeyx99/201/base 2025-12-04T08:57:03.7254237Z * [new branch] gh/janeyx99/201/head -> origin/gh/janeyx99/201/head 2025-12-04T08:57:03.7255773Z * [new branch] gh/janeyx99/201/orig -> origin/gh/janeyx99/201/orig 2025-12-04T08:57:03.7258072Z * [new branch] gh/janeyx99/225/base -> origin/gh/janeyx99/225/base 2025-12-04T08:57:03.7259685Z * [new branch] gh/janeyx99/225/head -> origin/gh/janeyx99/225/head 2025-12-04T08:57:03.7261332Z * [new branch] gh/janeyx99/225/orig -> origin/gh/janeyx99/225/orig 2025-12-04T08:57:03.7263569Z * [new branch] gh/janeyx99/299/base -> origin/gh/janeyx99/299/base 2025-12-04T08:57:03.7265285Z * [new branch] gh/janeyx99/299/head -> origin/gh/janeyx99/299/head 2025-12-04T08:57:03.7266970Z * [new branch] gh/janeyx99/299/orig -> origin/gh/janeyx99/299/orig 2025-12-04T08:57:03.7269218Z * [new branch] gh/janeyx99/302/base -> origin/gh/janeyx99/302/base 2025-12-04T08:57:03.7270768Z * [new branch] gh/janeyx99/302/head -> origin/gh/janeyx99/302/head 2025-12-04T08:57:03.7272676Z * [new branch] gh/janeyx99/303/base -> origin/gh/janeyx99/303/base 2025-12-04T08:57:03.7274413Z * [new branch] gh/janeyx99/303/head -> origin/gh/janeyx99/303/head 2025-12-04T08:57:03.7276552Z * [new branch] gh/janeyx99/305/base -> origin/gh/janeyx99/305/base 2025-12-04T08:57:03.7278171Z * [new branch] gh/janeyx99/305/head -> origin/gh/janeyx99/305/head 2025-12-04T08:57:03.7280178Z * [new branch] gh/janeyx99/306/base -> origin/gh/janeyx99/306/base 2025-12-04T08:57:03.7281757Z * [new branch] gh/janeyx99/306/head -> origin/gh/janeyx99/306/head 2025-12-04T08:57:03.7284053Z * [new branch] gh/janeyx99/314/base -> origin/gh/janeyx99/314/base 2025-12-04T08:57:03.7285796Z * [new branch] gh/janeyx99/314/head -> origin/gh/janeyx99/314/head 2025-12-04T08:57:03.7287321Z * [new branch] gh/janeyx99/314/orig -> origin/gh/janeyx99/314/orig 2025-12-04T08:57:03.7289461Z * [new branch] gh/janeyx99/315/base -> origin/gh/janeyx99/315/base 2025-12-04T08:57:03.7291019Z * [new branch] gh/janeyx99/315/head -> origin/gh/janeyx99/315/head 2025-12-04T08:57:03.7292559Z * [new branch] gh/janeyx99/315/orig -> origin/gh/janeyx99/315/orig 2025-12-04T08:57:03.7295154Z * [new branch] gh/janeyx99/316/base -> origin/gh/janeyx99/316/base 2025-12-04T08:57:03.7296665Z * [new branch] gh/janeyx99/316/head -> origin/gh/janeyx99/316/head 2025-12-04T08:57:03.7298286Z * [new branch] gh/janeyx99/316/orig -> origin/gh/janeyx99/316/orig 2025-12-04T08:57:03.7300498Z * [new branch] gh/janeyx99/317/base -> origin/gh/janeyx99/317/base 2025-12-04T08:57:03.7302101Z * [new branch] gh/janeyx99/317/head -> origin/gh/janeyx99/317/head 2025-12-04T08:57:03.7303630Z * [new branch] gh/janeyx99/317/orig -> origin/gh/janeyx99/317/orig 2025-12-04T08:57:03.7306125Z * [new branch] gh/janeyx99/325/base -> origin/gh/janeyx99/325/base 2025-12-04T08:57:03.7307692Z * [new branch] gh/janeyx99/325/head -> origin/gh/janeyx99/325/head 2025-12-04T08:57:03.7309226Z * [new branch] gh/janeyx99/325/orig -> origin/gh/janeyx99/325/orig 2025-12-04T08:57:03.7311350Z * [new branch] gh/janeyx99/327/base -> origin/gh/janeyx99/327/base 2025-12-04T08:57:03.7312930Z * [new branch] gh/janeyx99/327/head -> origin/gh/janeyx99/327/head 2025-12-04T08:57:03.7314446Z * [new branch] gh/janeyx99/327/orig -> origin/gh/janeyx99/327/orig 2025-12-04T08:57:03.7316515Z * [new branch] gh/janeyx99/328/base -> origin/gh/janeyx99/328/base 2025-12-04T08:57:03.7318171Z * [new branch] gh/janeyx99/328/head -> origin/gh/janeyx99/328/head 2025-12-04T08:57:03.7319780Z * [new branch] gh/janeyx99/328/orig -> origin/gh/janeyx99/328/orig 2025-12-04T08:57:03.7321701Z * [new branch] gh/janeyx99/329/base -> origin/gh/janeyx99/329/base 2025-12-04T08:57:03.7323391Z * [new branch] gh/janeyx99/329/head -> origin/gh/janeyx99/329/head 2025-12-04T08:57:03.7325001Z * [new branch] gh/janeyx99/329/orig -> origin/gh/janeyx99/329/orig 2025-12-04T08:57:03.7327653Z * [new branch] gh/janeyx99/330/base -> origin/gh/janeyx99/330/base 2025-12-04T08:57:03.7329247Z * [new branch] gh/janeyx99/330/head -> origin/gh/janeyx99/330/head 2025-12-04T08:57:03.7330907Z * [new branch] gh/janeyx99/330/orig -> origin/gh/janeyx99/330/orig 2025-12-04T08:57:03.7333169Z * [new branch] gh/janeyx99/331/base -> origin/gh/janeyx99/331/base 2025-12-04T08:57:03.7334849Z * [new branch] gh/janeyx99/331/head -> origin/gh/janeyx99/331/head 2025-12-04T08:57:03.7336351Z * [new branch] gh/janeyx99/331/orig -> origin/gh/janeyx99/331/orig 2025-12-04T08:57:03.7338414Z * [new branch] gh/janeyx99/332/base -> origin/gh/janeyx99/332/base 2025-12-04T08:57:03.7340000Z * [new branch] gh/janeyx99/332/head -> origin/gh/janeyx99/332/head 2025-12-04T08:57:03.7341504Z * [new branch] gh/janeyx99/332/orig -> origin/gh/janeyx99/332/orig 2025-12-04T08:57:03.7343717Z * [new branch] gh/janeyx99/333/base -> origin/gh/janeyx99/333/base 2025-12-04T08:57:03.7345782Z * [new branch] gh/janeyx99/333/head -> origin/gh/janeyx99/333/head 2025-12-04T08:57:03.7347087Z * [new branch] gh/janeyx99/333/orig -> origin/gh/janeyx99/333/orig 2025-12-04T08:57:03.7349201Z * [new branch] gh/janeyx99/88/base -> origin/gh/janeyx99/88/base 2025-12-04T08:57:03.7350737Z * [new branch] gh/janeyx99/88/head -> origin/gh/janeyx99/88/head 2025-12-04T08:57:03.7352270Z * [new branch] gh/janeyx99/88/orig -> origin/gh/janeyx99/88/orig 2025-12-04T08:57:03.7355599Z * [new branch] gh/jansel/360/base -> origin/gh/jansel/360/base 2025-12-04T08:57:03.7357171Z * [new branch] gh/jansel/360/head -> origin/gh/jansel/360/head 2025-12-04T08:57:03.7359260Z * [new branch] gh/jansel/451/base -> origin/gh/jansel/451/base 2025-12-04T08:57:03.7360795Z * [new branch] gh/jansel/451/head -> origin/gh/jansel/451/head 2025-12-04T08:57:03.7362591Z * [new branch] gh/jansel/451/orig -> origin/gh/jansel/451/orig 2025-12-04T08:57:03.7364632Z * [new branch] gh/jansel/462/base -> origin/gh/jansel/462/base 2025-12-04T08:57:03.7366292Z * [new branch] gh/jansel/462/head -> origin/gh/jansel/462/head 2025-12-04T08:57:03.7367801Z * [new branch] gh/jansel/462/orig -> origin/gh/jansel/462/orig 2025-12-04T08:57:03.7369968Z * [new branch] gh/jansel/533/base -> origin/gh/jansel/533/base 2025-12-04T08:57:03.7371597Z * [new branch] gh/jansel/533/head -> origin/gh/jansel/533/head 2025-12-04T08:57:03.7373272Z * [new branch] gh/jansel/533/orig -> origin/gh/jansel/533/orig 2025-12-04T08:57:03.7375423Z * [new branch] gh/jansel/552/base -> origin/gh/jansel/552/base 2025-12-04T08:57:03.7376929Z * [new branch] gh/jansel/552/head -> origin/gh/jansel/552/head 2025-12-04T08:57:03.7378488Z * [new branch] gh/jansel/552/orig -> origin/gh/jansel/552/orig 2025-12-04T08:57:03.7380578Z * [new branch] gh/jansel/553/base -> origin/gh/jansel/553/base 2025-12-04T08:57:03.7382139Z * [new branch] gh/jansel/553/head -> origin/gh/jansel/553/head 2025-12-04T08:57:03.7383592Z * [new branch] gh/jansel/553/orig -> origin/gh/jansel/553/orig 2025-12-04T08:57:03.7386188Z * [new branch] gh/jansel/554/base -> origin/gh/jansel/554/base 2025-12-04T08:57:03.7387834Z * [new branch] gh/jansel/554/head -> origin/gh/jansel/554/head 2025-12-04T08:57:03.7389397Z * [new branch] gh/jansel/554/orig -> origin/gh/jansel/554/orig 2025-12-04T08:57:03.7391452Z * [new branch] gh/jansel/555/base -> origin/gh/jansel/555/base 2025-12-04T08:57:03.7393002Z * [new branch] gh/jansel/555/head -> origin/gh/jansel/555/head 2025-12-04T08:57:03.7394597Z * [new branch] gh/jansel/555/orig -> origin/gh/jansel/555/orig 2025-12-04T08:57:03.7396613Z * [new branch] gh/jansel/556/base -> origin/gh/jansel/556/base 2025-12-04T08:57:03.7398158Z * [new branch] gh/jansel/556/head -> origin/gh/jansel/556/head 2025-12-04T08:57:03.7399727Z * [new branch] gh/jansel/556/orig -> origin/gh/jansel/556/orig 2025-12-04T08:57:03.7401976Z * [new branch] gh/jansel/557/base -> origin/gh/jansel/557/base 2025-12-04T08:57:03.7403429Z * [new branch] gh/jansel/557/head -> origin/gh/jansel/557/head 2025-12-04T08:57:03.7405283Z * [new branch] gh/jansel/557/orig -> origin/gh/jansel/557/orig 2025-12-04T08:57:03.7407469Z * [new branch] gh/jansel/558/base -> origin/gh/jansel/558/base 2025-12-04T08:57:03.7409114Z * [new branch] gh/jansel/558/head -> origin/gh/jansel/558/head 2025-12-04T08:57:03.7410660Z * [new branch] gh/jansel/558/orig -> origin/gh/jansel/558/orig 2025-12-04T08:57:03.7413261Z * [new branch] gh/jansel/559/base -> origin/gh/jansel/559/base 2025-12-04T08:57:03.7414818Z * [new branch] gh/jansel/559/head -> origin/gh/jansel/559/head 2025-12-04T08:57:03.7416397Z * [new branch] gh/jansel/559/orig -> origin/gh/jansel/559/orig 2025-12-04T08:57:03.7418430Z * [new branch] gh/jansel/560/base -> origin/gh/jansel/560/base 2025-12-04T08:57:03.7419924Z * [new branch] gh/jansel/560/head -> origin/gh/jansel/560/head 2025-12-04T08:57:03.7421467Z * [new branch] gh/jansel/560/orig -> origin/gh/jansel/560/orig 2025-12-04T08:57:03.7423623Z * [new branch] gh/jansel/561/base -> origin/gh/jansel/561/base 2025-12-04T08:57:03.7425306Z * [new branch] gh/jansel/561/head -> origin/gh/jansel/561/head 2025-12-04T08:57:03.7426634Z * [new branch] gh/jansel/561/orig -> origin/gh/jansel/561/orig 2025-12-04T08:57:03.7428685Z * [new branch] gh/jansel/562/base -> origin/gh/jansel/562/base 2025-12-04T08:57:03.7430407Z * [new branch] gh/jansel/562/head -> origin/gh/jansel/562/head 2025-12-04T08:57:03.7431949Z * [new branch] gh/jansel/562/orig -> origin/gh/jansel/562/orig 2025-12-04T08:57:03.7434041Z * [new branch] gh/jansel/563/base -> origin/gh/jansel/563/base 2025-12-04T08:57:03.7435735Z * [new branch] gh/jansel/563/head -> origin/gh/jansel/563/head 2025-12-04T08:57:03.7437255Z * [new branch] gh/jansel/563/orig -> origin/gh/jansel/563/orig 2025-12-04T08:57:03.7439767Z * [new branch] gh/jansel/564/base -> origin/gh/jansel/564/base 2025-12-04T08:57:03.7441337Z * [new branch] gh/jansel/564/head -> origin/gh/jansel/564/head 2025-12-04T08:57:03.7442915Z * [new branch] gh/jansel/564/orig -> origin/gh/jansel/564/orig 2025-12-04T08:57:03.7444985Z * [new branch] gh/jansel/565/base -> origin/gh/jansel/565/base 2025-12-04T08:57:03.7446506Z * [new branch] gh/jansel/565/head -> origin/gh/jansel/565/head 2025-12-04T08:57:03.7448058Z * [new branch] gh/jansel/565/orig -> origin/gh/jansel/565/orig 2025-12-04T08:57:03.7450246Z * [new branch] gh/jansel/566/base -> origin/gh/jansel/566/base 2025-12-04T08:57:03.7451819Z * [new branch] gh/jansel/566/head -> origin/gh/jansel/566/head 2025-12-04T08:57:03.7453445Z * [new branch] gh/jansel/566/orig -> origin/gh/jansel/566/orig 2025-12-04T08:57:03.7456137Z * [new branch] gh/jansel/567/base -> origin/gh/jansel/567/base 2025-12-04T08:57:03.7457699Z * [new branch] gh/jansel/567/head -> origin/gh/jansel/567/head 2025-12-04T08:57:03.7459400Z * [new branch] gh/jansel/567/orig -> origin/gh/jansel/567/orig 2025-12-04T08:57:03.7461509Z * [new branch] gh/jansel/568/base -> origin/gh/jansel/568/base 2025-12-04T08:57:03.7463106Z * [new branch] gh/jansel/568/head -> origin/gh/jansel/568/head 2025-12-04T08:57:03.7464669Z * [new branch] gh/jansel/568/orig -> origin/gh/jansel/568/orig 2025-12-04T08:57:03.7466798Z * [new branch] gh/jansel/569/base -> origin/gh/jansel/569/base 2025-12-04T08:57:03.7468309Z * [new branch] gh/jansel/569/head -> origin/gh/jansel/569/head 2025-12-04T08:57:03.7469879Z * [new branch] gh/jansel/569/orig -> origin/gh/jansel/569/orig 2025-12-04T08:57:03.7471979Z * [new branch] gh/jansel/570/base -> origin/gh/jansel/570/base 2025-12-04T08:57:03.7473662Z * [new branch] gh/jansel/570/head -> origin/gh/jansel/570/head 2025-12-04T08:57:03.7475355Z * [new branch] gh/jansel/570/orig -> origin/gh/jansel/570/orig 2025-12-04T08:57:03.7477477Z * [new branch] gh/jansel/571/base -> origin/gh/jansel/571/base 2025-12-04T08:57:03.7478979Z * [new branch] gh/jansel/571/head -> origin/gh/jansel/571/head 2025-12-04T08:57:03.7480486Z * [new branch] gh/jansel/571/orig -> origin/gh/jansel/571/orig 2025-12-04T08:57:03.7482542Z * [new branch] gh/jansel/572/base -> origin/gh/jansel/572/base 2025-12-04T08:57:03.7484124Z * [new branch] gh/jansel/572/head -> origin/gh/jansel/572/head 2025-12-04T08:57:03.7485642Z * [new branch] gh/jansel/572/orig -> origin/gh/jansel/572/orig 2025-12-04T08:57:03.7487878Z * [new branch] gh/jansel/573/base -> origin/gh/jansel/573/base 2025-12-04T08:57:03.7489474Z * [new branch] gh/jansel/573/head -> origin/gh/jansel/573/head 2025-12-04T08:57:03.7491063Z * [new branch] gh/jansel/573/orig -> origin/gh/jansel/573/orig 2025-12-04T08:57:03.7493259Z * [new branch] gh/jansel/574/base -> origin/gh/jansel/574/base 2025-12-04T08:57:03.7494959Z * [new branch] gh/jansel/574/head -> origin/gh/jansel/574/head 2025-12-04T08:57:03.7496471Z * [new branch] gh/jansel/574/orig -> origin/gh/jansel/574/orig 2025-12-04T08:57:03.7499165Z * [new branch] gh/jansel/575/base -> origin/gh/jansel/575/base 2025-12-04T08:57:03.7500720Z * [new branch] gh/jansel/575/head -> origin/gh/jansel/575/head 2025-12-04T08:57:03.7502311Z * [new branch] gh/jansel/575/orig -> origin/gh/jansel/575/orig 2025-12-04T08:57:03.7504664Z * [new branch] gh/jansel/576/base -> origin/gh/jansel/576/base 2025-12-04T08:57:03.7508059Z * [new branch] gh/jansel/576/head -> origin/gh/jansel/576/head 2025-12-04T08:57:03.7509590Z * [new branch] gh/jansel/576/orig -> origin/gh/jansel/576/orig 2025-12-04T08:57:03.7512162Z * [new branch] gh/jbschlosser/247/base -> origin/gh/jbschlosser/247/base 2025-12-04T08:57:03.7513733Z * [new branch] gh/jbschlosser/247/head -> origin/gh/jbschlosser/247/head 2025-12-04T08:57:03.7515365Z * [new branch] gh/jbschlosser/247/orig -> origin/gh/jbschlosser/247/orig 2025-12-04T08:57:03.7517616Z * [new branch] gh/jbschlosser/250/base -> origin/gh/jbschlosser/250/base 2025-12-04T08:57:03.7519109Z * [new branch] gh/jbschlosser/250/head -> origin/gh/jbschlosser/250/head 2025-12-04T08:57:03.7521209Z * [new branch] gh/jbschlosser/250/orig -> origin/gh/jbschlosser/250/orig 2025-12-04T08:57:03.7523796Z * [new branch] gh/jerryzh168/1/base -> origin/gh/jerryzh168/1/base 2025-12-04T08:57:03.7525417Z * [new branch] gh/jerryzh168/1/head -> origin/gh/jerryzh168/1/head 2025-12-04T08:57:03.7526852Z * [new branch] gh/jerryzh168/1/orig -> origin/gh/jerryzh168/1/orig 2025-12-04T08:57:03.7529671Z * [new branch] gh/jiayisunx/59/base -> origin/gh/jiayisunx/59/base 2025-12-04T08:57:03.7531269Z * [new branch] gh/jiayisunx/59/head -> origin/gh/jiayisunx/59/head 2025-12-04T08:57:03.7532877Z * [new branch] gh/jiayisunx/59/orig -> origin/gh/jiayisunx/59/orig 2025-12-04T08:57:03.7535342Z * [new branch] gh/jiayisunx/61/base -> origin/gh/jiayisunx/61/base 2025-12-04T08:57:03.7536936Z * [new branch] gh/jiayisunx/61/head -> origin/gh/jiayisunx/61/head 2025-12-04T08:57:03.7538524Z * [new branch] gh/jiayisunx/61/orig -> origin/gh/jiayisunx/61/orig 2025-12-04T08:57:03.7540574Z * [new branch] gh/jiayisunx/68/base -> origin/gh/jiayisunx/68/base 2025-12-04T08:57:03.7542101Z * [new branch] gh/jiayisunx/68/head -> origin/gh/jiayisunx/68/head 2025-12-04T08:57:03.7543845Z * [new branch] gh/jiayisunx/68/orig -> origin/gh/jiayisunx/68/orig 2025-12-04T08:57:03.7546165Z * [new branch] gh/jiayisunx/77/base -> origin/gh/jiayisunx/77/base 2025-12-04T08:57:03.7547785Z * [new branch] gh/jiayisunx/77/head -> origin/gh/jiayisunx/77/head 2025-12-04T08:57:03.7549311Z * [new branch] gh/jiayisunx/77/orig -> origin/gh/jiayisunx/77/orig 2025-12-04T08:57:03.7551604Z * [new branch] gh/jiayisunx/78/base -> origin/gh/jiayisunx/78/base 2025-12-04T08:57:03.7553199Z * [new branch] gh/jiayisunx/78/head -> origin/gh/jiayisunx/78/head 2025-12-04T08:57:03.7554696Z * [new branch] gh/jiayisunx/78/orig -> origin/gh/jiayisunx/78/orig 2025-12-04T08:57:03.7556760Z * [new branch] gh/jiayisunx/79/base -> origin/gh/jiayisunx/79/base 2025-12-04T08:57:03.7558309Z * [new branch] gh/jiayisunx/79/head -> origin/gh/jiayisunx/79/head 2025-12-04T08:57:03.7559950Z * [new branch] gh/jiayisunx/79/orig -> origin/gh/jiayisunx/79/orig 2025-12-04T08:57:03.7562046Z * [new branch] gh/jiayisunx/82/base -> origin/gh/jiayisunx/82/base 2025-12-04T08:57:03.7563608Z * [new branch] gh/jiayisunx/82/head -> origin/gh/jiayisunx/82/head 2025-12-04T08:57:03.7565193Z * [new branch] gh/jiayisunx/82/orig -> origin/gh/jiayisunx/82/orig 2025-12-04T08:57:03.7567224Z * [new branch] gh/jiayisunx/83/base -> origin/gh/jiayisunx/83/base 2025-12-04T08:57:03.7568834Z * [new branch] gh/jiayisunx/83/head -> origin/gh/jiayisunx/83/head 2025-12-04T08:57:03.7570374Z * [new branch] gh/jiayisunx/83/orig -> origin/gh/jiayisunx/83/orig 2025-12-04T08:57:03.7572411Z * [new branch] gh/jiayisunx/84/base -> origin/gh/jiayisunx/84/base 2025-12-04T08:57:03.7574122Z * [new branch] gh/jiayisunx/84/head -> origin/gh/jiayisunx/84/head 2025-12-04T08:57:03.7575643Z * [new branch] gh/jiayisunx/84/orig -> origin/gh/jiayisunx/84/orig 2025-12-04T08:57:03.7577711Z * [new branch] gh/jiayisunx/85/base -> origin/gh/jiayisunx/85/base 2025-12-04T08:57:03.7579287Z * [new branch] gh/jiayisunx/85/head -> origin/gh/jiayisunx/85/head 2025-12-04T08:57:03.7580890Z * [new branch] gh/jiayisunx/85/orig -> origin/gh/jiayisunx/85/orig 2025-12-04T08:57:03.7582876Z * [new branch] gh/jiayisunx/86/base -> origin/gh/jiayisunx/86/base 2025-12-04T08:57:03.7584459Z * [new branch] gh/jiayisunx/86/head -> origin/gh/jiayisunx/86/head 2025-12-04T08:57:03.7586029Z * [new branch] gh/jiayisunx/86/orig -> origin/gh/jiayisunx/86/orig 2025-12-04T08:57:03.7588231Z * [new branch] gh/jiayisunx/87/base -> origin/gh/jiayisunx/87/base 2025-12-04T08:57:03.7589688Z * [new branch] gh/jiayisunx/87/head -> origin/gh/jiayisunx/87/head 2025-12-04T08:57:03.7591280Z * [new branch] gh/jiayisunx/87/orig -> origin/gh/jiayisunx/87/orig 2025-12-04T08:57:03.7593396Z * [new branch] gh/jiayisunx/88/base -> origin/gh/jiayisunx/88/base 2025-12-04T08:57:03.7594900Z * [new branch] gh/jiayisunx/88/head -> origin/gh/jiayisunx/88/head 2025-12-04T08:57:03.7596517Z * [new branch] gh/jiayisunx/88/orig -> origin/gh/jiayisunx/88/orig 2025-12-04T08:57:03.7598703Z * [new branch] gh/jiayisunx/89/base -> origin/gh/jiayisunx/89/base 2025-12-04T08:57:03.7600291Z * [new branch] gh/jiayisunx/89/head -> origin/gh/jiayisunx/89/head 2025-12-04T08:57:03.7601801Z * [new branch] gh/jiayisunx/89/orig -> origin/gh/jiayisunx/89/orig 2025-12-04T08:57:03.7603830Z * [new branch] gh/jiayisunx/90/base -> origin/gh/jiayisunx/90/base 2025-12-04T08:57:03.7605687Z * [new branch] gh/jiayisunx/90/head -> origin/gh/jiayisunx/90/head 2025-12-04T08:57:03.7607291Z * [new branch] gh/jiayisunx/90/orig -> origin/gh/jiayisunx/90/orig 2025-12-04T08:57:03.7609719Z * [new branch] gh/jjwu@meta.com/1/base -> origin/gh/jjwu@meta.com/1/base 2025-12-04T08:57:03.7611235Z * [new branch] gh/jjwu@meta.com/1/head -> origin/gh/jjwu@meta.com/1/head 2025-12-04T08:57:03.7625739Z * [new branch] gh/jturney/1/base -> origin/gh/jturney/1/base 2025-12-04T08:57:03.7626183Z * [new branch] gh/jturney/1/head -> origin/gh/jturney/1/head 2025-12-04T08:57:03.7626385Z * [new branch] gh/jturney/1/orig -> origin/gh/jturney/1/orig 2025-12-04T08:57:03.7626584Z * [new branch] gh/jturney/2/base -> origin/gh/jturney/2/base 2025-12-04T08:57:03.7626783Z * [new branch] gh/jturney/2/head -> origin/gh/jturney/2/head 2025-12-04T08:57:03.7626970Z * [new branch] gh/jturney/2/orig -> origin/gh/jturney/2/orig 2025-12-04T08:57:03.7627173Z * [new branch] gh/karthickai/10/base -> origin/gh/karthickai/10/base 2025-12-04T08:57:03.7627363Z * [new branch] gh/karthickai/10/head -> origin/gh/karthickai/10/head 2025-12-04T08:57:03.7627746Z * [new branch] gh/karthickai/10/orig -> origin/gh/karthickai/10/orig 2025-12-04T08:57:03.7629999Z * [new branch] gh/karthickai/11/base -> origin/gh/karthickai/11/base 2025-12-04T08:57:03.7631607Z * [new branch] gh/karthickai/11/head -> origin/gh/karthickai/11/head 2025-12-04T08:57:03.7633172Z * [new branch] gh/karthickai/11/orig -> origin/gh/karthickai/11/orig 2025-12-04T08:57:03.7635501Z * [new branch] gh/karthickai/12/base -> origin/gh/karthickai/12/base 2025-12-04T08:57:03.7637161Z * [new branch] gh/karthickai/12/head -> origin/gh/karthickai/12/head 2025-12-04T08:57:03.7638748Z * [new branch] gh/karthickai/12/orig -> origin/gh/karthickai/12/orig 2025-12-04T08:57:03.7640896Z * [new branch] gh/karthickai/13/base -> origin/gh/karthickai/13/base 2025-12-04T08:57:03.7642471Z * [new branch] gh/karthickai/13/head -> origin/gh/karthickai/13/head 2025-12-04T08:57:03.7644026Z * [new branch] gh/karthickai/13/orig -> origin/gh/karthickai/13/orig 2025-12-04T08:57:03.7646196Z * [new branch] gh/karthickai/14/base -> origin/gh/karthickai/14/base 2025-12-04T08:57:03.7648278Z * [new branch] gh/karthickai/14/head -> origin/gh/karthickai/14/head 2025-12-04T08:57:03.7649845Z * [new branch] gh/karthickai/14/orig -> origin/gh/karthickai/14/orig 2025-12-04T08:57:03.7652364Z * [new branch] gh/karthickai/15/base -> origin/gh/karthickai/15/base 2025-12-04T08:57:03.7654201Z * [new branch] gh/karthickai/15/head -> origin/gh/karthickai/15/head 2025-12-04T08:57:03.7655731Z * [new branch] gh/karthickai/15/orig -> origin/gh/karthickai/15/orig 2025-12-04T08:57:03.7657695Z * [new branch] gh/karthickai/16/base -> origin/gh/karthickai/16/base 2025-12-04T08:57:03.7659394Z * [new branch] gh/karthickai/16/head -> origin/gh/karthickai/16/head 2025-12-04T08:57:03.7660931Z * [new branch] gh/karthickai/16/orig -> origin/gh/karthickai/16/orig 2025-12-04T08:57:03.7663019Z * [new branch] gh/karthickai/17/base -> origin/gh/karthickai/17/base 2025-12-04T08:57:03.7664398Z * [new branch] gh/karthickai/17/head -> origin/gh/karthickai/17/head 2025-12-04T08:57:03.7666400Z * [new branch] gh/karthickai/17/orig -> origin/gh/karthickai/17/orig 2025-12-04T08:57:03.7668548Z * [new branch] gh/karthickai/18/base -> origin/gh/karthickai/18/base 2025-12-04T08:57:03.7670299Z * [new branch] gh/karthickai/18/head -> origin/gh/karthickai/18/head 2025-12-04T08:57:03.7671980Z * [new branch] gh/karthickai/18/orig -> origin/gh/karthickai/18/orig 2025-12-04T08:57:03.7674332Z * [new branch] gh/karthickai/19/base -> origin/gh/karthickai/19/base 2025-12-04T08:57:03.7675891Z * [new branch] gh/karthickai/19/head -> origin/gh/karthickai/19/head 2025-12-04T08:57:03.7678019Z * [new branch] gh/karthickai/19/orig -> origin/gh/karthickai/19/orig 2025-12-04T08:57:03.7680704Z * [new branch] gh/karthickai/20/base -> origin/gh/karthickai/20/base 2025-12-04T08:57:03.7682776Z * [new branch] gh/karthickai/20/head -> origin/gh/karthickai/20/head 2025-12-04T08:57:03.7684381Z * [new branch] gh/karthickai/20/orig -> origin/gh/karthickai/20/orig 2025-12-04T08:57:03.7686632Z * [new branch] gh/karthickai/21/base -> origin/gh/karthickai/21/base 2025-12-04T08:57:03.7688401Z * [new branch] gh/karthickai/21/head -> origin/gh/karthickai/21/head 2025-12-04T08:57:03.7689961Z * [new branch] gh/karthickai/21/orig -> origin/gh/karthickai/21/orig 2025-12-04T08:57:03.7692207Z * [new branch] gh/karthickai/22/base -> origin/gh/karthickai/22/base 2025-12-04T08:57:03.7693760Z * [new branch] gh/karthickai/22/head -> origin/gh/karthickai/22/head 2025-12-04T08:57:03.7695320Z * [new branch] gh/karthickai/22/orig -> origin/gh/karthickai/22/orig 2025-12-04T08:57:03.7697508Z * [new branch] gh/karthickai/23/base -> origin/gh/karthickai/23/base 2025-12-04T08:57:03.7699164Z * [new branch] gh/karthickai/23/head -> origin/gh/karthickai/23/head 2025-12-04T08:57:03.7700691Z * [new branch] gh/karthickai/23/orig -> origin/gh/karthickai/23/orig 2025-12-04T08:57:03.7702849Z * [new branch] gh/karthickai/24/base -> origin/gh/karthickai/24/base 2025-12-04T08:57:03.7704448Z * [new branch] gh/karthickai/24/head -> origin/gh/karthickai/24/head 2025-12-04T08:57:03.7706368Z * [new branch] gh/karthickai/24/orig -> origin/gh/karthickai/24/orig 2025-12-04T08:57:03.7708948Z * [new branch] gh/karthickai/25/base -> origin/gh/karthickai/25/base 2025-12-04T08:57:03.7710560Z * [new branch] gh/karthickai/25/head -> origin/gh/karthickai/25/head 2025-12-04T08:57:03.7712085Z * [new branch] gh/karthickai/25/orig -> origin/gh/karthickai/25/orig 2025-12-04T08:57:03.7714093Z * [new branch] gh/karthickai/26/base -> origin/gh/karthickai/26/base 2025-12-04T08:57:03.7715929Z * [new branch] gh/karthickai/26/head -> origin/gh/karthickai/26/head 2025-12-04T08:57:03.7717649Z * [new branch] gh/karthickai/26/orig -> origin/gh/karthickai/26/orig 2025-12-04T08:57:03.7720674Z * [new branch] gh/karthickai/6/base -> origin/gh/karthickai/6/base 2025-12-04T08:57:03.7722639Z * [new branch] gh/karthickai/6/head -> origin/gh/karthickai/6/head 2025-12-04T08:57:03.7724479Z * [new branch] gh/karthickai/6/orig -> origin/gh/karthickai/6/orig 2025-12-04T08:57:03.7727096Z * [new branch] gh/krocki/1/base -> origin/gh/krocki/1/base 2025-12-04T08:57:03.7728654Z * [new branch] gh/krocki/1/head -> origin/gh/krocki/1/head 2025-12-04T08:57:03.7730628Z * [new branch] gh/krocki/1/orig -> origin/gh/krocki/1/orig 2025-12-04T08:57:03.7732819Z * [new branch] gh/krocki/2/base -> origin/gh/krocki/2/base 2025-12-04T08:57:03.7734576Z * [new branch] gh/krocki/2/head -> origin/gh/krocki/2/head 2025-12-04T08:57:03.7736083Z * [new branch] gh/krocki/2/orig -> origin/gh/krocki/2/orig 2025-12-04T08:57:03.7738682Z * [new branch] gh/kurtamohler/60/base -> origin/gh/kurtamohler/60/base 2025-12-04T08:57:03.7740386Z * [new branch] gh/kurtamohler/60/head -> origin/gh/kurtamohler/60/head 2025-12-04T08:57:03.7741953Z * [new branch] gh/kurtamohler/60/orig -> origin/gh/kurtamohler/60/orig 2025-12-04T08:57:03.7744060Z * [new branch] gh/kurtamohler/61/base -> origin/gh/kurtamohler/61/base 2025-12-04T08:57:03.7745621Z * [new branch] gh/kurtamohler/61/head -> origin/gh/kurtamohler/61/head 2025-12-04T08:57:03.7747157Z * [new branch] gh/kurtamohler/61/orig -> origin/gh/kurtamohler/61/orig 2025-12-04T08:57:03.7749219Z * [new branch] gh/kurtamohler/62/base -> origin/gh/kurtamohler/62/base 2025-12-04T08:57:03.7750743Z * [new branch] gh/kurtamohler/62/head -> origin/gh/kurtamohler/62/head 2025-12-04T08:57:03.7752353Z * [new branch] gh/kurtamohler/62/orig -> origin/gh/kurtamohler/62/orig 2025-12-04T08:57:03.7754384Z * [new branch] gh/kurtamohler/63/base -> origin/gh/kurtamohler/63/base 2025-12-04T08:57:03.7756012Z * [new branch] gh/kurtamohler/63/head -> origin/gh/kurtamohler/63/head 2025-12-04T08:57:03.7757581Z * [new branch] gh/kurtamohler/63/orig -> origin/gh/kurtamohler/63/orig 2025-12-04T08:57:03.7759609Z * [new branch] gh/kurtamohler/64/base -> origin/gh/kurtamohler/64/base 2025-12-04T08:57:03.7761175Z * [new branch] gh/kurtamohler/64/head -> origin/gh/kurtamohler/64/head 2025-12-04T08:57:03.7762662Z * [new branch] gh/kurtamohler/64/orig -> origin/gh/kurtamohler/64/orig 2025-12-04T08:57:03.7764776Z * [new branch] gh/kurtamohler/65/base -> origin/gh/kurtamohler/65/base 2025-12-04T08:57:03.7766347Z * [new branch] gh/kurtamohler/65/head -> origin/gh/kurtamohler/65/head 2025-12-04T08:57:03.7767920Z * [new branch] gh/kurtamohler/65/orig -> origin/gh/kurtamohler/65/orig 2025-12-04T08:57:03.7769887Z * [new branch] gh/kurtamohler/66/base -> origin/gh/kurtamohler/66/base 2025-12-04T08:57:03.7771466Z * [new branch] gh/kurtamohler/66/head -> origin/gh/kurtamohler/66/head 2025-12-04T08:57:03.7773078Z * [new branch] gh/kurtamohler/66/orig -> origin/gh/kurtamohler/66/orig 2025-12-04T08:57:03.7775419Z * [new branch] gh/kurtamohler/67/base -> origin/gh/kurtamohler/67/base 2025-12-04T08:57:03.7776973Z * [new branch] gh/kurtamohler/67/head -> origin/gh/kurtamohler/67/head 2025-12-04T08:57:03.7778489Z * [new branch] gh/kurtamohler/67/orig -> origin/gh/kurtamohler/67/orig 2025-12-04T08:57:03.7781054Z * [new branch] gh/kwen2501/130/base -> origin/gh/kwen2501/130/base 2025-12-04T08:57:03.7782859Z * [new branch] gh/kwen2501/130/head -> origin/gh/kwen2501/130/head 2025-12-04T08:57:03.7784384Z * [new branch] gh/kwen2501/130/orig -> origin/gh/kwen2501/130/orig 2025-12-04T08:57:03.7786501Z * [new branch] gh/kwen2501/170/base -> origin/gh/kwen2501/170/base 2025-12-04T08:57:03.7788037Z * [new branch] gh/kwen2501/170/head -> origin/gh/kwen2501/170/head 2025-12-04T08:57:03.7790296Z * [new branch] gh/kwen2501/187/base -> origin/gh/kwen2501/187/base 2025-12-04T08:57:03.7791878Z * [new branch] gh/kwen2501/187/head -> origin/gh/kwen2501/187/head 2025-12-04T08:57:03.7793547Z * [new branch] gh/kwen2501/187/orig -> origin/gh/kwen2501/187/orig 2025-12-04T08:57:03.7795737Z * [new branch] gh/kwen2501/188/base -> origin/gh/kwen2501/188/base 2025-12-04T08:57:03.7797273Z * [new branch] gh/kwen2501/188/head -> origin/gh/kwen2501/188/head 2025-12-04T08:57:03.7798814Z * [new branch] gh/kwen2501/188/orig -> origin/gh/kwen2501/188/orig 2025-12-04T08:57:03.7800846Z * [new branch] gh/kwen2501/211/base -> origin/gh/kwen2501/211/base 2025-12-04T08:57:03.7802401Z * [new branch] gh/kwen2501/211/head -> origin/gh/kwen2501/211/head 2025-12-04T08:57:03.7804391Z * [new branch] gh/kwen2501/224/base -> origin/gh/kwen2501/224/base 2025-12-04T08:57:03.7806300Z * [new branch] gh/kwen2501/224/head -> origin/gh/kwen2501/224/head 2025-12-04T08:57:03.7807856Z * [new branch] gh/kwen2501/224/orig -> origin/gh/kwen2501/224/orig 2025-12-04T08:57:03.7809933Z * [new branch] gh/kwen2501/228/base -> origin/gh/kwen2501/228/base 2025-12-04T08:57:03.7811478Z * [new branch] gh/kwen2501/228/head -> origin/gh/kwen2501/228/head 2025-12-04T08:57:03.7813016Z * [new branch] gh/kwen2501/228/orig -> origin/gh/kwen2501/228/orig 2025-12-04T08:57:03.7815465Z * [new branch] gh/kwen2501/234/base -> origin/gh/kwen2501/234/base 2025-12-04T08:57:03.7817024Z * [new branch] gh/kwen2501/234/head -> origin/gh/kwen2501/234/head 2025-12-04T08:57:03.7818575Z * [new branch] gh/kwen2501/234/orig -> origin/gh/kwen2501/234/orig 2025-12-04T08:57:03.7820637Z * [new branch] gh/kwen2501/235/base -> origin/gh/kwen2501/235/base 2025-12-04T08:57:03.7822180Z * [new branch] gh/kwen2501/235/head -> origin/gh/kwen2501/235/head 2025-12-04T08:57:03.7823654Z * [new branch] gh/kwen2501/235/orig -> origin/gh/kwen2501/235/orig 2025-12-04T08:57:03.7825873Z * [new branch] gh/kwen2501/236/base -> origin/gh/kwen2501/236/base 2025-12-04T08:57:03.7827405Z * [new branch] gh/kwen2501/236/head -> origin/gh/kwen2501/236/head 2025-12-04T08:57:03.7828952Z * [new branch] gh/kwen2501/236/orig -> origin/gh/kwen2501/236/orig 2025-12-04T08:57:03.7830973Z * [new branch] gh/kwen2501/237/base -> origin/gh/kwen2501/237/base 2025-12-04T08:57:03.7832565Z * [new branch] gh/kwen2501/237/head -> origin/gh/kwen2501/237/head 2025-12-04T08:57:03.7833997Z * [new branch] gh/kwen2501/237/orig -> origin/gh/kwen2501/237/orig 2025-12-04T08:57:03.7836141Z * [new branch] gh/kwen2501/238/base -> origin/gh/kwen2501/238/base 2025-12-04T08:57:03.7837745Z * [new branch] gh/kwen2501/238/head -> origin/gh/kwen2501/238/head 2025-12-04T08:57:03.7839259Z * [new branch] gh/kwen2501/238/orig -> origin/gh/kwen2501/238/orig 2025-12-04T08:57:03.7841896Z * [new branch] gh/kwen2501/240/base -> origin/gh/kwen2501/240/base 2025-12-04T08:57:03.7843514Z * [new branch] gh/kwen2501/240/head -> origin/gh/kwen2501/240/head 2025-12-04T08:57:03.7845236Z * [new branch] gh/kwen2501/240/orig -> origin/gh/kwen2501/240/orig 2025-12-04T08:57:03.7847366Z * [new branch] gh/kwen2501/241/base -> origin/gh/kwen2501/241/base 2025-12-04T08:57:03.7848683Z * [new branch] gh/kwen2501/241/head -> origin/gh/kwen2501/241/head 2025-12-04T08:57:03.7850188Z * [new branch] gh/kwen2501/241/orig -> origin/gh/kwen2501/241/orig 2025-12-04T08:57:03.7852277Z * [new branch] gh/kwen2501/247/base -> origin/gh/kwen2501/247/base 2025-12-04T08:57:03.7853875Z * [new branch] gh/kwen2501/247/head -> origin/gh/kwen2501/247/head 2025-12-04T08:57:03.7855439Z * [new branch] gh/kwen2501/247/orig -> origin/gh/kwen2501/247/orig 2025-12-04T08:57:03.7857578Z * [new branch] gh/kwen2501/252/base -> origin/gh/kwen2501/252/base 2025-12-04T08:57:03.7859143Z * [new branch] gh/kwen2501/252/head -> origin/gh/kwen2501/252/head 2025-12-04T08:57:03.7860665Z * [new branch] gh/kwen2501/252/orig -> origin/gh/kwen2501/252/orig 2025-12-04T08:57:03.7863257Z * [new branch] gh/kwen2501/259/base -> origin/gh/kwen2501/259/base 2025-12-04T08:57:03.7864992Z * [new branch] gh/kwen2501/259/head -> origin/gh/kwen2501/259/head 2025-12-04T08:57:03.7866540Z * [new branch] gh/kwen2501/259/orig -> origin/gh/kwen2501/259/orig 2025-12-04T08:57:03.7868742Z * [new branch] gh/kwen2501/260/base -> origin/gh/kwen2501/260/base 2025-12-04T08:57:03.7870389Z * [new branch] gh/kwen2501/260/head -> origin/gh/kwen2501/260/head 2025-12-04T08:57:03.7871990Z * [new branch] gh/kwen2501/260/orig -> origin/gh/kwen2501/260/orig 2025-12-04T08:57:03.7874043Z * [new branch] gh/kwen2501/268/base -> origin/gh/kwen2501/268/base 2025-12-04T08:57:03.7875628Z * [new branch] gh/kwen2501/268/head -> origin/gh/kwen2501/268/head 2025-12-04T08:57:03.7877217Z * [new branch] gh/kwen2501/268/orig -> origin/gh/kwen2501/268/orig 2025-12-04T08:57:03.7879493Z * [new branch] gh/kwen2501/269/base -> origin/gh/kwen2501/269/base 2025-12-04T08:57:03.7881139Z * [new branch] gh/kwen2501/269/head -> origin/gh/kwen2501/269/head 2025-12-04T08:57:03.7882653Z * [new branch] gh/kwen2501/269/orig -> origin/gh/kwen2501/269/orig 2025-12-04T08:57:03.7885200Z * [new branch] gh/kwen2501/270/base -> origin/gh/kwen2501/270/base 2025-12-04T08:57:03.7886853Z * [new branch] gh/kwen2501/270/head -> origin/gh/kwen2501/270/head 2025-12-04T08:57:03.7888349Z * [new branch] gh/kwen2501/270/orig -> origin/gh/kwen2501/270/orig 2025-12-04T08:57:03.7890563Z * [new branch] gh/kwen2501/271/base -> origin/gh/kwen2501/271/base 2025-12-04T08:57:03.7892161Z * [new branch] gh/kwen2501/271/head -> origin/gh/kwen2501/271/head 2025-12-04T08:57:03.7894057Z * [new branch] gh/kwen2501/271/orig -> origin/gh/kwen2501/271/orig 2025-12-04T08:57:03.7896926Z * [new branch] gh/kwen2501/274/base -> origin/gh/kwen2501/274/base 2025-12-04T08:57:03.7898568Z * [new branch] gh/kwen2501/274/head -> origin/gh/kwen2501/274/head 2025-12-04T08:57:03.7900145Z * [new branch] gh/kwen2501/274/orig -> origin/gh/kwen2501/274/orig 2025-12-04T08:57:03.7902463Z * [new branch] gh/kwen2501/275/base -> origin/gh/kwen2501/275/base 2025-12-04T08:57:03.7904133Z * [new branch] gh/kwen2501/275/head -> origin/gh/kwen2501/275/head 2025-12-04T08:57:03.7907571Z * [new branch] gh/kwen2501/275/orig -> origin/gh/kwen2501/275/orig 2025-12-04T08:57:03.7909785Z * [new branch] gh/kwen2501/276/base -> origin/gh/kwen2501/276/base 2025-12-04T08:57:03.7911601Z * [new branch] gh/kwen2501/276/head -> origin/gh/kwen2501/276/head 2025-12-04T08:57:03.7912824Z * [new branch] gh/kwen2501/276/orig -> origin/gh/kwen2501/276/orig 2025-12-04T08:57:03.7915017Z * [new branch] gh/kwen2501/277/base -> origin/gh/kwen2501/277/base 2025-12-04T08:57:03.7916694Z * [new branch] gh/kwen2501/277/head -> origin/gh/kwen2501/277/head 2025-12-04T08:57:03.7918270Z * [new branch] gh/kwen2501/277/orig -> origin/gh/kwen2501/277/orig 2025-12-04T08:57:03.7920387Z * [new branch] gh/kwen2501/278/base -> origin/gh/kwen2501/278/base 2025-12-04T08:57:03.7921967Z * [new branch] gh/kwen2501/278/head -> origin/gh/kwen2501/278/head 2025-12-04T08:57:03.7923441Z * [new branch] gh/kwen2501/278/orig -> origin/gh/kwen2501/278/orig 2025-12-04T08:57:03.7925767Z * [new branch] gh/kwen2501/279/base -> origin/gh/kwen2501/279/base 2025-12-04T08:57:03.7927519Z * [new branch] gh/kwen2501/279/head -> origin/gh/kwen2501/279/head 2025-12-04T08:57:03.7929128Z * [new branch] gh/kwen2501/279/orig -> origin/gh/kwen2501/279/orig 2025-12-04T08:57:03.7931332Z * [new branch] gh/kwen2501/280/base -> origin/gh/kwen2501/280/base 2025-12-04T08:57:03.7932924Z * [new branch] gh/kwen2501/280/head -> origin/gh/kwen2501/280/head 2025-12-04T08:57:03.7934708Z * [new branch] gh/kwen2501/280/orig -> origin/gh/kwen2501/280/orig 2025-12-04T08:57:03.7936817Z * [new branch] gh/kwen2501/281/base -> origin/gh/kwen2501/281/base 2025-12-04T08:57:03.7938362Z * [new branch] gh/kwen2501/281/head -> origin/gh/kwen2501/281/head 2025-12-04T08:57:03.7939911Z * [new branch] gh/kwen2501/281/orig -> origin/gh/kwen2501/281/orig 2025-12-04T08:57:03.7942074Z * [new branch] gh/kwen2501/282/base -> origin/gh/kwen2501/282/base 2025-12-04T08:57:03.7943689Z * [new branch] gh/kwen2501/282/head -> origin/gh/kwen2501/282/head 2025-12-04T08:57:03.7945350Z * [new branch] gh/kwen2501/282/orig -> origin/gh/kwen2501/282/orig 2025-12-04T08:57:03.7947597Z * [new branch] gh/kwen2501/283/base -> origin/gh/kwen2501/283/base 2025-12-04T08:57:03.7949744Z * [new branch] gh/kwen2501/283/head -> origin/gh/kwen2501/283/head 2025-12-04T08:57:03.7951323Z * [new branch] gh/kwen2501/283/orig -> origin/gh/kwen2501/283/orig 2025-12-04T08:57:03.7953460Z * [new branch] gh/kwen2501/284/base -> origin/gh/kwen2501/284/base 2025-12-04T08:57:03.7955043Z * [new branch] gh/kwen2501/284/head -> origin/gh/kwen2501/284/head 2025-12-04T08:57:03.7956683Z * [new branch] gh/kwen2501/284/orig -> origin/gh/kwen2501/284/orig 2025-12-04T08:57:03.7958828Z * [new branch] gh/kwen2501/285/base -> origin/gh/kwen2501/285/base 2025-12-04T08:57:03.7960351Z * [new branch] gh/kwen2501/285/head -> origin/gh/kwen2501/285/head 2025-12-04T08:57:03.7961893Z * [new branch] gh/kwen2501/285/orig -> origin/gh/kwen2501/285/orig 2025-12-04T08:57:03.7963959Z * [new branch] gh/kwen2501/286/base -> origin/gh/kwen2501/286/base 2025-12-04T08:57:03.7966069Z * [new branch] gh/kwen2501/286/head -> origin/gh/kwen2501/286/head 2025-12-04T08:57:03.7967624Z * [new branch] gh/kwen2501/286/orig -> origin/gh/kwen2501/286/orig 2025-12-04T08:57:03.7969781Z * [new branch] gh/kwen2501/287/base -> origin/gh/kwen2501/287/base 2025-12-04T08:57:03.7971303Z * [new branch] gh/kwen2501/287/head -> origin/gh/kwen2501/287/head 2025-12-04T08:57:03.7972875Z * [new branch] gh/kwen2501/287/orig -> origin/gh/kwen2501/287/orig 2025-12-04T08:57:03.7975830Z * [new branch] gh/kwen2501/288/base -> origin/gh/kwen2501/288/base 2025-12-04T08:57:03.7977371Z * [new branch] gh/kwen2501/288/head -> origin/gh/kwen2501/288/head 2025-12-04T08:57:03.7978992Z * [new branch] gh/kwen2501/288/orig -> origin/gh/kwen2501/288/orig 2025-12-04T08:57:03.7981506Z * [new branch] gh/laithsakka/251/base -> origin/gh/laithsakka/251/base 2025-12-04T08:57:03.7983056Z * [new branch] gh/laithsakka/251/head -> origin/gh/laithsakka/251/head 2025-12-04T08:57:03.7984593Z * [new branch] gh/laithsakka/251/orig -> origin/gh/laithsakka/251/orig 2025-12-04T08:57:03.7986656Z * [new branch] gh/laithsakka/276/base -> origin/gh/laithsakka/276/base 2025-12-04T08:57:03.7988165Z * [new branch] gh/laithsakka/276/head -> origin/gh/laithsakka/276/head 2025-12-04T08:57:03.7989797Z * [new branch] gh/laithsakka/276/orig -> origin/gh/laithsakka/276/orig 2025-12-04T08:57:03.7992050Z * [new branch] gh/laithsakka/28/base -> origin/gh/laithsakka/28/base 2025-12-04T08:57:03.7994042Z * [new branch] gh/laithsakka/29/base -> origin/gh/laithsakka/29/base 2025-12-04T08:57:03.7995996Z * [new branch] gh/laithsakka/30/base -> origin/gh/laithsakka/30/base 2025-12-04T08:57:03.7997587Z * [new branch] gh/laithsakka/30/head -> origin/gh/laithsakka/30/head 2025-12-04T08:57:03.7999527Z * [new branch] gh/laithsakka/31/base -> origin/gh/laithsakka/31/base 2025-12-04T08:57:03.8001034Z * [new branch] gh/laithsakka/31/head -> origin/gh/laithsakka/31/head 2025-12-04T08:57:03.8003210Z * [new branch] gh/laithsakka/313/base -> origin/gh/laithsakka/313/base 2025-12-04T08:57:03.8005043Z * [new branch] gh/laithsakka/313/head -> origin/gh/laithsakka/313/head 2025-12-04T08:57:03.8006760Z * [new branch] gh/laithsakka/313/orig -> origin/gh/laithsakka/313/orig 2025-12-04T08:57:03.8009047Z * [new branch] gh/laithsakka/316/base -> origin/gh/laithsakka/316/base 2025-12-04T08:57:03.8010526Z * [new branch] gh/laithsakka/316/head -> origin/gh/laithsakka/316/head 2025-12-04T08:57:03.8012132Z * [new branch] gh/laithsakka/316/orig -> origin/gh/laithsakka/316/orig 2025-12-04T08:57:03.8014686Z * [new branch] gh/laithsakka/317/base -> origin/gh/laithsakka/317/base 2025-12-04T08:57:03.8016245Z * [new branch] gh/laithsakka/317/head -> origin/gh/laithsakka/317/head 2025-12-04T08:57:03.8017722Z * [new branch] gh/laithsakka/317/orig -> origin/gh/laithsakka/317/orig 2025-12-04T08:57:03.8019873Z * [new branch] gh/laithsakka/319/base -> origin/gh/laithsakka/319/base 2025-12-04T08:57:03.8021382Z * [new branch] gh/laithsakka/319/head -> origin/gh/laithsakka/319/head 2025-12-04T08:57:03.8022960Z * [new branch] gh/laithsakka/319/orig -> origin/gh/laithsakka/319/orig 2025-12-04T08:57:03.8024853Z * [new branch] gh/laithsakka/32/base -> origin/gh/laithsakka/32/base 2025-12-04T08:57:03.8026767Z * [new branch] gh/laithsakka/32/head -> origin/gh/laithsakka/32/head 2025-12-04T08:57:03.8028935Z * [new branch] gh/laithsakka/320/base -> origin/gh/laithsakka/320/base 2025-12-04T08:57:03.8030510Z * [new branch] gh/laithsakka/320/head -> origin/gh/laithsakka/320/head 2025-12-04T08:57:03.8032001Z * [new branch] gh/laithsakka/320/orig -> origin/gh/laithsakka/320/orig 2025-12-04T08:57:03.8034085Z * [new branch] gh/laithsakka/321/base -> origin/gh/laithsakka/321/base 2025-12-04T08:57:03.8035776Z * [new branch] gh/laithsakka/321/head -> origin/gh/laithsakka/321/head 2025-12-04T08:57:03.8037483Z * [new branch] gh/laithsakka/321/orig -> origin/gh/laithsakka/321/orig 2025-12-04T08:57:03.8039569Z * [new branch] gh/laithsakka/322/base -> origin/gh/laithsakka/322/base 2025-12-04T08:57:03.8041160Z * [new branch] gh/laithsakka/322/head -> origin/gh/laithsakka/322/head 2025-12-04T08:57:03.8042677Z * [new branch] gh/laithsakka/322/orig -> origin/gh/laithsakka/322/orig 2025-12-04T08:57:03.8045011Z * [new branch] gh/laithsakka/323/base -> origin/gh/laithsakka/323/base 2025-12-04T08:57:03.8046533Z * [new branch] gh/laithsakka/323/head -> origin/gh/laithsakka/323/head 2025-12-04T08:57:03.8048354Z * [new branch] gh/laithsakka/323/orig -> origin/gh/laithsakka/323/orig 2025-12-04T08:57:03.8050526Z * [new branch] gh/laithsakka/324/base -> origin/gh/laithsakka/324/base 2025-12-04T08:57:03.8052113Z * [new branch] gh/laithsakka/324/head -> origin/gh/laithsakka/324/head 2025-12-04T08:57:03.8053768Z * [new branch] gh/laithsakka/324/orig -> origin/gh/laithsakka/324/orig 2025-12-04T08:57:03.8056060Z * [new branch] gh/laithsakka/325/base -> origin/gh/laithsakka/325/base 2025-12-04T08:57:03.8057610Z * [new branch] gh/laithsakka/325/head -> origin/gh/laithsakka/325/head 2025-12-04T08:57:03.8059122Z * [new branch] gh/laithsakka/325/orig -> origin/gh/laithsakka/325/orig 2025-12-04T08:57:03.8061578Z * [new branch] gh/laithsakka/326/base -> origin/gh/laithsakka/326/base 2025-12-04T08:57:03.8063254Z * [new branch] gh/laithsakka/326/head -> origin/gh/laithsakka/326/head 2025-12-04T08:57:03.8065076Z * [new branch] gh/laithsakka/326/orig -> origin/gh/laithsakka/326/orig 2025-12-04T08:57:03.8067541Z * [new branch] gh/laithsakka/327/base -> origin/gh/laithsakka/327/base 2025-12-04T08:57:03.8069131Z * [new branch] gh/laithsakka/327/head -> origin/gh/laithsakka/327/head 2025-12-04T08:57:03.8070732Z * [new branch] gh/laithsakka/327/orig -> origin/gh/laithsakka/327/orig 2025-12-04T08:57:03.8073298Z * [new branch] gh/laithsakka/328/base -> origin/gh/laithsakka/328/base 2025-12-04T08:57:03.8074946Z * [new branch] gh/laithsakka/328/head -> origin/gh/laithsakka/328/head 2025-12-04T08:57:03.8076516Z * [new branch] gh/laithsakka/328/orig -> origin/gh/laithsakka/328/orig 2025-12-04T08:57:03.8078970Z * [new branch] gh/liangel/4/base -> origin/gh/liangel/4/base 2025-12-04T08:57:03.8080531Z * [new branch] gh/liangel/4/head -> origin/gh/liangel/4/head 2025-12-04T08:57:03.8082045Z * [new branch] gh/liangel/4/orig -> origin/gh/liangel/4/orig 2025-12-04T08:57:03.8086417Z * [new branch] gh/lucaskabela/1/base -> origin/gh/lucaskabela/1/base 2025-12-04T08:57:03.8088018Z * [new branch] gh/lucaskabela/1/head -> origin/gh/lucaskabela/1/head 2025-12-04T08:57:03.8090436Z * [new branch] gh/lw/4/base -> origin/gh/lw/4/base 2025-12-04T08:57:03.8091979Z * [new branch] gh/lw/4/head -> origin/gh/lw/4/head 2025-12-04T08:57:03.8093639Z * [new branch] gh/lw/4/orig -> origin/gh/lw/4/orig 2025-12-04T08:57:03.8095789Z * [new branch] gh/lw/5/base -> origin/gh/lw/5/base 2025-12-04T08:57:03.8097327Z * [new branch] gh/lw/5/head -> origin/gh/lw/5/head 2025-12-04T08:57:03.8098836Z * [new branch] gh/lw/5/orig -> origin/gh/lw/5/orig 2025-12-04T08:57:03.8100926Z * [new branch] gh/lw/6/base -> origin/gh/lw/6/base 2025-12-04T08:57:03.8102456Z * [new branch] gh/lw/6/head -> origin/gh/lw/6/head 2025-12-04T08:57:03.8103975Z * [new branch] gh/lw/6/orig -> origin/gh/lw/6/orig 2025-12-04T08:57:03.8107040Z * [new branch] gh/malfet/14/base -> origin/gh/malfet/14/base 2025-12-04T08:57:03.8109047Z * [new branch] gh/malfet/417/base -> origin/gh/malfet/417/base 2025-12-04T08:57:03.8110527Z * [new branch] gh/malfet/417/head -> origin/gh/malfet/417/head 2025-12-04T08:57:03.8112043Z * [new branch] gh/malfet/417/orig -> origin/gh/malfet/417/orig 2025-12-04T08:57:03.8114206Z * [new branch] gh/malfet/506/base -> origin/gh/malfet/506/base 2025-12-04T08:57:03.8115786Z * [new branch] gh/malfet/506/head -> origin/gh/malfet/506/head 2025-12-04T08:57:03.8117277Z * [new branch] gh/malfet/506/orig -> origin/gh/malfet/506/orig 2025-12-04T08:57:03.8119535Z * [new branch] gh/malfet/517/base -> origin/gh/malfet/517/base 2025-12-04T08:57:03.8121123Z * [new branch] gh/malfet/517/head -> origin/gh/malfet/517/head 2025-12-04T08:57:03.8123176Z * [new branch] gh/malfet/528/base -> origin/gh/malfet/528/base 2025-12-04T08:57:03.8124696Z * [new branch] gh/malfet/528/head -> origin/gh/malfet/528/head 2025-12-04T08:57:03.8126219Z * [new branch] gh/malfet/528/orig -> origin/gh/malfet/528/orig 2025-12-04T08:57:03.8128282Z * [new branch] gh/malfet/537/base -> origin/gh/malfet/537/base 2025-12-04T08:57:03.8129854Z * [new branch] gh/malfet/537/head -> origin/gh/malfet/537/head 2025-12-04T08:57:03.8131431Z * [new branch] gh/malfet/537/orig -> origin/gh/malfet/537/orig 2025-12-04T08:57:03.8133581Z * [new branch] gh/malfet/546/base -> origin/gh/malfet/546/base 2025-12-04T08:57:03.8135321Z * [new branch] gh/malfet/546/head -> origin/gh/malfet/546/head 2025-12-04T08:57:03.8136794Z * [new branch] gh/malfet/546/orig -> origin/gh/malfet/546/orig 2025-12-04T08:57:03.8138874Z * [new branch] gh/malfet/565/base -> origin/gh/malfet/565/base 2025-12-04T08:57:03.8140420Z * [new branch] gh/malfet/565/head -> origin/gh/malfet/565/head 2025-12-04T08:57:03.8141989Z * [new branch] gh/malfet/565/orig -> origin/gh/malfet/565/orig 2025-12-04T08:57:03.8144169Z * [new branch] gh/malfet/575/base -> origin/gh/malfet/575/base 2025-12-04T08:57:03.8145701Z * [new branch] gh/malfet/575/head -> origin/gh/malfet/575/head 2025-12-04T08:57:03.8147709Z * [new branch] gh/malfet/575/orig -> origin/gh/malfet/575/orig 2025-12-04T08:57:03.8149767Z * [new branch] gh/malfet/580/base -> origin/gh/malfet/580/base 2025-12-04T08:57:03.8151313Z * [new branch] gh/malfet/580/head -> origin/gh/malfet/580/head 2025-12-04T08:57:03.8152953Z * [new branch] gh/malfet/580/orig -> origin/gh/malfet/580/orig 2025-12-04T08:57:03.8155003Z * [new branch] gh/malfet/581/base -> origin/gh/malfet/581/base 2025-12-04T08:57:03.8156778Z * [new branch] gh/malfet/581/head -> origin/gh/malfet/581/head 2025-12-04T08:57:03.8158301Z * [new branch] gh/malfet/581/orig -> origin/gh/malfet/581/orig 2025-12-04T08:57:03.8160353Z * [new branch] gh/malfet/583/base -> origin/gh/malfet/583/base 2025-12-04T08:57:03.8161894Z * [new branch] gh/malfet/583/head -> origin/gh/malfet/583/head 2025-12-04T08:57:03.8163457Z * [new branch] gh/malfet/583/orig -> origin/gh/malfet/583/orig 2025-12-04T08:57:03.8165380Z * [new branch] gh/malfet/586/base -> origin/gh/malfet/586/base 2025-12-04T08:57:03.8166934Z * [new branch] gh/malfet/586/head -> origin/gh/malfet/586/head 2025-12-04T08:57:03.8168577Z * [new branch] gh/malfet/586/orig -> origin/gh/malfet/586/orig 2025-12-04T08:57:03.8170579Z * [new branch] gh/malfet/587/base -> origin/gh/malfet/587/base 2025-12-04T08:57:03.8172102Z * [new branch] gh/malfet/587/head -> origin/gh/malfet/587/head 2025-12-04T08:57:03.8173773Z * [new branch] gh/malfet/587/orig -> origin/gh/malfet/587/orig 2025-12-04T08:57:03.8175858Z * [new branch] gh/malfet/588/base -> origin/gh/malfet/588/base 2025-12-04T08:57:03.8177503Z * [new branch] gh/malfet/588/head -> origin/gh/malfet/588/head 2025-12-04T08:57:03.8179178Z * [new branch] gh/malfet/588/orig -> origin/gh/malfet/588/orig 2025-12-04T08:57:03.8181249Z * [new branch] gh/malfet/589/base -> origin/gh/malfet/589/base 2025-12-04T08:57:03.8182788Z * [new branch] gh/malfet/589/head -> origin/gh/malfet/589/head 2025-12-04T08:57:03.8184352Z * [new branch] gh/malfet/589/orig -> origin/gh/malfet/589/orig 2025-12-04T08:57:03.8186416Z * [new branch] gh/malfet/590/base -> origin/gh/malfet/590/base 2025-12-04T08:57:03.8187941Z * [new branch] gh/malfet/590/head -> origin/gh/malfet/590/head 2025-12-04T08:57:03.8189540Z * [new branch] gh/malfet/590/orig -> origin/gh/malfet/590/orig 2025-12-04T08:57:03.8192154Z * [new branch] gh/malfet/591/base -> origin/gh/malfet/591/base 2025-12-04T08:57:03.8193679Z * [new branch] gh/malfet/591/head -> origin/gh/malfet/591/head 2025-12-04T08:57:03.8195231Z * [new branch] gh/malfet/591/orig -> origin/gh/malfet/591/orig 2025-12-04T08:57:03.8197369Z * [new branch] gh/malfet/592/base -> origin/gh/malfet/592/base 2025-12-04T08:57:03.8199020Z * [new branch] gh/malfet/592/head -> origin/gh/malfet/592/head 2025-12-04T08:57:03.8200574Z * [new branch] gh/malfet/592/orig -> origin/gh/malfet/592/orig 2025-12-04T08:57:03.8202666Z * [new branch] gh/malfet/593/base -> origin/gh/malfet/593/base 2025-12-04T08:57:03.8204268Z * [new branch] gh/malfet/593/head -> origin/gh/malfet/593/head 2025-12-04T08:57:03.8206217Z * [new branch] gh/malfet/593/orig -> origin/gh/malfet/593/orig 2025-12-04T08:57:03.8208361Z * [new branch] gh/malfet/594/base -> origin/gh/malfet/594/base 2025-12-04T08:57:03.8209893Z * [new branch] gh/malfet/594/head -> origin/gh/malfet/594/head 2025-12-04T08:57:03.8211498Z * [new branch] gh/malfet/594/orig -> origin/gh/malfet/594/orig 2025-12-04T08:57:03.8213831Z * [new branch] gh/malfet/595/base -> origin/gh/malfet/595/base 2025-12-04T08:57:03.8215275Z * [new branch] gh/malfet/595/head -> origin/gh/malfet/595/head 2025-12-04T08:57:03.8216868Z * [new branch] gh/malfet/595/orig -> origin/gh/malfet/595/orig 2025-12-04T08:57:03.8218968Z * [new branch] gh/malfet/596/base -> origin/gh/malfet/596/base 2025-12-04T08:57:03.8220618Z * [new branch] gh/malfet/596/head -> origin/gh/malfet/596/head 2025-12-04T08:57:03.8222119Z * [new branch] gh/malfet/596/orig -> origin/gh/malfet/596/orig 2025-12-04T08:57:03.8224226Z * [new branch] gh/malfet/597/base -> origin/gh/malfet/597/base 2025-12-04T08:57:03.8225776Z * [new branch] gh/malfet/597/head -> origin/gh/malfet/597/head 2025-12-04T08:57:03.8227335Z * [new branch] gh/malfet/597/orig -> origin/gh/malfet/597/orig 2025-12-04T08:57:03.8229458Z * [new branch] gh/malfet/598/base -> origin/gh/malfet/598/base 2025-12-04T08:57:03.8231030Z * [new branch] gh/malfet/598/head -> origin/gh/malfet/598/head 2025-12-04T08:57:03.8232767Z * [new branch] gh/malfet/598/orig -> origin/gh/malfet/598/orig 2025-12-04T08:57:03.8234782Z * [new branch] gh/malfet/599/base -> origin/gh/malfet/599/base 2025-12-04T08:57:03.8236307Z * [new branch] gh/malfet/599/head -> origin/gh/malfet/599/head 2025-12-04T08:57:03.8237858Z * [new branch] gh/malfet/599/orig -> origin/gh/malfet/599/orig 2025-12-04T08:57:03.8239993Z * [new branch] gh/malfet/600/base -> origin/gh/malfet/600/base 2025-12-04T08:57:03.8241646Z * [new branch] gh/malfet/600/head -> origin/gh/malfet/600/head 2025-12-04T08:57:03.8243191Z * [new branch] gh/malfet/600/orig -> origin/gh/malfet/600/orig 2025-12-04T08:57:03.8245299Z * [new branch] gh/malfet/601/base -> origin/gh/malfet/601/base 2025-12-04T08:57:03.8246930Z * [new branch] gh/malfet/601/head -> origin/gh/malfet/601/head 2025-12-04T08:57:03.8248547Z * [new branch] gh/malfet/601/orig -> origin/gh/malfet/601/orig 2025-12-04T08:57:03.8250692Z * [new branch] gh/malfet/602/base -> origin/gh/malfet/602/base 2025-12-04T08:57:03.8252257Z * [new branch] gh/malfet/602/head -> origin/gh/malfet/602/head 2025-12-04T08:57:03.8253991Z * [new branch] gh/malfet/602/orig -> origin/gh/malfet/602/orig 2025-12-04T08:57:03.8256060Z * [new branch] gh/malfet/603/base -> origin/gh/malfet/603/base 2025-12-04T08:57:03.8257549Z * [new branch] gh/malfet/603/head -> origin/gh/malfet/603/head 2025-12-04T08:57:03.8259134Z * [new branch] gh/malfet/603/orig -> origin/gh/malfet/603/orig 2025-12-04T08:57:03.8261215Z * [new branch] gh/malfet/604/base -> origin/gh/malfet/604/base 2025-12-04T08:57:03.8262925Z * [new branch] gh/malfet/604/head -> origin/gh/malfet/604/head 2025-12-04T08:57:03.8264458Z * [new branch] gh/malfet/604/orig -> origin/gh/malfet/604/orig 2025-12-04T08:57:03.8266684Z * [new branch] gh/malfet/605/base -> origin/gh/malfet/605/base 2025-12-04T08:57:03.8268200Z * [new branch] gh/malfet/605/head -> origin/gh/malfet/605/head 2025-12-04T08:57:03.8269821Z * [new branch] gh/malfet/605/orig -> origin/gh/malfet/605/orig 2025-12-04T08:57:03.8271932Z * [new branch] gh/malfet/606/base -> origin/gh/malfet/606/base 2025-12-04T08:57:03.8273603Z * [new branch] gh/malfet/606/head -> origin/gh/malfet/606/head 2025-12-04T08:57:03.8275108Z * [new branch] gh/malfet/606/orig -> origin/gh/malfet/606/orig 2025-12-04T08:57:03.8277265Z * [new branch] gh/malfet/607/base -> origin/gh/malfet/607/base 2025-12-04T08:57:03.8278784Z * [new branch] gh/malfet/607/head -> origin/gh/malfet/607/head 2025-12-04T08:57:03.8280391Z * [new branch] gh/malfet/607/orig -> origin/gh/malfet/607/orig 2025-12-04T08:57:03.8282530Z * [new branch] gh/malfet/608/base -> origin/gh/malfet/608/base 2025-12-04T08:57:03.8284203Z * [new branch] gh/malfet/608/head -> origin/gh/malfet/608/head 2025-12-04T08:57:03.8285738Z * [new branch] gh/malfet/608/orig -> origin/gh/malfet/608/orig 2025-12-04T08:57:03.8287908Z * [new branch] gh/malfet/609/base -> origin/gh/malfet/609/base 2025-12-04T08:57:03.8289964Z * [new branch] gh/malfet/609/head -> origin/gh/malfet/609/head 2025-12-04T08:57:03.8291503Z * [new branch] gh/malfet/609/orig -> origin/gh/malfet/609/orig 2025-12-04T08:57:03.8293773Z * [new branch] gh/malfet/610/base -> origin/gh/malfet/610/base 2025-12-04T08:57:03.8295523Z * [new branch] gh/malfet/610/head -> origin/gh/malfet/610/head 2025-12-04T08:57:03.8297024Z * [new branch] gh/malfet/610/orig -> origin/gh/malfet/610/orig 2025-12-04T08:57:03.8299122Z * [new branch] gh/malfet/611/base -> origin/gh/malfet/611/base 2025-12-04T08:57:03.8300822Z * [new branch] gh/malfet/611/head -> origin/gh/malfet/611/head 2025-12-04T08:57:03.8302348Z * [new branch] gh/malfet/611/orig -> origin/gh/malfet/611/orig 2025-12-04T08:57:03.8304880Z * [new branch] gh/malfet/612/base -> origin/gh/malfet/612/base 2025-12-04T08:57:03.8308247Z * [new branch] gh/malfet/612/head -> origin/gh/malfet/612/head 2025-12-04T08:57:03.8309825Z * [new branch] gh/malfet/612/orig -> origin/gh/malfet/612/orig 2025-12-04T08:57:03.8311967Z * [new branch] gh/malfet/64/base -> origin/gh/malfet/64/base 2025-12-04T08:57:03.8313524Z * [new branch] gh/malfet/64/head -> origin/gh/malfet/64/head 2025-12-04T08:57:03.8316433Z * [new branch] gh/manuelcandales/11/base -> origin/gh/manuelcandales/11/base 2025-12-04T08:57:03.8318265Z * [new branch] gh/manuelcandales/11/head -> origin/gh/manuelcandales/11/head 2025-12-04T08:57:03.8319770Z * [new branch] gh/manuelcandales/11/orig -> origin/gh/manuelcandales/11/orig 2025-12-04T08:57:03.8322371Z * [new branch] gh/markkm/1/base -> origin/gh/markkm/1/base 2025-12-04T08:57:03.8324989Z * [new branch] gh/masnesral/1/base -> origin/gh/masnesral/1/base 2025-12-04T08:57:03.8326620Z * [new branch] gh/masnesral/1/head -> origin/gh/masnesral/1/head 2025-12-04T08:57:03.8328172Z * [new branch] gh/masnesral/1/orig -> origin/gh/masnesral/1/orig 2025-12-04T08:57:03.8330761Z * [new branch] gh/mhorowitz/0/base -> origin/gh/mhorowitz/0/base 2025-12-04T08:57:03.8332368Z * [new branch] gh/mhorowitz/0/head -> origin/gh/mhorowitz/0/head 2025-12-04T08:57:03.8334396Z * [new branch] gh/mhorowitz/1/base -> origin/gh/mhorowitz/1/base 2025-12-04T08:57:03.8336066Z * [new branch] gh/mhorowitz/1/head -> origin/gh/mhorowitz/1/head 2025-12-04T08:57:03.8338012Z * [new branch] gh/mhorowitz/2/base -> origin/gh/mhorowitz/2/base 2025-12-04T08:57:03.8339610Z * [new branch] gh/mhorowitz/2/head -> origin/gh/mhorowitz/2/head 2025-12-04T08:57:03.8341569Z * [new branch] gh/mhorowitz/3/base -> origin/gh/mhorowitz/3/base 2025-12-04T08:57:03.8343108Z * [new branch] gh/mhorowitz/3/head -> origin/gh/mhorowitz/3/head 2025-12-04T08:57:03.8344990Z * [new branch] gh/mhorowitz/4/base -> origin/gh/mhorowitz/4/base 2025-12-04T08:57:03.8346485Z * [new branch] gh/mhorowitz/4/head -> origin/gh/mhorowitz/4/head 2025-12-04T08:57:03.8348424Z * [new branch] gh/mhorowitz/5/base -> origin/gh/mhorowitz/5/base 2025-12-04T08:57:03.8349914Z * [new branch] gh/mhorowitz/5/head -> origin/gh/mhorowitz/5/head 2025-12-04T08:57:03.8351867Z * [new branch] gh/mhorowitz/6/base -> origin/gh/mhorowitz/6/base 2025-12-04T08:57:03.8353329Z * [new branch] gh/mhorowitz/6/head -> origin/gh/mhorowitz/6/head 2025-12-04T08:57:03.8355951Z * [new branch] gh/mikaylagawarecki/234/base -> origin/gh/mikaylagawarecki/234/base 2025-12-04T08:57:03.8357555Z * [new branch] gh/mikaylagawarecki/234/head -> origin/gh/mikaylagawarecki/234/head 2025-12-04T08:57:03.8359607Z * [new branch] gh/mikaylagawarecki/235/base -> origin/gh/mikaylagawarecki/235/base 2025-12-04T08:57:03.8361100Z * [new branch] gh/mikaylagawarecki/235/head -> origin/gh/mikaylagawarecki/235/head 2025-12-04T08:57:03.8363400Z * [new branch] gh/mikaylagawarecki/236/base -> origin/gh/mikaylagawarecki/236/base 2025-12-04T08:57:03.8364758Z * [new branch] gh/mikaylagawarecki/236/head -> origin/gh/mikaylagawarecki/236/head 2025-12-04T08:57:03.8366733Z * [new branch] gh/mikaylagawarecki/237/base -> origin/gh/mikaylagawarecki/237/base 2025-12-04T08:57:03.8368220Z * [new branch] gh/mikaylagawarecki/237/head -> origin/gh/mikaylagawarecki/237/head 2025-12-04T08:57:03.8370311Z * [new branch] gh/mikaylagawarecki/238/base -> origin/gh/mikaylagawarecki/238/base 2025-12-04T08:57:03.8372026Z * [new branch] gh/mikaylagawarecki/238/head -> origin/gh/mikaylagawarecki/238/head 2025-12-04T08:57:03.8374177Z * [new branch] gh/mikaylagawarecki/336/base -> origin/gh/mikaylagawarecki/336/base 2025-12-04T08:57:03.8375703Z * [new branch] gh/mikaylagawarecki/336/head -> origin/gh/mikaylagawarecki/336/head 2025-12-04T08:57:03.8377221Z * [new branch] gh/mikaylagawarecki/336/orig -> origin/gh/mikaylagawarecki/336/orig 2025-12-04T08:57:03.8379415Z * [new branch] gh/mikaylagawarecki/341/base -> origin/gh/mikaylagawarecki/341/base 2025-12-04T08:57:03.8380907Z * [new branch] gh/mikaylagawarecki/341/head -> origin/gh/mikaylagawarecki/341/head 2025-12-04T08:57:03.8382474Z * [new branch] gh/mikaylagawarecki/341/orig -> origin/gh/mikaylagawarecki/341/orig 2025-12-04T08:57:03.8384818Z * [new branch] gh/mikaylagawarecki/342/base -> origin/gh/mikaylagawarecki/342/base 2025-12-04T08:57:03.8386373Z * [new branch] gh/mikaylagawarecki/342/head -> origin/gh/mikaylagawarecki/342/head 2025-12-04T08:57:03.8387945Z * [new branch] gh/mikaylagawarecki/342/orig -> origin/gh/mikaylagawarecki/342/orig 2025-12-04T08:57:03.8390189Z * [new branch] gh/mikaylagawarecki/345/base -> origin/gh/mikaylagawarecki/345/base 2025-12-04T08:57:03.8391727Z * [new branch] gh/mikaylagawarecki/345/head -> origin/gh/mikaylagawarecki/345/head 2025-12-04T08:57:03.8393326Z * [new branch] gh/mikaylagawarecki/345/orig -> origin/gh/mikaylagawarecki/345/orig 2025-12-04T08:57:03.8395462Z * [new branch] gh/mikaylagawarecki/346/base -> origin/gh/mikaylagawarecki/346/base 2025-12-04T08:57:03.8397003Z * [new branch] gh/mikaylagawarecki/346/head -> origin/gh/mikaylagawarecki/346/head 2025-12-04T08:57:03.8398576Z * [new branch] gh/mikaylagawarecki/346/orig -> origin/gh/mikaylagawarecki/346/orig 2025-12-04T08:57:03.8400732Z * [new branch] gh/mikaylagawarecki/347/base -> origin/gh/mikaylagawarecki/347/base 2025-12-04T08:57:03.8402210Z * [new branch] gh/mikaylagawarecki/347/head -> origin/gh/mikaylagawarecki/347/head 2025-12-04T08:57:03.8403809Z * [new branch] gh/mikaylagawarecki/347/orig -> origin/gh/mikaylagawarecki/347/orig 2025-12-04T08:57:03.8406460Z * [new branch] gh/mikaylagawarecki/350/base -> origin/gh/mikaylagawarecki/350/base 2025-12-04T08:57:03.8408032Z * [new branch] gh/mikaylagawarecki/350/head -> origin/gh/mikaylagawarecki/350/head 2025-12-04T08:57:03.8409577Z * [new branch] gh/mikaylagawarecki/350/orig -> origin/gh/mikaylagawarecki/350/orig 2025-12-04T08:57:03.8412031Z * [new branch] gh/mikaylagawarecki/351/base -> origin/gh/mikaylagawarecki/351/base 2025-12-04T08:57:03.8413750Z * [new branch] gh/mikaylagawarecki/351/head -> origin/gh/mikaylagawarecki/351/head 2025-12-04T08:57:03.8415528Z * [new branch] gh/mikaylagawarecki/351/orig -> origin/gh/mikaylagawarecki/351/orig 2025-12-04T08:57:03.8417808Z * [new branch] gh/mikaylagawarecki/352/base -> origin/gh/mikaylagawarecki/352/base 2025-12-04T08:57:03.8419446Z * [new branch] gh/mikaylagawarecki/352/head -> origin/gh/mikaylagawarecki/352/head 2025-12-04T08:57:03.8421026Z * [new branch] gh/mikaylagawarecki/352/orig -> origin/gh/mikaylagawarecki/352/orig 2025-12-04T08:57:03.8423493Z * [new branch] gh/mikaylagawarecki/353/base -> origin/gh/mikaylagawarecki/353/base 2025-12-04T08:57:03.8425042Z * [new branch] gh/mikaylagawarecki/353/head -> origin/gh/mikaylagawarecki/353/head 2025-12-04T08:57:03.8426528Z * [new branch] gh/mikaylagawarecki/353/orig -> origin/gh/mikaylagawarecki/353/orig 2025-12-04T08:57:03.8428658Z * [new branch] gh/mikaylagawarecki/354/base -> origin/gh/mikaylagawarecki/354/base 2025-12-04T08:57:03.8430176Z * [new branch] gh/mikaylagawarecki/354/head -> origin/gh/mikaylagawarecki/354/head 2025-12-04T08:57:03.8431745Z * [new branch] gh/mikaylagawarecki/354/orig -> origin/gh/mikaylagawarecki/354/orig 2025-12-04T08:57:03.8434374Z * [new branch] gh/mikaylagawarecki/356/base -> origin/gh/mikaylagawarecki/356/base 2025-12-04T08:57:03.8436142Z * [new branch] gh/mikaylagawarecki/356/head -> origin/gh/mikaylagawarecki/356/head 2025-12-04T08:57:03.8437655Z * [new branch] gh/mikaylagawarecki/356/orig -> origin/gh/mikaylagawarecki/356/orig 2025-12-04T08:57:03.8439740Z * [new branch] gh/mikaylagawarecki/357/base -> origin/gh/mikaylagawarecki/357/base 2025-12-04T08:57:03.8441306Z * [new branch] gh/mikaylagawarecki/357/head -> origin/gh/mikaylagawarecki/357/head 2025-12-04T08:57:03.8442946Z * [new branch] gh/mikaylagawarecki/357/orig -> origin/gh/mikaylagawarecki/357/orig 2025-12-04T08:57:03.8445167Z * [new branch] gh/mikaylagawarecki/359/base -> origin/gh/mikaylagawarecki/359/base 2025-12-04T08:57:03.8446808Z * [new branch] gh/mikaylagawarecki/359/head -> origin/gh/mikaylagawarecki/359/head 2025-12-04T08:57:03.8448375Z * [new branch] gh/mikaylagawarecki/359/orig -> origin/gh/mikaylagawarecki/359/orig 2025-12-04T08:57:03.8450575Z * [new branch] gh/mikaylagawarecki/360/base -> origin/gh/mikaylagawarecki/360/base 2025-12-04T08:57:03.8452147Z * [new branch] gh/mikaylagawarecki/360/head -> origin/gh/mikaylagawarecki/360/head 2025-12-04T08:57:03.8453829Z * [new branch] gh/mikaylagawarecki/360/orig -> origin/gh/mikaylagawarecki/360/orig 2025-12-04T08:57:03.8456062Z * [new branch] gh/mikaylagawarecki/361/base -> origin/gh/mikaylagawarecki/361/base 2025-12-04T08:57:03.8457645Z * [new branch] gh/mikaylagawarecki/361/head -> origin/gh/mikaylagawarecki/361/head 2025-12-04T08:57:03.8459243Z * [new branch] gh/mikaylagawarecki/361/orig -> origin/gh/mikaylagawarecki/361/orig 2025-12-04T08:57:03.8461461Z * [new branch] gh/mikaylagawarecki/362/base -> origin/gh/mikaylagawarecki/362/base 2025-12-04T08:57:03.8463170Z * [new branch] gh/mikaylagawarecki/362/head -> origin/gh/mikaylagawarecki/362/head 2025-12-04T08:57:03.8464682Z * [new branch] gh/mikaylagawarecki/362/orig -> origin/gh/mikaylagawarecki/362/orig 2025-12-04T08:57:03.8467082Z * [new branch] gh/mikaylagawarecki/363/base -> origin/gh/mikaylagawarecki/363/base 2025-12-04T08:57:03.8468873Z * [new branch] gh/mikaylagawarecki/363/head -> origin/gh/mikaylagawarecki/363/head 2025-12-04T08:57:03.8470432Z * [new branch] gh/mikaylagawarecki/363/orig -> origin/gh/mikaylagawarecki/363/orig 2025-12-04T08:57:03.8473086Z * [new branch] gh/mikaylagawarecki/364/base -> origin/gh/mikaylagawarecki/364/base 2025-12-04T08:57:03.8474628Z * [new branch] gh/mikaylagawarecki/364/head -> origin/gh/mikaylagawarecki/364/head 2025-12-04T08:57:03.8476208Z * [new branch] gh/mikaylagawarecki/364/orig -> origin/gh/mikaylagawarecki/364/orig 2025-12-04T08:57:03.8478475Z * [new branch] gh/mikaylagawarecki/365/base -> origin/gh/mikaylagawarecki/365/base 2025-12-04T08:57:03.8480102Z * [new branch] gh/mikaylagawarecki/365/head -> origin/gh/mikaylagawarecki/365/head 2025-12-04T08:57:03.8481863Z * [new branch] gh/mikaylagawarecki/365/orig -> origin/gh/mikaylagawarecki/365/orig 2025-12-04T08:57:03.8484038Z * [new branch] gh/mikaylagawarecki/366/base -> origin/gh/mikaylagawarecki/366/base 2025-12-04T08:57:03.8485533Z * [new branch] gh/mikaylagawarecki/366/head -> origin/gh/mikaylagawarecki/366/head 2025-12-04T08:57:03.8487069Z * [new branch] gh/mikaylagawarecki/366/orig -> origin/gh/mikaylagawarecki/366/orig 2025-12-04T08:57:03.8489191Z * [new branch] gh/mikaylagawarecki/367/base -> origin/gh/mikaylagawarecki/367/base 2025-12-04T08:57:03.8490769Z * [new branch] gh/mikaylagawarecki/367/head -> origin/gh/mikaylagawarecki/367/head 2025-12-04T08:57:03.8492492Z * [new branch] gh/mikaylagawarecki/367/orig -> origin/gh/mikaylagawarecki/367/orig 2025-12-04T08:57:03.8494912Z * [new branch] gh/mikaylagawarecki/368/base -> origin/gh/mikaylagawarecki/368/base 2025-12-04T08:57:03.8496439Z * [new branch] gh/mikaylagawarecki/368/head -> origin/gh/mikaylagawarecki/368/head 2025-12-04T08:57:03.8497971Z * [new branch] gh/mikaylagawarecki/368/orig -> origin/gh/mikaylagawarecki/368/orig 2025-12-04T08:57:03.8500210Z * [new branch] gh/mikaylagawarecki/369/base -> origin/gh/mikaylagawarecki/369/base 2025-12-04T08:57:03.8502294Z * [new branch] gh/mikaylagawarecki/369/head -> origin/gh/mikaylagawarecki/369/head 2025-12-04T08:57:03.8503820Z * [new branch] gh/mikaylagawarecki/369/orig -> origin/gh/mikaylagawarecki/369/orig 2025-12-04T08:57:03.8506341Z * [new branch] gh/mikaylagawarecki/370/base -> origin/gh/mikaylagawarecki/370/base 2025-12-04T08:57:03.8507965Z * [new branch] gh/mikaylagawarecki/370/head -> origin/gh/mikaylagawarecki/370/head 2025-12-04T08:57:03.8509519Z * [new branch] gh/mikaylagawarecki/370/orig -> origin/gh/mikaylagawarecki/370/orig 2025-12-04T08:57:03.8511691Z * [new branch] gh/mikaylagawarecki/371/base -> origin/gh/mikaylagawarecki/371/base 2025-12-04T08:57:03.8513211Z * [new branch] gh/mikaylagawarecki/371/head -> origin/gh/mikaylagawarecki/371/head 2025-12-04T08:57:03.8514800Z * [new branch] gh/mikaylagawarecki/371/orig -> origin/gh/mikaylagawarecki/371/orig 2025-12-04T08:57:03.8517196Z * [new branch] gh/mikaylagawarecki/372/base -> origin/gh/mikaylagawarecki/372/base 2025-12-04T08:57:03.8518725Z * [new branch] gh/mikaylagawarecki/372/head -> origin/gh/mikaylagawarecki/372/head 2025-12-04T08:57:03.8520258Z * [new branch] gh/mikaylagawarecki/372/orig -> origin/gh/mikaylagawarecki/372/orig 2025-12-04T08:57:03.8522400Z * [new branch] gh/mikaylagawarecki/373/base -> origin/gh/mikaylagawarecki/373/base 2025-12-04T08:57:03.8523899Z * [new branch] gh/mikaylagawarecki/373/head -> origin/gh/mikaylagawarecki/373/head 2025-12-04T08:57:03.8525455Z * [new branch] gh/mikaylagawarecki/373/orig -> origin/gh/mikaylagawarecki/373/orig 2025-12-04T08:57:03.8527602Z * [new branch] gh/mikaylagawarecki/374/base -> origin/gh/mikaylagawarecki/374/base 2025-12-04T08:57:03.8529234Z * [new branch] gh/mikaylagawarecki/374/head -> origin/gh/mikaylagawarecki/374/head 2025-12-04T08:57:03.8530754Z * [new branch] gh/mikaylagawarecki/374/orig -> origin/gh/mikaylagawarecki/374/orig 2025-12-04T08:57:03.8532855Z * [new branch] gh/mikaylagawarecki/375/base -> origin/gh/mikaylagawarecki/375/base 2025-12-04T08:57:03.8534599Z * [new branch] gh/mikaylagawarecki/375/head -> origin/gh/mikaylagawarecki/375/head 2025-12-04T08:57:03.8536188Z * [new branch] gh/mikaylagawarecki/375/orig -> origin/gh/mikaylagawarecki/375/orig 2025-12-04T08:57:03.8538396Z * [new branch] gh/mikaylagawarecki/376/base -> origin/gh/mikaylagawarecki/376/base 2025-12-04T08:57:03.8540534Z * [new branch] gh/mikaylagawarecki/376/head -> origin/gh/mikaylagawarecki/376/head 2025-12-04T08:57:03.8542250Z * [new branch] gh/mikaylagawarecki/376/orig -> origin/gh/mikaylagawarecki/376/orig 2025-12-04T08:57:03.8544429Z * [new branch] gh/mikaylagawarecki/377/base -> origin/gh/mikaylagawarecki/377/base 2025-12-04T08:57:03.8545956Z * [new branch] gh/mikaylagawarecki/377/head -> origin/gh/mikaylagawarecki/377/head 2025-12-04T08:57:03.8547551Z * [new branch] gh/mikaylagawarecki/377/orig -> origin/gh/mikaylagawarecki/377/orig 2025-12-04T08:57:03.8549708Z * [new branch] gh/mikaylagawarecki/378/base -> origin/gh/mikaylagawarecki/378/base 2025-12-04T08:57:03.8551300Z * [new branch] gh/mikaylagawarecki/378/head -> origin/gh/mikaylagawarecki/378/head 2025-12-04T08:57:03.8552837Z * [new branch] gh/mikaylagawarecki/378/orig -> origin/gh/mikaylagawarecki/378/orig 2025-12-04T08:57:03.8554896Z * [new branch] gh/mikaylagawarecki/379/base -> origin/gh/mikaylagawarecki/379/base 2025-12-04T08:57:03.8556515Z * [new branch] gh/mikaylagawarecki/379/head -> origin/gh/mikaylagawarecki/379/head 2025-12-04T08:57:03.8558012Z * [new branch] gh/mikaylagawarecki/379/orig -> origin/gh/mikaylagawarecki/379/orig 2025-12-04T08:57:03.8560602Z * [new branch] gh/mikaylagawarecki/380/base -> origin/gh/mikaylagawarecki/380/base 2025-12-04T08:57:03.8562219Z * [new branch] gh/mikaylagawarecki/380/head -> origin/gh/mikaylagawarecki/380/head 2025-12-04T08:57:03.8563829Z * [new branch] gh/mikaylagawarecki/380/orig -> origin/gh/mikaylagawarecki/380/orig 2025-12-04T08:57:03.8565837Z * [new branch] gh/mikaylagawarecki/381/base -> origin/gh/mikaylagawarecki/381/base 2025-12-04T08:57:03.8567432Z * [new branch] gh/mikaylagawarecki/381/head -> origin/gh/mikaylagawarecki/381/head 2025-12-04T08:57:03.8568932Z * [new branch] gh/mikaylagawarecki/381/orig -> origin/gh/mikaylagawarecki/381/orig 2025-12-04T08:57:03.8570960Z * [new branch] gh/mikaylagawarecki/382/base -> origin/gh/mikaylagawarecki/382/base 2025-12-04T08:57:03.8572486Z * [new branch] gh/mikaylagawarecki/382/head -> origin/gh/mikaylagawarecki/382/head 2025-12-04T08:57:03.8574122Z * [new branch] gh/mikaylagawarecki/382/orig -> origin/gh/mikaylagawarecki/382/orig 2025-12-04T08:57:03.8576271Z * [new branch] gh/mikaylagawarecki/383/base -> origin/gh/mikaylagawarecki/383/base 2025-12-04T08:57:03.8577915Z * [new branch] gh/mikaylagawarecki/383/head -> origin/gh/mikaylagawarecki/383/head 2025-12-04T08:57:03.8579490Z * [new branch] gh/mikaylagawarecki/383/orig -> origin/gh/mikaylagawarecki/383/orig 2025-12-04T08:57:03.8581757Z * [new branch] gh/mikaylagawarecki/384/base -> origin/gh/mikaylagawarecki/384/base 2025-12-04T08:57:03.8583758Z * [new branch] gh/mikaylagawarecki/384/head -> origin/gh/mikaylagawarecki/384/head 2025-12-04T08:57:03.8585557Z * [new branch] gh/mikaylagawarecki/384/orig -> origin/gh/mikaylagawarecki/384/orig 2025-12-04T08:57:03.8587624Z * [new branch] gh/mikaylagawarecki/385/base -> origin/gh/mikaylagawarecki/385/base 2025-12-04T08:57:03.8589222Z * [new branch] gh/mikaylagawarecki/385/head -> origin/gh/mikaylagawarecki/385/head 2025-12-04T08:57:03.8590805Z * [new branch] gh/mikaylagawarecki/385/orig -> origin/gh/mikaylagawarecki/385/orig 2025-12-04T08:57:03.8593082Z * [new branch] gh/mikaylagawarecki/386/base -> origin/gh/mikaylagawarecki/386/base 2025-12-04T08:57:03.8594586Z * [new branch] gh/mikaylagawarecki/386/head -> origin/gh/mikaylagawarecki/386/head 2025-12-04T08:57:03.8596173Z * [new branch] gh/mikaylagawarecki/386/orig -> origin/gh/mikaylagawarecki/386/orig 2025-12-04T08:57:03.8598382Z * [new branch] gh/mikaylagawarecki/387/base -> origin/gh/mikaylagawarecki/387/base 2025-12-04T08:57:03.8600079Z * [new branch] gh/mikaylagawarecki/387/head -> origin/gh/mikaylagawarecki/387/head 2025-12-04T08:57:03.8601570Z * [new branch] gh/mikaylagawarecki/387/orig -> origin/gh/mikaylagawarecki/387/orig 2025-12-04T08:57:03.8603631Z * [new branch] gh/mikaylagawarecki/388/base -> origin/gh/mikaylagawarecki/388/base 2025-12-04T08:57:03.8605460Z * [new branch] gh/mikaylagawarecki/388/head -> origin/gh/mikaylagawarecki/388/head 2025-12-04T08:57:03.8607138Z * [new branch] gh/mikaylagawarecki/388/orig -> origin/gh/mikaylagawarecki/388/orig 2025-12-04T08:57:03.8609381Z * [new branch] gh/mikaylagawarecki/389/base -> origin/gh/mikaylagawarecki/389/base 2025-12-04T08:57:03.8610937Z * [new branch] gh/mikaylagawarecki/389/head -> origin/gh/mikaylagawarecki/389/head 2025-12-04T08:57:03.8612549Z * [new branch] gh/mikaylagawarecki/389/orig -> origin/gh/mikaylagawarecki/389/orig 2025-12-04T08:57:03.8614888Z * [new branch] gh/mikaylagawarecki/390/base -> origin/gh/mikaylagawarecki/390/base 2025-12-04T08:57:03.8616372Z * [new branch] gh/mikaylagawarecki/390/head -> origin/gh/mikaylagawarecki/390/head 2025-12-04T08:57:03.8617951Z * [new branch] gh/mikaylagawarecki/390/orig -> origin/gh/mikaylagawarecki/390/orig 2025-12-04T08:57:03.8620281Z * [new branch] gh/mikaylagawarecki/391/base -> origin/gh/mikaylagawarecki/391/base 2025-12-04T08:57:03.8621815Z * [new branch] gh/mikaylagawarecki/391/head -> origin/gh/mikaylagawarecki/391/head 2025-12-04T08:57:03.8623468Z * [new branch] gh/mikaylagawarecki/391/orig -> origin/gh/mikaylagawarecki/391/orig 2025-12-04T08:57:03.8626856Z * [new branch] gh/mikaylagawarecki/392/base -> origin/gh/mikaylagawarecki/392/base 2025-12-04T08:57:03.8628435Z * [new branch] gh/mikaylagawarecki/392/head -> origin/gh/mikaylagawarecki/392/head 2025-12-04T08:57:03.8629977Z * [new branch] gh/mikaylagawarecki/392/orig -> origin/gh/mikaylagawarecki/392/orig 2025-12-04T08:57:03.8632473Z * [new branch] gh/mlazos/41/base -> origin/gh/mlazos/41/base 2025-12-04T08:57:03.8633996Z * [new branch] gh/mlazos/41/head -> origin/gh/mlazos/41/head 2025-12-04T08:57:03.8635550Z * [new branch] gh/mlazos/41/orig -> origin/gh/mlazos/41/orig 2025-12-04T08:57:03.8637702Z * [new branch] gh/mlazos/42/base -> origin/gh/mlazos/42/base 2025-12-04T08:57:03.8639191Z * [new branch] gh/mlazos/42/head -> origin/gh/mlazos/42/head 2025-12-04T08:57:03.8640727Z * [new branch] gh/mlazos/42/orig -> origin/gh/mlazos/42/orig 2025-12-04T08:57:03.8643213Z * [new branch] gh/mlazos/43/base -> origin/gh/mlazos/43/base 2025-12-04T08:57:03.8644734Z * [new branch] gh/mlazos/43/head -> origin/gh/mlazos/43/head 2025-12-04T08:57:03.8646427Z * [new branch] gh/mlazos/43/orig -> origin/gh/mlazos/43/orig 2025-12-04T08:57:03.8648420Z * [new branch] gh/mlazos/44/base -> origin/gh/mlazos/44/base 2025-12-04T08:57:03.8649957Z * [new branch] gh/mlazos/44/head -> origin/gh/mlazos/44/head 2025-12-04T08:57:03.8651530Z * [new branch] gh/mlazos/44/orig -> origin/gh/mlazos/44/orig 2025-12-04T08:57:03.8653691Z * [new branch] gh/mlazos/47/base -> origin/gh/mlazos/47/base 2025-12-04T08:57:03.8655761Z * [new branch] gh/mlazos/47/head -> origin/gh/mlazos/47/head 2025-12-04T08:57:03.8657353Z * [new branch] gh/mlazos/47/orig -> origin/gh/mlazos/47/orig 2025-12-04T08:57:03.8659340Z * [new branch] gh/mlazos/48/base -> origin/gh/mlazos/48/base 2025-12-04T08:57:03.8660953Z * [new branch] gh/mlazos/48/head -> origin/gh/mlazos/48/head 2025-12-04T08:57:03.8662674Z * [new branch] gh/mlazos/48/orig -> origin/gh/mlazos/48/orig 2025-12-04T08:57:03.8664624Z * [new branch] gh/mlazos/49/base -> origin/gh/mlazos/49/base 2025-12-04T08:57:03.8666416Z * [new branch] gh/mlazos/49/head -> origin/gh/mlazos/49/head 2025-12-04T08:57:03.8668001Z * [new branch] gh/mlazos/49/orig -> origin/gh/mlazos/49/orig 2025-12-04T08:57:03.8670490Z * [new branch] gh/mlazos/50/base -> origin/gh/mlazos/50/base 2025-12-04T08:57:03.8672066Z * [new branch] gh/mlazos/50/head -> origin/gh/mlazos/50/head 2025-12-04T08:57:03.8673502Z * [new branch] gh/mlazos/50/orig -> origin/gh/mlazos/50/orig 2025-12-04T08:57:03.8675569Z * [new branch] gh/mlazos/51/base -> origin/gh/mlazos/51/base 2025-12-04T08:57:03.8677119Z * [new branch] gh/mlazos/51/head -> origin/gh/mlazos/51/head 2025-12-04T08:57:03.8678778Z * [new branch] gh/mlazos/51/orig -> origin/gh/mlazos/51/orig 2025-12-04T08:57:03.8680827Z * [new branch] gh/mlazos/52/base -> origin/gh/mlazos/52/base 2025-12-04T08:57:03.8682361Z * [new branch] gh/mlazos/52/head -> origin/gh/mlazos/52/head 2025-12-04T08:57:03.8683915Z * [new branch] gh/mlazos/52/orig -> origin/gh/mlazos/52/orig 2025-12-04T08:57:03.8685968Z * [new branch] gh/mlazos/53/base -> origin/gh/mlazos/53/base 2025-12-04T08:57:03.8687494Z * [new branch] gh/mlazos/53/head -> origin/gh/mlazos/53/head 2025-12-04T08:57:03.8689143Z * [new branch] gh/mlazos/53/orig -> origin/gh/mlazos/53/orig 2025-12-04T08:57:03.8691311Z * [new branch] gh/mlazos/54/base -> origin/gh/mlazos/54/base 2025-12-04T08:57:03.8692856Z * [new branch] gh/mlazos/54/head -> origin/gh/mlazos/54/head 2025-12-04T08:57:03.8694535Z * [new branch] gh/mlazos/54/orig -> origin/gh/mlazos/54/orig 2025-12-04T08:57:03.8696550Z * [new branch] gh/mlazos/55/base -> origin/gh/mlazos/55/base 2025-12-04T08:57:03.8698176Z * [new branch] gh/mlazos/55/head -> origin/gh/mlazos/55/head 2025-12-04T08:57:03.8699695Z * [new branch] gh/mlazos/55/orig -> origin/gh/mlazos/55/orig 2025-12-04T08:57:03.8701769Z * [new branch] gh/mlazos/56/base -> origin/gh/mlazos/56/base 2025-12-04T08:57:03.8703345Z * [new branch] gh/mlazos/56/head -> origin/gh/mlazos/56/head 2025-12-04T08:57:03.8705107Z * [new branch] gh/mlazos/56/orig -> origin/gh/mlazos/56/orig 2025-12-04T08:57:03.8708850Z * [new branch] gh/mlazos/57/base -> origin/gh/mlazos/57/base 2025-12-04T08:57:03.8710396Z * [new branch] gh/mlazos/57/head -> origin/gh/mlazos/57/head 2025-12-04T08:57:03.8712024Z * [new branch] gh/mlazos/57/orig -> origin/gh/mlazos/57/orig 2025-12-04T08:57:03.8714094Z * [new branch] gh/mlazos/58/base -> origin/gh/mlazos/58/base 2025-12-04T08:57:03.8715687Z * [new branch] gh/mlazos/58/head -> origin/gh/mlazos/58/head 2025-12-04T08:57:03.8717276Z * [new branch] gh/mlazos/58/orig -> origin/gh/mlazos/58/orig 2025-12-04T08:57:03.8719333Z * [new branch] gh/mlazos/59/base -> origin/gh/mlazos/59/base 2025-12-04T08:57:03.8720904Z * [new branch] gh/mlazos/59/head -> origin/gh/mlazos/59/head 2025-12-04T08:57:03.8722359Z * [new branch] gh/mlazos/59/orig -> origin/gh/mlazos/59/orig 2025-12-04T08:57:03.8724584Z * [new branch] gh/mlazos/60/base -> origin/gh/mlazos/60/base 2025-12-04T08:57:03.8726153Z * [new branch] gh/mlazos/60/head -> origin/gh/mlazos/60/head 2025-12-04T08:57:03.8727839Z * [new branch] gh/mlazos/60/orig -> origin/gh/mlazos/60/orig 2025-12-04T08:57:03.8730325Z * [new branch] gh/mlazos/61/base -> origin/gh/mlazos/61/base 2025-12-04T08:57:03.8731902Z * [new branch] gh/mlazos/61/head -> origin/gh/mlazos/61/head 2025-12-04T08:57:03.8733648Z * [new branch] gh/mlazos/61/orig -> origin/gh/mlazos/61/orig 2025-12-04T08:57:03.8736513Z * [new branch] gh/mlazos/62/base -> origin/gh/mlazos/62/base 2025-12-04T08:57:03.8737324Z * [new branch] gh/mlazos/62/head -> origin/gh/mlazos/62/head 2025-12-04T08:57:03.8738896Z * [new branch] gh/mlazos/62/orig -> origin/gh/mlazos/62/orig 2025-12-04T08:57:03.8741158Z * [new branch] gh/mlazos/63/base -> origin/gh/mlazos/63/base 2025-12-04T08:57:03.8742760Z * [new branch] gh/mlazos/63/head -> origin/gh/mlazos/63/head 2025-12-04T08:57:03.8744307Z * [new branch] gh/mlazos/63/orig -> origin/gh/mlazos/63/orig 2025-12-04T08:57:03.8746372Z * [new branch] gh/mlazos/64/base -> origin/gh/mlazos/64/base 2025-12-04T08:57:03.8747921Z * [new branch] gh/mlazos/64/head -> origin/gh/mlazos/64/head 2025-12-04T08:57:03.8749452Z * [new branch] gh/mlazos/64/orig -> origin/gh/mlazos/64/orig 2025-12-04T08:57:03.8751598Z * [new branch] gh/mlazos/65/base -> origin/gh/mlazos/65/base 2025-12-04T08:57:03.8753202Z * [new branch] gh/mlazos/65/head -> origin/gh/mlazos/65/head 2025-12-04T08:57:03.8754765Z * [new branch] gh/mlazos/65/orig -> origin/gh/mlazos/65/orig 2025-12-04T08:57:03.8756905Z * [new branch] gh/mlazos/66/base -> origin/gh/mlazos/66/base 2025-12-04T08:57:03.8758462Z * [new branch] gh/mlazos/66/head -> origin/gh/mlazos/66/head 2025-12-04T08:57:03.8760011Z * [new branch] gh/mlazos/66/orig -> origin/gh/mlazos/66/orig 2025-12-04T08:57:03.8762648Z * [new branch] gh/mlazos/67/base -> origin/gh/mlazos/67/base 2025-12-04T08:57:03.8764232Z * [new branch] gh/mlazos/67/head -> origin/gh/mlazos/67/head 2025-12-04T08:57:03.8765801Z * [new branch] gh/mlazos/67/orig -> origin/gh/mlazos/67/orig 2025-12-04T08:57:03.8767938Z * [new branch] gh/mlazos/68/base -> origin/gh/mlazos/68/base 2025-12-04T08:57:03.8769483Z * [new branch] gh/mlazos/68/head -> origin/gh/mlazos/68/head 2025-12-04T08:57:03.8771087Z * [new branch] gh/mlazos/68/orig -> origin/gh/mlazos/68/orig 2025-12-04T08:57:03.8773301Z * [new branch] gh/mlazos/69/base -> origin/gh/mlazos/69/base 2025-12-04T08:57:03.8774970Z * [new branch] gh/mlazos/69/head -> origin/gh/mlazos/69/head 2025-12-04T08:57:03.8776682Z * [new branch] gh/mlazos/69/orig -> origin/gh/mlazos/69/orig 2025-12-04T08:57:03.8778814Z * [new branch] gh/mlazos/70/base -> origin/gh/mlazos/70/base 2025-12-04T08:57:03.8780382Z * [new branch] gh/mlazos/70/head -> origin/gh/mlazos/70/head 2025-12-04T08:57:03.8781965Z * [new branch] gh/mlazos/70/orig -> origin/gh/mlazos/70/orig 2025-12-04T08:57:03.8784244Z * [new branch] gh/mlazos/71/base -> origin/gh/mlazos/71/base 2025-12-04T08:57:03.8785793Z * [new branch] gh/mlazos/71/head -> origin/gh/mlazos/71/head 2025-12-04T08:57:03.8787329Z * [new branch] gh/mlazos/71/orig -> origin/gh/mlazos/71/orig 2025-12-04T08:57:03.8789488Z * [new branch] gh/mlazos/72/base -> origin/gh/mlazos/72/base 2025-12-04T08:57:03.8790985Z * [new branch] gh/mlazos/72/head -> origin/gh/mlazos/72/head 2025-12-04T08:57:03.8792690Z * [new branch] gh/mlazos/72/orig -> origin/gh/mlazos/72/orig 2025-12-04T08:57:03.8795164Z * [new branch] gh/mlazos/73/base -> origin/gh/mlazos/73/base 2025-12-04T08:57:03.8796739Z * [new branch] gh/mlazos/73/head -> origin/gh/mlazos/73/head 2025-12-04T08:57:03.8798416Z * [new branch] gh/mlazos/73/orig -> origin/gh/mlazos/73/orig 2025-12-04T08:57:03.8801053Z * [new branch] gh/mrmiywj/1/base -> origin/gh/mrmiywj/1/base 2025-12-04T08:57:03.8802724Z * [new branch] gh/mrmiywj/1/head -> origin/gh/mrmiywj/1/head 2025-12-04T08:57:03.8805534Z * [new branch] gh/muchulee8/73/base -> origin/gh/muchulee8/73/base 2025-12-04T08:57:03.8807249Z * [new branch] gh/muchulee8/73/head -> origin/gh/muchulee8/73/head 2025-12-04T08:57:03.8808841Z * [new branch] gh/muchulee8/73/orig -> origin/gh/muchulee8/73/orig 2025-12-04T08:57:03.8811489Z * [new branch] gh/naveenthangudu/1/base -> origin/gh/naveenthangudu/1/base 2025-12-04T08:57:03.8813107Z * [new branch] gh/naveenthangudu/1/head -> origin/gh/naveenthangudu/1/head 2025-12-04T08:57:03.8814815Z * [new branch] gh/naveenthangudu/1/orig -> origin/gh/naveenthangudu/1/orig 2025-12-04T08:57:03.8817083Z * [new branch] gh/naveenthangudu/2/base -> origin/gh/naveenthangudu/2/base 2025-12-04T08:57:03.8818609Z * [new branch] gh/naveenthangudu/2/head -> origin/gh/naveenthangudu/2/head 2025-12-04T08:57:03.8820234Z * [new branch] gh/naveenthangudu/2/orig -> origin/gh/naveenthangudu/2/orig 2025-12-04T08:57:03.8822659Z * [new branch] gh/naveenthangudu/3/base -> origin/gh/naveenthangudu/3/base 2025-12-04T08:57:03.8824155Z * [new branch] gh/naveenthangudu/3/head -> origin/gh/naveenthangudu/3/head 2025-12-04T08:57:03.8825747Z * [new branch] gh/naveenthangudu/3/orig -> origin/gh/naveenthangudu/3/orig 2025-12-04T08:57:03.8827810Z * [new branch] gh/naveenthangudu/4/base -> origin/gh/naveenthangudu/4/base 2025-12-04T08:57:03.8829441Z * [new branch] gh/naveenthangudu/4/head -> origin/gh/naveenthangudu/4/head 2025-12-04T08:57:03.8831067Z * [new branch] gh/naveenthangudu/4/orig -> origin/gh/naveenthangudu/4/orig 2025-12-04T08:57:03.8833128Z * [new branch] gh/naveenthangudu/5/base -> origin/gh/naveenthangudu/5/base 2025-12-04T08:57:03.8834743Z * [new branch] gh/naveenthangudu/5/head -> origin/gh/naveenthangudu/5/head 2025-12-04T08:57:03.8836439Z * [new branch] gh/naveenthangudu/5/orig -> origin/gh/naveenthangudu/5/orig 2025-12-04T08:57:03.8838591Z * [new branch] gh/naveenthangudu/6/base -> origin/gh/naveenthangudu/6/base 2025-12-04T08:57:03.8840703Z * [new branch] gh/naveenthangudu/6/head -> origin/gh/naveenthangudu/6/head 2025-12-04T08:57:03.8842252Z * [new branch] gh/naveenthangudu/6/orig -> origin/gh/naveenthangudu/6/orig 2025-12-04T08:57:03.8844358Z * [new branch] gh/naveenthangudu/7/base -> origin/gh/naveenthangudu/7/base 2025-12-04T08:57:03.8845906Z * [new branch] gh/naveenthangudu/7/head -> origin/gh/naveenthangudu/7/head 2025-12-04T08:57:03.8847407Z * [new branch] gh/naveenthangudu/7/orig -> origin/gh/naveenthangudu/7/orig 2025-12-04T08:57:03.8849462Z * [new branch] gh/naveenthangudu/8/base -> origin/gh/naveenthangudu/8/base 2025-12-04T08:57:03.8851060Z * [new branch] gh/naveenthangudu/8/head -> origin/gh/naveenthangudu/8/head 2025-12-04T08:57:03.8852712Z * [new branch] gh/naveenthangudu/8/orig -> origin/gh/naveenthangudu/8/orig 2025-12-04T08:57:03.8854930Z * [new branch] gh/naveenthangudu/9/base -> origin/gh/naveenthangudu/9/base 2025-12-04T08:57:03.8856497Z * [new branch] gh/naveenthangudu/9/head -> origin/gh/naveenthangudu/9/head 2025-12-04T08:57:03.8858678Z * [new branch] gh/naveenthangudu/9/orig -> origin/gh/naveenthangudu/9/orig 2025-12-04T08:57:03.8861192Z * [new branch] gh/nikitaved/1/base -> origin/gh/nikitaved/1/base 2025-12-04T08:57:03.8862720Z * [new branch] gh/nikitaved/1/head -> origin/gh/nikitaved/1/head 2025-12-04T08:57:03.8864278Z * [new branch] gh/nikitaved/1/orig -> origin/gh/nikitaved/1/orig 2025-12-04T08:57:03.8866412Z * [new branch] gh/nikitaved/10/base -> origin/gh/nikitaved/10/base 2025-12-04T08:57:03.8868195Z * [new branch] gh/nikitaved/10/head -> origin/gh/nikitaved/10/head 2025-12-04T08:57:03.8869753Z * [new branch] gh/nikitaved/10/orig -> origin/gh/nikitaved/10/orig 2025-12-04T08:57:03.8871802Z * [new branch] gh/nikitaved/11/base -> origin/gh/nikitaved/11/base 2025-12-04T08:57:03.8873405Z * [new branch] gh/nikitaved/11/head -> origin/gh/nikitaved/11/head 2025-12-04T08:57:03.8875473Z * [new branch] gh/nikitaved/11/orig -> origin/gh/nikitaved/11/orig 2025-12-04T08:57:03.8877474Z * [new branch] gh/nikitaved/12/base -> origin/gh/nikitaved/12/base 2025-12-04T08:57:03.8879068Z * [new branch] gh/nikitaved/12/head -> origin/gh/nikitaved/12/head 2025-12-04T08:57:03.8880709Z * [new branch] gh/nikitaved/12/orig -> origin/gh/nikitaved/12/orig 2025-12-04T08:57:03.8882909Z * [new branch] gh/nikitaved/13/base -> origin/gh/nikitaved/13/base 2025-12-04T08:57:03.8884506Z * [new branch] gh/nikitaved/13/head -> origin/gh/nikitaved/13/head 2025-12-04T08:57:03.8886078Z * [new branch] gh/nikitaved/13/orig -> origin/gh/nikitaved/13/orig 2025-12-04T08:57:03.8888209Z * [new branch] gh/nikitaved/14/base -> origin/gh/nikitaved/14/base 2025-12-04T08:57:03.8889786Z * [new branch] gh/nikitaved/14/head -> origin/gh/nikitaved/14/head 2025-12-04T08:57:03.8891403Z * [new branch] gh/nikitaved/14/orig -> origin/gh/nikitaved/14/orig 2025-12-04T08:57:03.8893871Z * [new branch] gh/nikitaved/15/base -> origin/gh/nikitaved/15/base 2025-12-04T08:57:03.8895575Z * [new branch] gh/nikitaved/15/head -> origin/gh/nikitaved/15/head 2025-12-04T08:57:03.8897089Z * [new branch] gh/nikitaved/15/orig -> origin/gh/nikitaved/15/orig 2025-12-04T08:57:03.8899171Z * [new branch] gh/nikitaved/16/base -> origin/gh/nikitaved/16/base 2025-12-04T08:57:03.8901193Z * [new branch] gh/nikitaved/16/head -> origin/gh/nikitaved/16/head 2025-12-04T08:57:03.8902781Z * [new branch] gh/nikitaved/16/orig -> origin/gh/nikitaved/16/orig 2025-12-04T08:57:03.8905179Z * [new branch] gh/nikitaved/2/base -> origin/gh/nikitaved/2/base 2025-12-04T08:57:03.8908295Z * [new branch] gh/nikitaved/2/head -> origin/gh/nikitaved/2/head 2025-12-04T08:57:03.8909787Z * [new branch] gh/nikitaved/2/orig -> origin/gh/nikitaved/2/orig 2025-12-04T08:57:03.8911796Z * [new branch] gh/nikitaved/4/base -> origin/gh/nikitaved/4/base 2025-12-04T08:57:03.8913365Z * [new branch] gh/nikitaved/4/head -> origin/gh/nikitaved/4/head 2025-12-04T08:57:03.8915022Z * [new branch] gh/nikitaved/4/orig -> origin/gh/nikitaved/4/orig 2025-12-04T08:57:03.8917620Z * [new branch] gh/nikitaved/5/base -> origin/gh/nikitaved/5/base 2025-12-04T08:57:03.8919154Z * [new branch] gh/nikitaved/5/head -> origin/gh/nikitaved/5/head 2025-12-04T08:57:03.8920699Z * [new branch] gh/nikitaved/5/orig -> origin/gh/nikitaved/5/orig 2025-12-04T08:57:03.8922747Z * [new branch] gh/nikitaved/6/base -> origin/gh/nikitaved/6/base 2025-12-04T08:57:03.8924372Z * [new branch] gh/nikitaved/6/head -> origin/gh/nikitaved/6/head 2025-12-04T08:57:03.8925847Z * [new branch] gh/nikitaved/6/orig -> origin/gh/nikitaved/6/orig 2025-12-04T08:57:03.8927954Z * [new branch] gh/nikitaved/8/base -> origin/gh/nikitaved/8/base 2025-12-04T08:57:03.8929526Z * [new branch] gh/nikitaved/8/head -> origin/gh/nikitaved/8/head 2025-12-04T08:57:03.8931040Z * [new branch] gh/nikitaved/8/orig -> origin/gh/nikitaved/8/orig 2025-12-04T08:57:03.8933170Z * [new branch] gh/nikitaved/9/base -> origin/gh/nikitaved/9/base 2025-12-04T08:57:03.8934922Z * [new branch] gh/nikitaved/9/head -> origin/gh/nikitaved/9/head 2025-12-04T08:57:03.8936454Z * [new branch] gh/nikitaved/9/orig -> origin/gh/nikitaved/9/orig 2025-12-04T08:57:03.8938947Z * [new branch] gh/oulgen/10/base -> origin/gh/oulgen/10/base 2025-12-04T08:57:03.8940922Z * [new branch] gh/oulgen/10/head -> origin/gh/oulgen/10/head 2025-12-04T08:57:03.8942473Z * [new branch] gh/oulgen/10/orig -> origin/gh/oulgen/10/orig 2025-12-04T08:57:03.8944532Z * [new branch] gh/oulgen/11/base -> origin/gh/oulgen/11/base 2025-12-04T08:57:03.8946061Z * [new branch] gh/oulgen/11/head -> origin/gh/oulgen/11/head 2025-12-04T08:57:03.8947668Z * [new branch] gh/oulgen/11/orig -> origin/gh/oulgen/11/orig 2025-12-04T08:57:03.8949709Z * [new branch] gh/oulgen/12/base -> origin/gh/oulgen/12/base 2025-12-04T08:57:03.8951239Z * [new branch] gh/oulgen/12/head -> origin/gh/oulgen/12/head 2025-12-04T08:57:03.8952789Z * [new branch] gh/oulgen/12/orig -> origin/gh/oulgen/12/orig 2025-12-04T08:57:03.8954814Z * [new branch] gh/oulgen/13/base -> origin/gh/oulgen/13/base 2025-12-04T08:57:03.8956351Z * [new branch] gh/oulgen/13/head -> origin/gh/oulgen/13/head 2025-12-04T08:57:03.8957925Z * [new branch] gh/oulgen/13/orig -> origin/gh/oulgen/13/orig 2025-12-04T08:57:03.8959957Z * [new branch] gh/oulgen/14/base -> origin/gh/oulgen/14/base 2025-12-04T08:57:03.8961576Z * [new branch] gh/oulgen/14/head -> origin/gh/oulgen/14/head 2025-12-04T08:57:03.8963143Z * [new branch] gh/oulgen/14/orig -> origin/gh/oulgen/14/orig 2025-12-04T08:57:03.8965359Z * [new branch] gh/oulgen/15/base -> origin/gh/oulgen/15/base 2025-12-04T08:57:03.8967174Z * [new branch] gh/oulgen/15/head -> origin/gh/oulgen/15/head 2025-12-04T08:57:03.8968682Z * [new branch] gh/oulgen/15/orig -> origin/gh/oulgen/15/orig 2025-12-04T08:57:03.8970713Z * [new branch] gh/oulgen/16/base -> origin/gh/oulgen/16/base 2025-12-04T08:57:03.8972181Z * [new branch] gh/oulgen/16/head -> origin/gh/oulgen/16/head 2025-12-04T08:57:03.8973818Z * [new branch] gh/oulgen/16/orig -> origin/gh/oulgen/16/orig 2025-12-04T08:57:03.8975942Z * [new branch] gh/oulgen/17/base -> origin/gh/oulgen/17/base 2025-12-04T08:57:03.8977497Z * [new branch] gh/oulgen/17/head -> origin/gh/oulgen/17/head 2025-12-04T08:57:03.8979008Z * [new branch] gh/oulgen/17/orig -> origin/gh/oulgen/17/orig 2025-12-04T08:57:03.8981157Z * [new branch] gh/oulgen/18/base -> origin/gh/oulgen/18/base 2025-12-04T08:57:03.8982685Z * [new branch] gh/oulgen/18/head -> origin/gh/oulgen/18/head 2025-12-04T08:57:03.8984568Z * [new branch] gh/oulgen/18/orig -> origin/gh/oulgen/18/orig 2025-12-04T08:57:03.8986412Z * [new branch] gh/oulgen/19/base -> origin/gh/oulgen/19/base 2025-12-04T08:57:03.8987834Z * [new branch] gh/oulgen/19/head -> origin/gh/oulgen/19/head 2025-12-04T08:57:03.8989565Z * [new branch] gh/oulgen/19/orig -> origin/gh/oulgen/19/orig 2025-12-04T08:57:03.8991621Z * [new branch] gh/oulgen/20/base -> origin/gh/oulgen/20/base 2025-12-04T08:57:03.8993171Z * [new branch] gh/oulgen/20/head -> origin/gh/oulgen/20/head 2025-12-04T08:57:03.8994705Z * [new branch] gh/oulgen/20/orig -> origin/gh/oulgen/20/orig 2025-12-04T08:57:03.8996714Z * [new branch] gh/oulgen/21/base -> origin/gh/oulgen/21/base 2025-12-04T08:57:03.8998255Z * [new branch] gh/oulgen/21/head -> origin/gh/oulgen/21/head 2025-12-04T08:57:03.8999784Z * [new branch] gh/oulgen/21/orig -> origin/gh/oulgen/21/orig 2025-12-04T08:57:03.9001851Z * [new branch] gh/oulgen/22/base -> origin/gh/oulgen/22/base 2025-12-04T08:57:03.9003398Z * [new branch] gh/oulgen/22/head -> origin/gh/oulgen/22/head 2025-12-04T08:57:03.9004930Z * [new branch] gh/oulgen/22/orig -> origin/gh/oulgen/22/orig 2025-12-04T08:57:03.9007330Z * [new branch] gh/oulgen/23/base -> origin/gh/oulgen/23/base 2025-12-04T08:57:03.9008889Z * [new branch] gh/oulgen/23/head -> origin/gh/oulgen/23/head 2025-12-04T08:57:03.9010542Z * [new branch] gh/oulgen/23/orig -> origin/gh/oulgen/23/orig 2025-12-04T08:57:03.9012551Z * [new branch] gh/oulgen/24/base -> origin/gh/oulgen/24/base 2025-12-04T08:57:03.9014252Z * [new branch] gh/oulgen/24/head -> origin/gh/oulgen/24/head 2025-12-04T08:57:03.9015804Z * [new branch] gh/oulgen/24/orig -> origin/gh/oulgen/24/orig 2025-12-04T08:57:03.9017830Z * [new branch] gh/oulgen/25/base -> origin/gh/oulgen/25/base 2025-12-04T08:57:03.9019361Z * [new branch] gh/oulgen/25/head -> origin/gh/oulgen/25/head 2025-12-04T08:57:03.9020940Z * [new branch] gh/oulgen/25/orig -> origin/gh/oulgen/25/orig 2025-12-04T08:57:03.9022967Z * [new branch] gh/oulgen/26/base -> origin/gh/oulgen/26/base 2025-12-04T08:57:03.9024546Z * [new branch] gh/oulgen/26/head -> origin/gh/oulgen/26/head 2025-12-04T08:57:03.9026522Z * [new branch] gh/oulgen/26/orig -> origin/gh/oulgen/26/orig 2025-12-04T08:57:03.9028665Z * [new branch] gh/oulgen/4/base -> origin/gh/oulgen/4/base 2025-12-04T08:57:03.9030183Z * [new branch] gh/oulgen/4/head -> origin/gh/oulgen/4/head 2025-12-04T08:57:03.9031811Z * [new branch] gh/oulgen/4/orig -> origin/gh/oulgen/4/orig 2025-12-04T08:57:03.9034362Z * [new branch] gh/oulgen/7/base -> origin/gh/oulgen/7/base 2025-12-04T08:57:03.9035934Z * [new branch] gh/oulgen/7/head -> origin/gh/oulgen/7/head 2025-12-04T08:57:03.9037547Z * [new branch] gh/oulgen/7/orig -> origin/gh/oulgen/7/orig 2025-12-04T08:57:03.9039751Z * [new branch] gh/oulgen/8/base -> origin/gh/oulgen/8/base 2025-12-04T08:57:03.9041313Z * [new branch] gh/oulgen/8/head -> origin/gh/oulgen/8/head 2025-12-04T08:57:03.9042836Z * [new branch] gh/oulgen/8/orig -> origin/gh/oulgen/8/orig 2025-12-04T08:57:03.9044880Z * [new branch] gh/oulgen/9/base -> origin/gh/oulgen/9/base 2025-12-04T08:57:03.9046476Z * [new branch] gh/oulgen/9/head -> origin/gh/oulgen/9/head 2025-12-04T08:57:03.9048018Z * [new branch] gh/oulgen/9/orig -> origin/gh/oulgen/9/orig 2025-12-04T08:57:03.9050330Z * [new branch] gh/patvig/mtia-serialization -> origin/gh/patvig/mtia-serialization 2025-12-04T08:57:03.9052873Z * [new branch] gh/pearu/108/base -> origin/gh/pearu/108/base 2025-12-04T08:57:03.9054770Z * [new branch] gh/pearu/108/head -> origin/gh/pearu/108/head 2025-12-04T08:57:03.9056480Z * [new branch] gh/pearu/108/orig -> origin/gh/pearu/108/orig 2025-12-04T08:57:03.9058493Z * [new branch] gh/pearu/109/base -> origin/gh/pearu/109/base 2025-12-04T08:57:03.9060110Z * [new branch] gh/pearu/109/head -> origin/gh/pearu/109/head 2025-12-04T08:57:03.9061668Z * [new branch] gh/pearu/109/orig -> origin/gh/pearu/109/orig 2025-12-04T08:57:03.9063762Z * [new branch] gh/pearu/110/base -> origin/gh/pearu/110/base 2025-12-04T08:57:03.9065266Z * [new branch] gh/pearu/110/head -> origin/gh/pearu/110/head 2025-12-04T08:57:03.9066883Z * [new branch] gh/pearu/110/orig -> origin/gh/pearu/110/orig 2025-12-04T08:57:03.9068926Z * [new branch] gh/pearu/111/base -> origin/gh/pearu/111/base 2025-12-04T08:57:03.9070506Z * [new branch] gh/pearu/111/head -> origin/gh/pearu/111/head 2025-12-04T08:57:03.9072103Z * [new branch] gh/pearu/111/orig -> origin/gh/pearu/111/orig 2025-12-04T08:57:03.9074434Z * [new branch] gh/pearu/112/base -> origin/gh/pearu/112/base 2025-12-04T08:57:03.9076001Z * [new branch] gh/pearu/112/head -> origin/gh/pearu/112/head 2025-12-04T08:57:03.9078034Z * [new branch] gh/pearu/112/orig -> origin/gh/pearu/112/orig 2025-12-04T08:57:03.9080076Z * [new branch] gh/pearu/115/base -> origin/gh/pearu/115/base 2025-12-04T08:57:03.9081685Z * [new branch] gh/pearu/115/head -> origin/gh/pearu/115/head 2025-12-04T08:57:03.9083296Z * [new branch] gh/pearu/115/orig -> origin/gh/pearu/115/orig 2025-12-04T08:57:03.9085283Z * [new branch] gh/pearu/116/base -> origin/gh/pearu/116/base 2025-12-04T08:57:03.9086861Z * [new branch] gh/pearu/116/head -> origin/gh/pearu/116/head 2025-12-04T08:57:03.9088447Z * [new branch] gh/pearu/116/orig -> origin/gh/pearu/116/orig 2025-12-04T08:57:03.9090548Z * [new branch] gh/pearu/117/base -> origin/gh/pearu/117/base 2025-12-04T08:57:03.9092183Z * [new branch] gh/pearu/117/head -> origin/gh/pearu/117/head 2025-12-04T08:57:03.9093810Z * [new branch] gh/pearu/117/orig -> origin/gh/pearu/117/orig 2025-12-04T08:57:03.9096093Z * [new branch] gh/pearu/118/base -> origin/gh/pearu/118/base 2025-12-04T08:57:03.9098464Z * [new branch] gh/pearu/118/head -> origin/gh/pearu/118/head 2025-12-04T08:57:03.9100020Z * [new branch] gh/pearu/118/orig -> origin/gh/pearu/118/orig 2025-12-04T08:57:03.9102057Z * [new branch] gh/pearu/119/base -> origin/gh/pearu/119/base 2025-12-04T08:57:03.9103725Z * [new branch] gh/pearu/119/head -> origin/gh/pearu/119/head 2025-12-04T08:57:03.9105290Z * [new branch] gh/pearu/119/orig -> origin/gh/pearu/119/orig 2025-12-04T08:57:03.9107641Z * [new branch] gh/pearu/139/base -> origin/gh/pearu/139/base 2025-12-04T08:57:03.9109203Z * [new branch] gh/pearu/139/head -> origin/gh/pearu/139/head 2025-12-04T08:57:03.9110804Z * [new branch] gh/pearu/139/orig -> origin/gh/pearu/139/orig 2025-12-04T08:57:03.9112881Z * [new branch] gh/pearu/140/base -> origin/gh/pearu/140/base 2025-12-04T08:57:03.9114347Z * [new branch] gh/pearu/140/head -> origin/gh/pearu/140/head 2025-12-04T08:57:03.9116065Z * [new branch] gh/pearu/140/orig -> origin/gh/pearu/140/orig 2025-12-04T08:57:03.9118117Z * [new branch] gh/pearu/142/base -> origin/gh/pearu/142/base 2025-12-04T08:57:03.9119607Z * [new branch] gh/pearu/142/head -> origin/gh/pearu/142/head 2025-12-04T08:57:03.9121135Z * [new branch] gh/pearu/142/orig -> origin/gh/pearu/142/orig 2025-12-04T08:57:03.9123230Z * [new branch] gh/pearu/143/base -> origin/gh/pearu/143/base 2025-12-04T08:57:03.9124794Z * [new branch] gh/pearu/143/head -> origin/gh/pearu/143/head 2025-12-04T08:57:03.9126342Z * [new branch] gh/pearu/143/orig -> origin/gh/pearu/143/orig 2025-12-04T08:57:03.9128501Z * [new branch] gh/pearu/147/base -> origin/gh/pearu/147/base 2025-12-04T08:57:03.9130057Z * [new branch] gh/pearu/147/head -> origin/gh/pearu/147/head 2025-12-04T08:57:03.9131666Z * [new branch] gh/pearu/147/orig -> origin/gh/pearu/147/orig 2025-12-04T08:57:03.9133811Z * [new branch] gh/pearu/149/base -> origin/gh/pearu/149/base 2025-12-04T08:57:03.9135389Z * [new branch] gh/pearu/149/head -> origin/gh/pearu/149/head 2025-12-04T08:57:03.9136883Z * [new branch] gh/pearu/149/orig -> origin/gh/pearu/149/orig 2025-12-04T08:57:03.9139493Z * [new branch] gh/pearu/150/base -> origin/gh/pearu/150/base 2025-12-04T08:57:03.9141092Z * [new branch] gh/pearu/150/head -> origin/gh/pearu/150/head 2025-12-04T08:57:03.9142638Z * [new branch] gh/pearu/150/orig -> origin/gh/pearu/150/orig 2025-12-04T08:57:03.9144977Z * [new branch] gh/pearu/151/base -> origin/gh/pearu/151/base 2025-12-04T08:57:03.9146510Z * [new branch] gh/pearu/151/head -> origin/gh/pearu/151/head 2025-12-04T08:57:03.9148139Z * [new branch] gh/pearu/151/orig -> origin/gh/pearu/151/orig 2025-12-04T08:57:03.9150404Z * [new branch] gh/pearu/152/base -> origin/gh/pearu/152/base 2025-12-04T08:57:03.9151937Z * [new branch] gh/pearu/152/head -> origin/gh/pearu/152/head 2025-12-04T08:57:03.9153420Z * [new branch] gh/pearu/152/orig -> origin/gh/pearu/152/orig 2025-12-04T08:57:03.9155544Z * [new branch] gh/pearu/153/base -> origin/gh/pearu/153/base 2025-12-04T08:57:03.9157068Z * [new branch] gh/pearu/153/head -> origin/gh/pearu/153/head 2025-12-04T08:57:03.9158650Z * [new branch] gh/pearu/153/orig -> origin/gh/pearu/153/orig 2025-12-04T08:57:03.9160845Z * [new branch] gh/pearu/154/base -> origin/gh/pearu/154/base 2025-12-04T08:57:03.9162392Z * [new branch] gh/pearu/154/head -> origin/gh/pearu/154/head 2025-12-04T08:57:03.9164003Z * [new branch] gh/pearu/154/orig -> origin/gh/pearu/154/orig 2025-12-04T08:57:03.9166202Z * [new branch] gh/pearu/155/base -> origin/gh/pearu/155/base 2025-12-04T08:57:03.9167670Z * [new branch] gh/pearu/155/head -> origin/gh/pearu/155/head 2025-12-04T08:57:03.9169216Z * [new branch] gh/pearu/155/orig -> origin/gh/pearu/155/orig 2025-12-04T08:57:03.9171437Z * [new branch] gh/pearu/156/base -> origin/gh/pearu/156/base 2025-12-04T08:57:03.9173024Z * [new branch] gh/pearu/156/head -> origin/gh/pearu/156/head 2025-12-04T08:57:03.9174677Z * [new branch] gh/pearu/156/orig -> origin/gh/pearu/156/orig 2025-12-04T08:57:03.9177169Z * [new branch] gh/pearu/56/base -> origin/gh/pearu/56/base 2025-12-04T08:57:03.9178862Z * [new branch] gh/pearu/56/head -> origin/gh/pearu/56/head 2025-12-04T08:57:03.9180623Z * [new branch] gh/pearu/56/orig -> origin/gh/pearu/56/orig 2025-12-04T08:57:03.9182921Z * [new branch] gh/pearu/97/base -> origin/gh/pearu/97/base 2025-12-04T08:57:03.9184519Z * [new branch] gh/pearu/97/head -> origin/gh/pearu/97/head 2025-12-04T08:57:03.9186152Z * [new branch] gh/pearu/97/orig -> origin/gh/pearu/97/orig 2025-12-04T08:57:03.9188722Z * [new branch] gh/pianpwk/21/base -> origin/gh/pianpwk/21/base 2025-12-04T08:57:03.9190249Z * [new branch] gh/pianpwk/21/head -> origin/gh/pianpwk/21/head 2025-12-04T08:57:03.9192365Z * [new branch] gh/pianpwk/28/base -> origin/gh/pianpwk/28/base 2025-12-04T08:57:03.9194282Z * [new branch] gh/pianpwk/28/head -> origin/gh/pianpwk/28/head 2025-12-04T08:57:03.9195822Z * [new branch] gh/pianpwk/28/orig -> origin/gh/pianpwk/28/orig 2025-12-04T08:57:03.9197946Z * [new branch] gh/pianpwk/29/base -> origin/gh/pianpwk/29/base 2025-12-04T08:57:03.9199658Z * [new branch] gh/pianpwk/29/head -> origin/gh/pianpwk/29/head 2025-12-04T08:57:03.9201196Z * [new branch] gh/pianpwk/29/orig -> origin/gh/pianpwk/29/orig 2025-12-04T08:57:03.9203483Z * [new branch] gh/pianpwk/30/base -> origin/gh/pianpwk/30/base 2025-12-04T08:57:03.9204921Z * [new branch] gh/pianpwk/30/head -> origin/gh/pianpwk/30/head 2025-12-04T08:57:03.9206820Z * [new branch] gh/pianpwk/30/orig -> origin/gh/pianpwk/30/orig 2025-12-04T08:57:03.9208909Z * [new branch] gh/pianpwk/31/base -> origin/gh/pianpwk/31/base 2025-12-04T08:57:03.9210527Z * [new branch] gh/pianpwk/31/head -> origin/gh/pianpwk/31/head 2025-12-04T08:57:03.9212080Z * [new branch] gh/pianpwk/31/orig -> origin/gh/pianpwk/31/orig 2025-12-04T08:57:03.9214180Z * [new branch] gh/pianpwk/32/base -> origin/gh/pianpwk/32/base 2025-12-04T08:57:03.9215830Z * [new branch] gh/pianpwk/32/head -> origin/gh/pianpwk/32/head 2025-12-04T08:57:03.9217299Z * [new branch] gh/pianpwk/32/orig -> origin/gh/pianpwk/32/orig 2025-12-04T08:57:03.9219225Z * [new branch] gh/pianpwk/33/base -> origin/gh/pianpwk/33/base 2025-12-04T08:57:03.9220748Z * [new branch] gh/pianpwk/33/head -> origin/gh/pianpwk/33/head 2025-12-04T08:57:03.9222325Z * [new branch] gh/pianpwk/33/orig -> origin/gh/pianpwk/33/orig 2025-12-04T08:57:03.9224701Z * [new branch] gh/pianpwk/34/base -> origin/gh/pianpwk/34/base 2025-12-04T08:57:03.9226450Z * [new branch] gh/pianpwk/34/head -> origin/gh/pianpwk/34/head 2025-12-04T08:57:03.9228123Z * [new branch] gh/pianpwk/34/orig -> origin/gh/pianpwk/34/orig 2025-12-04T08:57:03.9230188Z * [new branch] gh/pianpwk/35/base -> origin/gh/pianpwk/35/base 2025-12-04T08:57:03.9231898Z * [new branch] gh/pianpwk/35/head -> origin/gh/pianpwk/35/head 2025-12-04T08:57:03.9233533Z * [new branch] gh/pianpwk/35/orig -> origin/gh/pianpwk/35/orig 2025-12-04T08:57:03.9236047Z * [new branch] gh/rec/141/base -> origin/gh/rec/141/base 2025-12-04T08:57:03.9237571Z * [new branch] gh/rec/141/head -> origin/gh/rec/141/head 2025-12-04T08:57:03.9239809Z * [new branch] gh/rec/153/base -> origin/gh/rec/153/base 2025-12-04T08:57:03.9241304Z * [new branch] gh/rec/153/head -> origin/gh/rec/153/head 2025-12-04T08:57:03.9242940Z * [new branch] gh/rec/153/orig -> origin/gh/rec/153/orig 2025-12-04T08:57:03.9245033Z * [new branch] gh/rec/154/base -> origin/gh/rec/154/base 2025-12-04T08:57:03.9246745Z * [new branch] gh/rec/154/head -> origin/gh/rec/154/head 2025-12-04T08:57:03.9248165Z * [new branch] gh/rec/154/orig -> origin/gh/rec/154/orig 2025-12-04T08:57:03.9250283Z * [new branch] gh/rec/164/base -> origin/gh/rec/164/base 2025-12-04T08:57:03.9251808Z * [new branch] gh/rec/164/head -> origin/gh/rec/164/head 2025-12-04T08:57:03.9253424Z * [new branch] gh/rec/164/orig -> origin/gh/rec/164/orig 2025-12-04T08:57:03.9255694Z * [new branch] gh/rec/166/base -> origin/gh/rec/166/base 2025-12-04T08:57:03.9257611Z * [new branch] gh/rec/166/head -> origin/gh/rec/166/head 2025-12-04T08:57:03.9259227Z * [new branch] gh/rec/166/orig -> origin/gh/rec/166/orig 2025-12-04T08:57:03.9261295Z * [new branch] gh/rec/167/base -> origin/gh/rec/167/base 2025-12-04T08:57:03.9262825Z * [new branch] gh/rec/167/head -> origin/gh/rec/167/head 2025-12-04T08:57:03.9264296Z * [new branch] gh/rec/167/orig -> origin/gh/rec/167/orig 2025-12-04T08:57:03.9267006Z * [new branch] gh/rec/168/base -> origin/gh/rec/168/base 2025-12-04T08:57:03.9268612Z * [new branch] gh/rec/168/head -> origin/gh/rec/168/head 2025-12-04T08:57:03.9270184Z * [new branch] gh/rec/168/orig -> origin/gh/rec/168/orig 2025-12-04T08:57:03.9272302Z * [new branch] gh/rec/169/base -> origin/gh/rec/169/base 2025-12-04T08:57:03.9273913Z * [new branch] gh/rec/169/head -> origin/gh/rec/169/head 2025-12-04T08:57:03.9275426Z * [new branch] gh/rec/169/orig -> origin/gh/rec/169/orig 2025-12-04T08:57:03.9277567Z * [new branch] gh/rec/170/base -> origin/gh/rec/170/base 2025-12-04T08:57:03.9279097Z * [new branch] gh/rec/170/head -> origin/gh/rec/170/head 2025-12-04T08:57:03.9280689Z * [new branch] gh/rec/170/orig -> origin/gh/rec/170/orig 2025-12-04T08:57:03.9282738Z * [new branch] gh/rec/171/base -> origin/gh/rec/171/base 2025-12-04T08:57:03.9284306Z * [new branch] gh/rec/171/head -> origin/gh/rec/171/head 2025-12-04T08:57:03.9285899Z * [new branch] gh/rec/171/orig -> origin/gh/rec/171/orig 2025-12-04T08:57:03.9288041Z * [new branch] gh/rec/172/base -> origin/gh/rec/172/base 2025-12-04T08:57:03.9289582Z * [new branch] gh/rec/172/head -> origin/gh/rec/172/head 2025-12-04T08:57:03.9291076Z * [new branch] gh/rec/172/orig -> origin/gh/rec/172/orig 2025-12-04T08:57:03.9293166Z * [new branch] gh/rec/173/base -> origin/gh/rec/173/base 2025-12-04T08:57:03.9295027Z * [new branch] gh/rec/173/head -> origin/gh/rec/173/head 2025-12-04T08:57:03.9296570Z * [new branch] gh/rec/173/orig -> origin/gh/rec/173/orig 2025-12-04T08:57:03.9298592Z * [new branch] gh/rec/174/base -> origin/gh/rec/174/base 2025-12-04T08:57:03.9300133Z * [new branch] gh/rec/174/head -> origin/gh/rec/174/head 2025-12-04T08:57:03.9301682Z * [new branch] gh/rec/174/orig -> origin/gh/rec/174/orig 2025-12-04T08:57:03.9303737Z * [new branch] gh/rec/175/base -> origin/gh/rec/175/base 2025-12-04T08:57:03.9306975Z * [new branch] gh/rec/175/head -> origin/gh/rec/175/head 2025-12-04T08:57:03.9308572Z * [new branch] gh/rec/175/orig -> origin/gh/rec/175/orig 2025-12-04T08:57:03.9310703Z * [new branch] gh/rec/176/base -> origin/gh/rec/176/base 2025-12-04T08:57:03.9312247Z * [new branch] gh/rec/176/head -> origin/gh/rec/176/head 2025-12-04T08:57:03.9313861Z * [new branch] gh/rec/176/orig -> origin/gh/rec/176/orig 2025-12-04T08:57:03.9316271Z * [new branch] gh/rec/177/base -> origin/gh/rec/177/base 2025-12-04T08:57:03.9317830Z * [new branch] gh/rec/177/head -> origin/gh/rec/177/head 2025-12-04T08:57:03.9319341Z * [new branch] gh/rec/177/orig -> origin/gh/rec/177/orig 2025-12-04T08:57:03.9321892Z * [new branch] gh/robert-hardwick/3/base -> origin/gh/robert-hardwick/3/base 2025-12-04T08:57:03.9323486Z * [new branch] gh/robert-hardwick/3/head -> origin/gh/robert-hardwick/3/head 2025-12-04T08:57:03.9325113Z * [new branch] gh/robert-hardwick/3/orig -> origin/gh/robert-hardwick/3/orig 2025-12-04T08:57:03.9327171Z * [new branch] gh/robert-hardwick/4/base -> origin/gh/robert-hardwick/4/base 2025-12-04T08:57:03.9328705Z * [new branch] gh/robert-hardwick/4/head -> origin/gh/robert-hardwick/4/head 2025-12-04T08:57:03.9330370Z * [new branch] gh/robert-hardwick/4/orig -> origin/gh/robert-hardwick/4/orig 2025-12-04T08:57:03.9332429Z * [new branch] gh/robert-hardwick/5/base -> origin/gh/robert-hardwick/5/base 2025-12-04T08:57:03.9334122Z * [new branch] gh/robert-hardwick/5/head -> origin/gh/robert-hardwick/5/head 2025-12-04T08:57:03.9335782Z * [new branch] gh/robert-hardwick/5/orig -> origin/gh/robert-hardwick/5/orig 2025-12-04T08:57:03.9337873Z * [new branch] gh/robert-hardwick/6/base -> origin/gh/robert-hardwick/6/base 2025-12-04T08:57:03.9339892Z * [new branch] gh/robert-hardwick/6/head -> origin/gh/robert-hardwick/6/head 2025-12-04T08:57:03.9341459Z * [new branch] gh/robert-hardwick/6/orig -> origin/gh/robert-hardwick/6/orig 2025-12-04T08:57:03.9343552Z * [new branch] gh/robert-hardwick/7/base -> origin/gh/robert-hardwick/7/base 2025-12-04T08:57:03.9345088Z * [new branch] gh/robert-hardwick/7/head -> origin/gh/robert-hardwick/7/head 2025-12-04T08:57:03.9346674Z * [new branch] gh/robert-hardwick/7/orig -> origin/gh/robert-hardwick/7/orig 2025-12-04T08:57:03.9348806Z * [new branch] gh/robert-hardwick/8/base -> origin/gh/robert-hardwick/8/base 2025-12-04T08:57:03.9350406Z * [new branch] gh/robert-hardwick/8/head -> origin/gh/robert-hardwick/8/head 2025-12-04T08:57:03.9352020Z * [new branch] gh/robert-hardwick/8/orig -> origin/gh/robert-hardwick/8/orig 2025-12-04T08:57:03.9354188Z * [new branch] gh/robert-hardwick/9/base -> origin/gh/robert-hardwick/9/base 2025-12-04T08:57:03.9355781Z * [new branch] gh/robert-hardwick/9/head -> origin/gh/robert-hardwick/9/head 2025-12-04T08:57:03.9357371Z * [new branch] gh/robert-hardwick/9/orig -> origin/gh/robert-hardwick/9/orig 2025-12-04T08:57:03.9359978Z * [new branch] gh/rtimpe/1/base -> origin/gh/rtimpe/1/base 2025-12-04T08:57:03.9361502Z * [new branch] gh/rtimpe/1/head -> origin/gh/rtimpe/1/head 2025-12-04T08:57:03.9363490Z * [new branch] gh/rtimpe/2/base -> origin/gh/rtimpe/2/base 2025-12-04T08:57:03.9364961Z * [new branch] gh/rtimpe/2/head -> origin/gh/rtimpe/2/head 2025-12-04T08:57:03.9367109Z * [new branch] gh/rtimpe/22/base -> origin/gh/rtimpe/22/base 2025-12-04T08:57:03.9368642Z * [new branch] gh/rtimpe/22/head -> origin/gh/rtimpe/22/head 2025-12-04T08:57:03.9370207Z * [new branch] gh/rtimpe/22/orig -> origin/gh/rtimpe/22/orig 2025-12-04T08:57:03.9372276Z * [new branch] gh/rtimpe/23/base -> origin/gh/rtimpe/23/base 2025-12-04T08:57:03.9374087Z * [new branch] gh/rtimpe/23/head -> origin/gh/rtimpe/23/head 2025-12-04T08:57:03.9375740Z * [new branch] gh/rtimpe/23/orig -> origin/gh/rtimpe/23/orig 2025-12-04T08:57:03.9377674Z * [new branch] gh/rtimpe/24/base -> origin/gh/rtimpe/24/base 2025-12-04T08:57:03.9379187Z * [new branch] gh/rtimpe/24/head -> origin/gh/rtimpe/24/head 2025-12-04T08:57:03.9380763Z * [new branch] gh/rtimpe/24/orig -> origin/gh/rtimpe/24/orig 2025-12-04T08:57:03.9382790Z * [new branch] gh/rtimpe/25/base -> origin/gh/rtimpe/25/base 2025-12-04T08:57:03.9384311Z * [new branch] gh/rtimpe/25/head -> origin/gh/rtimpe/25/head 2025-12-04T08:57:03.9385891Z * [new branch] gh/rtimpe/25/orig -> origin/gh/rtimpe/25/orig 2025-12-04T08:57:03.9387985Z * [new branch] gh/rtimpe/26/base -> origin/gh/rtimpe/26/base 2025-12-04T08:57:03.9389511Z * [new branch] gh/rtimpe/26/head -> origin/gh/rtimpe/26/head 2025-12-04T08:57:03.9391045Z * [new branch] gh/rtimpe/26/orig -> origin/gh/rtimpe/26/orig 2025-12-04T08:57:03.9393161Z * [new branch] gh/rtimpe/27/base -> origin/gh/rtimpe/27/base 2025-12-04T08:57:03.9394817Z * [new branch] gh/rtimpe/27/head -> origin/gh/rtimpe/27/head 2025-12-04T08:57:03.9396269Z * [new branch] gh/rtimpe/27/orig -> origin/gh/rtimpe/27/orig 2025-12-04T08:57:03.9398352Z * [new branch] gh/rtimpe/28/base -> origin/gh/rtimpe/28/base 2025-12-04T08:57:03.9399876Z * [new branch] gh/rtimpe/28/head -> origin/gh/rtimpe/28/head 2025-12-04T08:57:03.9401508Z * [new branch] gh/rtimpe/28/orig -> origin/gh/rtimpe/28/orig 2025-12-04T08:57:03.9403688Z * [new branch] gh/rtimpe/29/base -> origin/gh/rtimpe/29/base 2025-12-04T08:57:03.9405092Z * [new branch] gh/rtimpe/29/head -> origin/gh/rtimpe/29/head 2025-12-04T08:57:03.9407103Z * [new branch] gh/rtimpe/29/orig -> origin/gh/rtimpe/29/orig 2025-12-04T08:57:03.9409157Z * [new branch] gh/rtimpe/3/base -> origin/gh/rtimpe/3/base 2025-12-04T08:57:03.9410655Z * [new branch] gh/rtimpe/3/head -> origin/gh/rtimpe/3/head 2025-12-04T08:57:03.9412777Z * [new branch] gh/rtimpe/30/base -> origin/gh/rtimpe/30/base 2025-12-04T08:57:03.9414405Z * [new branch] gh/rtimpe/30/head -> origin/gh/rtimpe/30/head 2025-12-04T08:57:03.9415965Z * [new branch] gh/rtimpe/30/orig -> origin/gh/rtimpe/30/orig 2025-12-04T08:57:03.9417967Z * [new branch] gh/rtimpe/31/base -> origin/gh/rtimpe/31/base 2025-12-04T08:57:03.9419546Z * [new branch] gh/rtimpe/31/head -> origin/gh/rtimpe/31/head 2025-12-04T08:57:03.9421212Z * [new branch] gh/rtimpe/31/orig -> origin/gh/rtimpe/31/orig 2025-12-04T08:57:03.9423274Z * [new branch] gh/rtimpe/32/base -> origin/gh/rtimpe/32/base 2025-12-04T08:57:03.9424958Z * [new branch] gh/rtimpe/32/head -> origin/gh/rtimpe/32/head 2025-12-04T08:57:03.9426490Z * [new branch] gh/rtimpe/32/orig -> origin/gh/rtimpe/32/orig 2025-12-04T08:57:03.9428548Z * [new branch] gh/rtimpe/33/base -> origin/gh/rtimpe/33/base 2025-12-04T08:57:03.9430152Z * [new branch] gh/rtimpe/33/head -> origin/gh/rtimpe/33/head 2025-12-04T08:57:03.9431701Z * [new branch] gh/rtimpe/33/orig -> origin/gh/rtimpe/33/orig 2025-12-04T08:57:03.9433748Z * [new branch] gh/rtimpe/34/base -> origin/gh/rtimpe/34/base 2025-12-04T08:57:03.9435425Z * [new branch] gh/rtimpe/34/head -> origin/gh/rtimpe/34/head 2025-12-04T08:57:03.9437340Z * [new branch] gh/rtimpe/34/orig -> origin/gh/rtimpe/34/orig 2025-12-04T08:57:03.9439644Z * [new branch] gh/rtimpe/35/base -> origin/gh/rtimpe/35/base 2025-12-04T08:57:03.9440925Z * [new branch] gh/rtimpe/35/head -> origin/gh/rtimpe/35/head 2025-12-04T08:57:03.9442658Z * [new branch] gh/rtimpe/35/orig -> origin/gh/rtimpe/35/orig 2025-12-04T08:57:03.9444737Z * [new branch] gh/rtimpe/4/base -> origin/gh/rtimpe/4/base 2025-12-04T08:57:03.9446251Z * [new branch] gh/rtimpe/4/head -> origin/gh/rtimpe/4/head 2025-12-04T08:57:03.9448828Z * [new branch] gh/ruisizhang123/1/base -> origin/gh/ruisizhang123/1/base 2025-12-04T08:57:03.9450382Z * [new branch] gh/ruisizhang123/1/head -> origin/gh/ruisizhang123/1/head 2025-12-04T08:57:03.9452073Z * [new branch] gh/ruisizhang123/1/orig -> origin/gh/ruisizhang123/1/orig 2025-12-04T08:57:03.9454412Z * [new branch] gh/ruisizhang123/4/base -> origin/gh/ruisizhang123/4/base 2025-12-04T08:57:03.9456120Z * [new branch] gh/ruisizhang123/4/head -> origin/gh/ruisizhang123/4/head 2025-12-04T08:57:03.9457644Z * [new branch] gh/ruisizhang123/4/orig -> origin/gh/ruisizhang123/4/orig 2025-12-04T08:57:03.9459776Z * [new branch] gh/ruisizhang123/5/base -> origin/gh/ruisizhang123/5/base 2025-12-04T08:57:03.9461330Z * [new branch] gh/ruisizhang123/5/head -> origin/gh/ruisizhang123/5/head 2025-12-04T08:57:03.9462855Z * [new branch] gh/ruisizhang123/5/orig -> origin/gh/ruisizhang123/5/orig 2025-12-04T08:57:03.9464961Z * [new branch] gh/ruisizhang123/6/base -> origin/gh/ruisizhang123/6/base 2025-12-04T08:57:03.9466474Z * [new branch] gh/ruisizhang123/6/head -> origin/gh/ruisizhang123/6/head 2025-12-04T08:57:03.9468004Z * [new branch] gh/ruisizhang123/6/orig -> origin/gh/ruisizhang123/6/orig 2025-12-04T08:57:03.9470066Z * [new branch] gh/ruisizhang123/7/base -> origin/gh/ruisizhang123/7/base 2025-12-04T08:57:03.9471713Z * [new branch] gh/ruisizhang123/7/head -> origin/gh/ruisizhang123/7/head 2025-12-04T08:57:03.9473232Z * [new branch] gh/ruisizhang123/7/orig -> origin/gh/ruisizhang123/7/orig 2025-12-04T08:57:03.9475298Z * [new branch] gh/ruisizhang123/8/base -> origin/gh/ruisizhang123/8/base 2025-12-04T08:57:03.9476872Z * [new branch] gh/ruisizhang123/8/head -> origin/gh/ruisizhang123/8/head 2025-12-04T08:57:03.9478410Z * [new branch] gh/ruisizhang123/8/orig -> origin/gh/ruisizhang123/8/orig 2025-12-04T08:57:03.9480854Z * [new branch] gh/ruisizhang123/9/base -> origin/gh/ruisizhang123/9/base 2025-12-04T08:57:03.9482434Z * [new branch] gh/ruisizhang123/9/head -> origin/gh/ruisizhang123/9/head 2025-12-04T08:57:03.9484001Z * [new branch] gh/ruisizhang123/9/orig -> origin/gh/ruisizhang123/9/orig 2025-12-04T08:57:03.9486683Z * [new branch] gh/seemethere/52/base -> origin/gh/seemethere/52/base 2025-12-04T08:57:03.9488175Z * [new branch] gh/seemethere/52/head -> origin/gh/seemethere/52/head 2025-12-04T08:57:03.9489732Z * [new branch] gh/seemethere/52/orig -> origin/gh/seemethere/52/orig 2025-12-04T08:57:03.9491778Z * [new branch] gh/seemethere/53/base -> origin/gh/seemethere/53/base 2025-12-04T08:57:03.9493376Z * [new branch] gh/seemethere/53/head -> origin/gh/seemethere/53/head 2025-12-04T08:57:03.9495015Z * [new branch] gh/seemethere/53/orig -> origin/gh/seemethere/53/orig 2025-12-04T08:57:03.9497189Z * [new branch] gh/seemethere/54/base -> origin/gh/seemethere/54/base 2025-12-04T08:57:03.9498788Z * [new branch] gh/seemethere/54/head -> origin/gh/seemethere/54/head 2025-12-04T08:57:03.9500371Z * [new branch] gh/seemethere/54/orig -> origin/gh/seemethere/54/orig 2025-12-04T08:57:03.9502496Z * [new branch] gh/seemethere/55/base -> origin/gh/seemethere/55/base 2025-12-04T08:57:03.9503988Z * [new branch] gh/seemethere/55/head -> origin/gh/seemethere/55/head 2025-12-04T08:57:03.9505828Z * [new branch] gh/seemethere/55/orig -> origin/gh/seemethere/55/orig 2025-12-04T08:57:03.9507866Z * [new branch] gh/seemethere/59/base -> origin/gh/seemethere/59/base 2025-12-04T08:57:03.9509401Z * [new branch] gh/seemethere/59/head -> origin/gh/seemethere/59/head 2025-12-04T08:57:03.9510942Z * [new branch] gh/seemethere/59/orig -> origin/gh/seemethere/59/orig 2025-12-04T08:57:03.9513027Z * [new branch] gh/seemethere/62/base -> origin/gh/seemethere/62/base 2025-12-04T08:57:03.9514580Z * [new branch] gh/seemethere/62/head -> origin/gh/seemethere/62/head 2025-12-04T08:57:03.9516144Z * [new branch] gh/seemethere/62/orig -> origin/gh/seemethere/62/orig 2025-12-04T08:57:03.9518288Z * [new branch] gh/seemethere/63/base -> origin/gh/seemethere/63/base 2025-12-04T08:57:03.9519829Z * [new branch] gh/seemethere/63/head -> origin/gh/seemethere/63/head 2025-12-04T08:57:03.9521464Z * [new branch] gh/seemethere/63/orig -> origin/gh/seemethere/63/orig 2025-12-04T08:57:03.9523553Z * [new branch] gh/seemethere/71/base -> origin/gh/seemethere/71/base 2025-12-04T08:57:03.9525510Z * [new branch] gh/seemethere/71/head -> origin/gh/seemethere/71/head 2025-12-04T08:57:03.9527095Z * [new branch] gh/seemethere/71/orig -> origin/gh/seemethere/71/orig 2025-12-04T08:57:03.9529196Z * [new branch] gh/seemethere/72/base -> origin/gh/seemethere/72/base 2025-12-04T08:57:03.9530719Z * [new branch] gh/seemethere/72/head -> origin/gh/seemethere/72/head 2025-12-04T08:57:03.9532307Z * [new branch] gh/seemethere/72/orig -> origin/gh/seemethere/72/orig 2025-12-04T08:57:03.9534475Z * [new branch] gh/seemethere/73/base -> origin/gh/seemethere/73/base 2025-12-04T08:57:03.9536019Z * [new branch] gh/seemethere/73/head -> origin/gh/seemethere/73/head 2025-12-04T08:57:03.9537663Z * [new branch] gh/seemethere/73/orig -> origin/gh/seemethere/73/orig 2025-12-04T08:57:03.9540289Z * [new branch] gh/seemethere/74/base -> origin/gh/seemethere/74/base 2025-12-04T08:57:03.9541893Z * [new branch] gh/seemethere/74/head -> origin/gh/seemethere/74/head 2025-12-04T08:57:03.9543537Z * [new branch] gh/seemethere/74/orig -> origin/gh/seemethere/74/orig 2025-12-04T08:57:03.9545586Z * [new branch] gh/seemethere/75/base -> origin/gh/seemethere/75/base 2025-12-04T08:57:03.9547156Z * [new branch] gh/seemethere/75/head -> origin/gh/seemethere/75/head 2025-12-04T08:57:03.9548756Z * [new branch] gh/seemethere/75/orig -> origin/gh/seemethere/75/orig 2025-12-04T08:57:03.9550812Z * [new branch] gh/seemethere/76/base -> origin/gh/seemethere/76/base 2025-12-04T08:57:03.9552362Z * [new branch] gh/seemethere/76/head -> origin/gh/seemethere/76/head 2025-12-04T08:57:03.9553870Z * [new branch] gh/seemethere/76/orig -> origin/gh/seemethere/76/orig 2025-12-04T08:57:03.9556611Z * [new branch] gh/shunting314/145/base -> origin/gh/shunting314/145/base 2025-12-04T08:57:03.9558333Z * [new branch] gh/shunting314/145/head -> origin/gh/shunting314/145/head 2025-12-04T08:57:03.9560014Z * [new branch] gh/shunting314/145/orig -> origin/gh/shunting314/145/orig 2025-12-04T08:57:03.9562295Z * [new branch] gh/shunting314/176/base -> origin/gh/shunting314/176/base 2025-12-04T08:57:03.9564180Z * [new branch] gh/shunting314/176/head -> origin/gh/shunting314/176/head 2025-12-04T08:57:03.9565637Z * [new branch] gh/shunting314/176/orig -> origin/gh/shunting314/176/orig 2025-12-04T08:57:03.9567867Z * [new branch] gh/shunting314/249/base -> origin/gh/shunting314/249/base 2025-12-04T08:57:03.9569462Z * [new branch] gh/shunting314/249/head -> origin/gh/shunting314/249/head 2025-12-04T08:57:03.9571056Z * [new branch] gh/shunting314/249/orig -> origin/gh/shunting314/249/orig 2025-12-04T08:57:03.9573251Z * [new branch] gh/shunting314/253/base -> origin/gh/shunting314/253/base 2025-12-04T08:57:03.9574770Z * [new branch] gh/shunting314/253/head -> origin/gh/shunting314/253/head 2025-12-04T08:57:03.9576325Z * [new branch] gh/shunting314/253/orig -> origin/gh/shunting314/253/orig 2025-12-04T08:57:03.9578451Z * [new branch] gh/shunting314/256/base -> origin/gh/shunting314/256/base 2025-12-04T08:57:03.9580001Z * [new branch] gh/shunting314/256/head -> origin/gh/shunting314/256/head 2025-12-04T08:57:03.9581609Z * [new branch] gh/shunting314/256/orig -> origin/gh/shunting314/256/orig 2025-12-04T08:57:03.9583880Z * [new branch] gh/shunting314/257/base -> origin/gh/shunting314/257/base 2025-12-04T08:57:03.9585600Z * [new branch] gh/shunting314/257/head -> origin/gh/shunting314/257/head 2025-12-04T08:57:03.9587180Z * [new branch] gh/shunting314/257/orig -> origin/gh/shunting314/257/orig 2025-12-04T08:57:03.9589377Z * [new branch] gh/shunting314/258/base -> origin/gh/shunting314/258/base 2025-12-04T08:57:03.9590954Z * [new branch] gh/shunting314/258/head -> origin/gh/shunting314/258/head 2025-12-04T08:57:03.9592549Z * [new branch] gh/shunting314/258/orig -> origin/gh/shunting314/258/orig 2025-12-04T08:57:03.9594470Z * [new branch] gh/shunting314/259/base -> origin/gh/shunting314/259/base 2025-12-04T08:57:03.9596028Z * [new branch] gh/shunting314/259/head -> origin/gh/shunting314/259/head 2025-12-04T08:57:03.9597601Z * [new branch] gh/shunting314/259/orig -> origin/gh/shunting314/259/orig 2025-12-04T08:57:03.9599772Z * [new branch] gh/shunting314/260/base -> origin/gh/shunting314/260/base 2025-12-04T08:57:03.9601311Z * [new branch] gh/shunting314/260/head -> origin/gh/shunting314/260/head 2025-12-04T08:57:03.9603090Z * [new branch] gh/shunting314/260/orig -> origin/gh/shunting314/260/orig 2025-12-04T08:57:03.9605511Z * [new branch] gh/shunting314/261/base -> origin/gh/shunting314/261/base 2025-12-04T08:57:03.9607121Z * [new branch] gh/shunting314/261/head -> origin/gh/shunting314/261/head 2025-12-04T08:57:03.9608851Z * [new branch] gh/shunting314/261/orig -> origin/gh/shunting314/261/orig 2025-12-04T08:57:03.9611011Z * [new branch] gh/shunting314/262/base -> origin/gh/shunting314/262/base 2025-12-04T08:57:03.9612684Z * [new branch] gh/shunting314/262/head -> origin/gh/shunting314/262/head 2025-12-04T08:57:03.9614469Z * [new branch] gh/shunting314/262/orig -> origin/gh/shunting314/262/orig 2025-12-04T08:57:03.9616572Z * [new branch] gh/shunting314/263/base -> origin/gh/shunting314/263/base 2025-12-04T08:57:03.9618259Z * [new branch] gh/shunting314/263/head -> origin/gh/shunting314/263/head 2025-12-04T08:57:03.9619869Z * [new branch] gh/shunting314/263/orig -> origin/gh/shunting314/263/orig 2025-12-04T08:57:03.9621964Z * [new branch] gh/shunting314/264/base -> origin/gh/shunting314/264/base 2025-12-04T08:57:03.9623522Z * [new branch] gh/shunting314/264/head -> origin/gh/shunting314/264/head 2025-12-04T08:57:03.9625433Z * [new branch] gh/shunting314/264/orig -> origin/gh/shunting314/264/orig 2025-12-04T08:57:03.9627386Z * [new branch] gh/shunting314/265/base -> origin/gh/shunting314/265/base 2025-12-04T08:57:03.9628951Z * [new branch] gh/shunting314/265/head -> origin/gh/shunting314/265/head 2025-12-04T08:57:03.9630465Z * [new branch] gh/shunting314/265/orig -> origin/gh/shunting314/265/orig 2025-12-04T08:57:03.9632531Z * [new branch] gh/shunting314/266/base -> origin/gh/shunting314/266/base 2025-12-04T08:57:03.9634209Z * [new branch] gh/shunting314/266/head -> origin/gh/shunting314/266/head 2025-12-04T08:57:03.9635722Z * [new branch] gh/shunting314/266/orig -> origin/gh/shunting314/266/orig 2025-12-04T08:57:03.9638041Z * [new branch] gh/shunting314/267/base -> origin/gh/shunting314/267/base 2025-12-04T08:57:03.9639758Z * [new branch] gh/shunting314/267/head -> origin/gh/shunting314/267/head 2025-12-04T08:57:03.9641373Z * [new branch] gh/shunting314/267/orig -> origin/gh/shunting314/267/orig 2025-12-04T08:57:03.9643944Z * [new branch] gh/shunting314/268/base -> origin/gh/shunting314/268/base 2025-12-04T08:57:03.9645542Z * [new branch] gh/shunting314/268/head -> origin/gh/shunting314/268/head 2025-12-04T08:57:03.9647238Z * [new branch] gh/shunting314/268/orig -> origin/gh/shunting314/268/orig 2025-12-04T08:57:03.9649943Z * [new branch] gh/shunting314/269/base -> origin/gh/shunting314/269/base 2025-12-04T08:57:03.9651509Z * [new branch] gh/shunting314/269/head -> origin/gh/shunting314/269/head 2025-12-04T08:57:03.9653225Z * [new branch] gh/shunting314/269/orig -> origin/gh/shunting314/269/orig 2025-12-04T08:57:03.9655848Z * [new branch] gh/silverguo/1/base -> origin/gh/silverguo/1/base 2025-12-04T08:57:03.9657407Z * [new branch] gh/silverguo/1/head -> origin/gh/silverguo/1/head 2025-12-04T08:57:03.9659336Z * [new branch] gh/silverguo/2/base -> origin/gh/silverguo/2/base 2025-12-04T08:57:03.9660856Z * [new branch] gh/silverguo/2/head -> origin/gh/silverguo/2/head 2025-12-04T08:57:03.9662805Z * [new branch] gh/silverguo/3/base -> origin/gh/silverguo/3/base 2025-12-04T08:57:03.9664405Z * [new branch] gh/silverguo/3/head -> origin/gh/silverguo/3/head 2025-12-04T08:57:03.9666446Z * [new branch] gh/silverguo/4/base -> origin/gh/silverguo/4/base 2025-12-04T08:57:03.9667965Z * [new branch] gh/silverguo/4/head -> origin/gh/silverguo/4/head 2025-12-04T08:57:03.9670485Z * [new branch] gh/slayton58/39/base -> origin/gh/slayton58/39/base 2025-12-04T08:57:03.9672086Z * [new branch] gh/slayton58/39/head -> origin/gh/slayton58/39/head 2025-12-04T08:57:03.9673784Z * [new branch] gh/slayton58/39/orig -> origin/gh/slayton58/39/orig 2025-12-04T08:57:03.9675864Z * [new branch] gh/slayton58/42/base -> origin/gh/slayton58/42/base 2025-12-04T08:57:03.9677472Z * [new branch] gh/slayton58/42/head -> origin/gh/slayton58/42/head 2025-12-04T08:57:03.9679070Z * [new branch] gh/slayton58/42/orig -> origin/gh/slayton58/42/orig 2025-12-04T08:57:03.9681171Z * [new branch] gh/slayton58/43/base -> origin/gh/slayton58/43/base 2025-12-04T08:57:03.9682739Z * [new branch] gh/slayton58/43/head -> origin/gh/slayton58/43/head 2025-12-04T08:57:03.9684298Z * [new branch] gh/slayton58/43/orig -> origin/gh/slayton58/43/orig 2025-12-04T08:57:03.9686524Z * [new branch] gh/slayton58/44/base -> origin/gh/slayton58/44/base 2025-12-04T08:57:03.9688094Z * [new branch] gh/slayton58/44/head -> origin/gh/slayton58/44/head 2025-12-04T08:57:03.9689773Z * [new branch] gh/slayton58/44/orig -> origin/gh/slayton58/44/orig 2025-12-04T08:57:03.9691776Z * [new branch] gh/slayton58/45/base -> origin/gh/slayton58/45/base 2025-12-04T08:57:03.9693333Z * [new branch] gh/slayton58/45/head -> origin/gh/slayton58/45/head 2025-12-04T08:57:03.9694930Z * [new branch] gh/slayton58/45/orig -> origin/gh/slayton58/45/orig 2025-12-04T08:57:03.9697504Z * [new branch] gh/slayton58/46/base -> origin/gh/slayton58/46/base 2025-12-04T08:57:03.9699203Z * [new branch] gh/slayton58/46/head -> origin/gh/slayton58/46/head 2025-12-04T08:57:03.9700841Z * [new branch] gh/slayton58/46/orig -> origin/gh/slayton58/46/orig 2025-12-04T08:57:03.9702945Z * [new branch] gh/slayton58/6/base -> origin/gh/slayton58/6/base 2025-12-04T08:57:03.9704740Z * [new branch] gh/slayton58/6/head -> origin/gh/slayton58/6/head 2025-12-04T08:57:03.9707774Z * [new branch] gh/slayton58/7/base -> origin/gh/slayton58/7/base 2025-12-04T08:57:03.9709295Z * [new branch] gh/slayton58/7/head -> origin/gh/slayton58/7/head 2025-12-04T08:57:03.9711958Z * [new branch] gh/soulitzer/269/base -> origin/gh/soulitzer/269/base 2025-12-04T08:57:03.9713557Z * [new branch] gh/soulitzer/269/head -> origin/gh/soulitzer/269/head 2025-12-04T08:57:03.9715175Z * [new branch] gh/soulitzer/269/orig -> origin/gh/soulitzer/269/orig 2025-12-04T08:57:03.9717332Z * [new branch] gh/soulitzer/276/base -> origin/gh/soulitzer/276/base 2025-12-04T08:57:03.9718868Z * [new branch] gh/soulitzer/276/head -> origin/gh/soulitzer/276/head 2025-12-04T08:57:03.9720472Z * [new branch] gh/soulitzer/276/orig -> origin/gh/soulitzer/276/orig 2025-12-04T08:57:03.9722833Z * [new branch] gh/soulitzer/287/base -> origin/gh/soulitzer/287/base 2025-12-04T08:57:03.9724400Z * [new branch] gh/soulitzer/287/head -> origin/gh/soulitzer/287/head 2025-12-04T08:57:03.9726019Z * [new branch] gh/soulitzer/287/orig -> origin/gh/soulitzer/287/orig 2025-12-04T08:57:03.9728267Z * [new branch] gh/soulitzer/296/base -> origin/gh/soulitzer/296/base 2025-12-04T08:57:03.9729962Z * [new branch] gh/soulitzer/296/head -> origin/gh/soulitzer/296/head 2025-12-04T08:57:03.9731606Z * [new branch] gh/soulitzer/296/orig -> origin/gh/soulitzer/296/orig 2025-12-04T08:57:03.9733843Z * [new branch] gh/soulitzer/299/base -> origin/gh/soulitzer/299/base 2025-12-04T08:57:03.9735861Z * [new branch] gh/soulitzer/299/head -> origin/gh/soulitzer/299/head 2025-12-04T08:57:03.9737495Z * [new branch] gh/soulitzer/299/orig -> origin/gh/soulitzer/299/orig 2025-12-04T08:57:03.9739595Z * [new branch] gh/soulitzer/300/base -> origin/gh/soulitzer/300/base 2025-12-04T08:57:03.9741397Z * [new branch] gh/soulitzer/300/head -> origin/gh/soulitzer/300/head 2025-12-04T08:57:03.9742926Z * [new branch] gh/soulitzer/300/orig -> origin/gh/soulitzer/300/orig 2025-12-04T08:57:03.9745196Z * [new branch] gh/soulitzer/301/base -> origin/gh/soulitzer/301/base 2025-12-04T08:57:03.9746765Z * [new branch] gh/soulitzer/301/head -> origin/gh/soulitzer/301/head 2025-12-04T08:57:03.9748387Z * [new branch] gh/soulitzer/301/orig -> origin/gh/soulitzer/301/orig 2025-12-04T08:57:03.9750465Z * [new branch] gh/soulitzer/313/base -> origin/gh/soulitzer/313/base 2025-12-04T08:57:03.9752144Z * [new branch] gh/soulitzer/313/head -> origin/gh/soulitzer/313/head 2025-12-04T08:57:03.9753704Z * [new branch] gh/soulitzer/313/orig -> origin/gh/soulitzer/313/orig 2025-12-04T08:57:03.9756198Z * [new branch] gh/soulitzer/319/base -> origin/gh/soulitzer/319/base 2025-12-04T08:57:03.9757526Z * [new branch] gh/soulitzer/319/head -> origin/gh/soulitzer/319/head 2025-12-04T08:57:03.9759215Z * [new branch] gh/soulitzer/319/orig -> origin/gh/soulitzer/319/orig 2025-12-04T08:57:03.9761380Z * [new branch] gh/soulitzer/320/base -> origin/gh/soulitzer/320/base 2025-12-04T08:57:03.9762872Z * [new branch] gh/soulitzer/320/head -> origin/gh/soulitzer/320/head 2025-12-04T08:57:03.9764332Z * [new branch] gh/soulitzer/320/orig -> origin/gh/soulitzer/320/orig 2025-12-04T08:57:03.9766492Z * [new branch] gh/soulitzer/336/base -> origin/gh/soulitzer/336/base 2025-12-04T08:57:03.9768007Z * [new branch] gh/soulitzer/336/head -> origin/gh/soulitzer/336/head 2025-12-04T08:57:03.9769573Z * [new branch] gh/soulitzer/336/orig -> origin/gh/soulitzer/336/orig 2025-12-04T08:57:03.9771729Z * [new branch] gh/soulitzer/347/base -> origin/gh/soulitzer/347/base 2025-12-04T08:57:03.9773378Z * [new branch] gh/soulitzer/347/head -> origin/gh/soulitzer/347/head 2025-12-04T08:57:03.9774941Z * [new branch] gh/soulitzer/347/orig -> origin/gh/soulitzer/347/orig 2025-12-04T08:57:03.9777203Z * [new branch] gh/soulitzer/349/base -> origin/gh/soulitzer/349/base 2025-12-04T08:57:03.9778882Z * [new branch] gh/soulitzer/349/head -> origin/gh/soulitzer/349/head 2025-12-04T08:57:03.9780429Z * [new branch] gh/soulitzer/349/orig -> origin/gh/soulitzer/349/orig 2025-12-04T08:57:03.9782494Z * [new branch] gh/soulitzer/350/base -> origin/gh/soulitzer/350/base 2025-12-04T08:57:03.9784074Z * [new branch] gh/soulitzer/350/head -> origin/gh/soulitzer/350/head 2025-12-04T08:57:03.9785633Z * [new branch] gh/soulitzer/350/orig -> origin/gh/soulitzer/350/orig 2025-12-04T08:57:03.9787685Z * [new branch] gh/soulitzer/351/base -> origin/gh/soulitzer/351/base 2025-12-04T08:57:03.9789266Z * [new branch] gh/soulitzer/351/head -> origin/gh/soulitzer/351/head 2025-12-04T08:57:03.9790815Z * [new branch] gh/soulitzer/351/orig -> origin/gh/soulitzer/351/orig 2025-12-04T08:57:03.9792923Z * [new branch] gh/soulitzer/353/base -> origin/gh/soulitzer/353/base 2025-12-04T08:57:03.9794648Z * [new branch] gh/soulitzer/353/head -> origin/gh/soulitzer/353/head 2025-12-04T08:57:03.9796223Z * [new branch] gh/soulitzer/353/orig -> origin/gh/soulitzer/353/orig 2025-12-04T08:57:03.9799077Z * [new branch] gh/soulitzer/358/base -> origin/gh/soulitzer/358/base 2025-12-04T08:57:03.9800700Z * [new branch] gh/soulitzer/358/head -> origin/gh/soulitzer/358/head 2025-12-04T08:57:03.9802264Z * [new branch] gh/soulitzer/358/orig -> origin/gh/soulitzer/358/orig 2025-12-04T08:57:03.9805035Z * [new branch] gh/soulitzer/359/base -> origin/gh/soulitzer/359/base 2025-12-04T08:57:03.9806814Z * [new branch] gh/soulitzer/359/head -> origin/gh/soulitzer/359/head 2025-12-04T08:57:03.9808483Z * [new branch] gh/soulitzer/359/orig -> origin/gh/soulitzer/359/orig 2025-12-04T08:57:03.9811022Z * [new branch] gh/soulitzer/374/base -> origin/gh/soulitzer/374/base 2025-12-04T08:57:03.9812605Z * [new branch] gh/soulitzer/374/head -> origin/gh/soulitzer/374/head 2025-12-04T08:57:03.9814253Z * [new branch] gh/soulitzer/374/orig -> origin/gh/soulitzer/374/orig 2025-12-04T08:57:03.9816413Z * [new branch] gh/soulitzer/375/base -> origin/gh/soulitzer/375/base 2025-12-04T08:57:03.9818056Z * [new branch] gh/soulitzer/375/head -> origin/gh/soulitzer/375/head 2025-12-04T08:57:03.9819746Z * [new branch] gh/soulitzer/375/orig -> origin/gh/soulitzer/375/orig 2025-12-04T08:57:03.9821776Z * [new branch] gh/soulitzer/380/base -> origin/gh/soulitzer/380/base 2025-12-04T08:57:03.9823367Z * [new branch] gh/soulitzer/380/head -> origin/gh/soulitzer/380/head 2025-12-04T08:57:03.9824896Z * [new branch] gh/soulitzer/380/orig -> origin/gh/soulitzer/380/orig 2025-12-04T08:57:03.9827069Z * [new branch] gh/soulitzer/385/base -> origin/gh/soulitzer/385/base 2025-12-04T08:57:03.9828656Z * [new branch] gh/soulitzer/385/head -> origin/gh/soulitzer/385/head 2025-12-04T08:57:03.9830269Z * [new branch] gh/soulitzer/385/orig -> origin/gh/soulitzer/385/orig 2025-12-04T08:57:03.9832444Z * [new branch] gh/soulitzer/386/base -> origin/gh/soulitzer/386/base 2025-12-04T08:57:03.9834107Z * [new branch] gh/soulitzer/386/head -> origin/gh/soulitzer/386/head 2025-12-04T08:57:03.9835675Z * [new branch] gh/soulitzer/386/orig -> origin/gh/soulitzer/386/orig 2025-12-04T08:57:03.9837901Z * [new branch] gh/soulitzer/387/base -> origin/gh/soulitzer/387/base 2025-12-04T08:57:03.9839510Z * [new branch] gh/soulitzer/387/head -> origin/gh/soulitzer/387/head 2025-12-04T08:57:03.9841064Z * [new branch] gh/soulitzer/387/orig -> origin/gh/soulitzer/387/orig 2025-12-04T08:57:03.9843239Z * [new branch] gh/soulitzer/388/base -> origin/gh/soulitzer/388/base 2025-12-04T08:57:03.9844803Z * [new branch] gh/soulitzer/388/head -> origin/gh/soulitzer/388/head 2025-12-04T08:57:03.9846402Z * [new branch] gh/soulitzer/388/orig -> origin/gh/soulitzer/388/orig 2025-12-04T08:57:03.9848507Z * [new branch] gh/soulitzer/389/base -> origin/gh/soulitzer/389/base 2025-12-04T08:57:03.9850218Z * [new branch] gh/soulitzer/389/head -> origin/gh/soulitzer/389/head 2025-12-04T08:57:03.9851604Z * [new branch] gh/soulitzer/389/orig -> origin/gh/soulitzer/389/orig 2025-12-04T08:57:03.9853892Z * [new branch] gh/soulitzer/390/base -> origin/gh/soulitzer/390/base 2025-12-04T08:57:03.9855433Z * [new branch] gh/soulitzer/390/head -> origin/gh/soulitzer/390/head 2025-12-04T08:57:03.9857069Z * [new branch] gh/soulitzer/390/orig -> origin/gh/soulitzer/390/orig 2025-12-04T08:57:03.9859209Z * [new branch] gh/soulitzer/391/base -> origin/gh/soulitzer/391/base 2025-12-04T08:57:03.9860873Z * [new branch] gh/soulitzer/391/head -> origin/gh/soulitzer/391/head 2025-12-04T08:57:03.9862416Z * [new branch] gh/soulitzer/391/orig -> origin/gh/soulitzer/391/orig 2025-12-04T08:57:03.9865101Z * [new branch] gh/soulitzer/392/base -> origin/gh/soulitzer/392/base 2025-12-04T08:57:03.9866692Z * [new branch] gh/soulitzer/392/head -> origin/gh/soulitzer/392/head 2025-12-04T08:57:03.9868219Z * [new branch] gh/soulitzer/392/orig -> origin/gh/soulitzer/392/orig 2025-12-04T08:57:03.9870721Z * [new branch] gh/swolchok/728/next -> origin/gh/swolchok/728/next 2025-12-04T08:57:03.9873071Z * [new branch] gh/swolchok/819/base -> origin/gh/swolchok/819/base 2025-12-04T08:57:03.9874543Z * [new branch] gh/swolchok/819/head -> origin/gh/swolchok/819/head 2025-12-04T08:57:03.9876096Z * [new branch] gh/swolchok/819/orig -> origin/gh/swolchok/819/orig 2025-12-04T08:57:03.9878210Z * [new branch] gh/swolchok/824/base -> origin/gh/swolchok/824/base 2025-12-04T08:57:03.9879776Z * [new branch] gh/swolchok/824/head -> origin/gh/swolchok/824/head 2025-12-04T08:57:03.9881397Z * [new branch] gh/swolchok/824/orig -> origin/gh/swolchok/824/orig 2025-12-04T08:57:03.9883631Z * [new branch] gh/swolchok/829/base -> origin/gh/swolchok/829/base 2025-12-04T08:57:03.9885013Z * [new branch] gh/swolchok/829/head -> origin/gh/swolchok/829/head 2025-12-04T08:57:03.9886587Z * [new branch] gh/swolchok/829/orig -> origin/gh/swolchok/829/orig 2025-12-04T08:57:03.9888760Z * [new branch] gh/swolchok/839/base -> origin/gh/swolchok/839/base 2025-12-04T08:57:03.9890265Z * [new branch] gh/swolchok/839/head -> origin/gh/swolchok/839/head 2025-12-04T08:57:03.9891829Z * [new branch] gh/swolchok/839/orig -> origin/gh/swolchok/839/orig 2025-12-04T08:57:03.9894073Z * [new branch] gh/swolchok/841/base -> origin/gh/swolchok/841/base 2025-12-04T08:57:03.9895590Z * [new branch] gh/swolchok/841/head -> origin/gh/swolchok/841/head 2025-12-04T08:57:03.9897224Z * [new branch] gh/swolchok/841/orig -> origin/gh/swolchok/841/orig 2025-12-04T08:57:03.9899378Z * [new branch] gh/swolchok/842/base -> origin/gh/swolchok/842/base 2025-12-04T08:57:03.9900952Z * [new branch] gh/swolchok/842/head -> origin/gh/swolchok/842/head 2025-12-04T08:57:03.9902562Z * [new branch] gh/swolchok/842/orig -> origin/gh/swolchok/842/orig 2025-12-04T08:57:03.9904782Z * [new branch] gh/swolchok/845/base -> origin/gh/swolchok/845/base 2025-12-04T08:57:03.9906525Z * [new branch] gh/swolchok/845/head -> origin/gh/swolchok/845/head 2025-12-04T08:57:03.9908184Z * [new branch] gh/swolchok/845/orig -> origin/gh/swolchok/845/orig 2025-12-04T08:57:03.9910297Z * [new branch] gh/swolchok/848/base -> origin/gh/swolchok/848/base 2025-12-04T08:57:03.9911934Z * [new branch] gh/swolchok/848/head -> origin/gh/swolchok/848/head 2025-12-04T08:57:03.9913527Z * [new branch] gh/swolchok/848/orig -> origin/gh/swolchok/848/orig 2025-12-04T08:57:03.9915595Z * [new branch] gh/swolchok/856/base -> origin/gh/swolchok/856/base 2025-12-04T08:57:03.9917269Z * [new branch] gh/swolchok/856/head -> origin/gh/swolchok/856/head 2025-12-04T08:57:03.9918797Z * [new branch] gh/swolchok/856/orig -> origin/gh/swolchok/856/orig 2025-12-04T08:57:03.9920976Z * [new branch] gh/swolchok/860/base -> origin/gh/swolchok/860/base 2025-12-04T08:57:03.9922587Z * [new branch] gh/swolchok/860/head -> origin/gh/swolchok/860/head 2025-12-04T08:57:03.9924101Z * [new branch] gh/swolchok/860/orig -> origin/gh/swolchok/860/orig 2025-12-04T08:57:03.9926406Z * [new branch] gh/swolchok/861/base -> origin/gh/swolchok/861/base 2025-12-04T08:57:03.9928020Z * [new branch] gh/swolchok/861/head -> origin/gh/swolchok/861/head 2025-12-04T08:57:03.9929620Z * [new branch] gh/swolchok/861/orig -> origin/gh/swolchok/861/orig 2025-12-04T08:57:03.9931777Z * [new branch] gh/swolchok/862/base -> origin/gh/swolchok/862/base 2025-12-04T08:57:03.9933342Z * [new branch] gh/swolchok/862/head -> origin/gh/swolchok/862/head 2025-12-04T08:57:03.9934862Z * [new branch] gh/swolchok/862/orig -> origin/gh/swolchok/862/orig 2025-12-04T08:57:03.9937541Z * [new branch] gh/swolchok/863/base -> origin/gh/swolchok/863/base 2025-12-04T08:57:03.9939065Z * [new branch] gh/swolchok/863/head -> origin/gh/swolchok/863/head 2025-12-04T08:57:03.9940748Z * [new branch] gh/swolchok/863/orig -> origin/gh/swolchok/863/orig 2025-12-04T08:57:03.9942933Z * [new branch] gh/swolchok/864/base -> origin/gh/swolchok/864/base 2025-12-04T08:57:03.9944556Z * [new branch] gh/swolchok/864/head -> origin/gh/swolchok/864/head 2025-12-04T08:57:03.9946338Z * [new branch] gh/swolchok/864/orig -> origin/gh/swolchok/864/orig 2025-12-04T08:57:03.9948330Z * [new branch] gh/swolchok/865/base -> origin/gh/swolchok/865/base 2025-12-04T08:57:03.9950048Z * [new branch] gh/swolchok/865/head -> origin/gh/swolchok/865/head 2025-12-04T08:57:03.9951587Z * [new branch] gh/swolchok/865/orig -> origin/gh/swolchok/865/orig 2025-12-04T08:57:03.9954206Z * [new branch] gh/swolchok/866/base -> origin/gh/swolchok/866/base 2025-12-04T08:57:03.9955894Z * [new branch] gh/swolchok/866/head -> origin/gh/swolchok/866/head 2025-12-04T08:57:03.9957478Z * [new branch] gh/swolchok/866/orig -> origin/gh/swolchok/866/orig 2025-12-04T08:57:03.9959543Z * [new branch] gh/swolchok/867/base -> origin/gh/swolchok/867/base 2025-12-04T08:57:03.9961170Z * [new branch] gh/swolchok/867/head -> origin/gh/swolchok/867/head 2025-12-04T08:57:03.9962743Z * [new branch] gh/swolchok/867/orig -> origin/gh/swolchok/867/orig 2025-12-04T08:57:03.9964944Z * [new branch] gh/swolchok/868/base -> origin/gh/swolchok/868/base 2025-12-04T08:57:03.9966517Z * [new branch] gh/swolchok/868/head -> origin/gh/swolchok/868/head 2025-12-04T08:57:03.9968150Z * [new branch] gh/swolchok/868/orig -> origin/gh/swolchok/868/orig 2025-12-04T08:57:03.9970333Z * [new branch] gh/swolchok/869/base -> origin/gh/swolchok/869/base 2025-12-04T08:57:03.9971940Z * [new branch] gh/swolchok/869/head -> origin/gh/swolchok/869/head 2025-12-04T08:57:03.9973625Z * [new branch] gh/swolchok/869/orig -> origin/gh/swolchok/869/orig 2025-12-04T08:57:03.9975877Z * [new branch] gh/swolchok/870/base -> origin/gh/swolchok/870/base 2025-12-04T08:57:03.9977375Z * [new branch] gh/swolchok/870/head -> origin/gh/swolchok/870/head 2025-12-04T08:57:03.9978979Z * [new branch] gh/swolchok/870/orig -> origin/gh/swolchok/870/orig 2025-12-04T08:57:03.9981159Z * [new branch] gh/swolchok/871/base -> origin/gh/swolchok/871/base 2025-12-04T08:57:03.9982756Z * [new branch] gh/swolchok/871/head -> origin/gh/swolchok/871/head 2025-12-04T08:57:03.9984317Z * [new branch] gh/swolchok/871/orig -> origin/gh/swolchok/871/orig 2025-12-04T08:57:03.9987159Z * [new branch] gh/teja-rao/4/base -> origin/gh/teja-rao/4/base 2025-12-04T08:57:03.9988729Z * [new branch] gh/teja-rao/4/head -> origin/gh/teja-rao/4/head 2025-12-04T08:57:03.9990271Z * [new branch] gh/teja-rao/4/orig -> origin/gh/teja-rao/4/orig 2025-12-04T08:57:03.9992800Z * [new branch] gh/tianyu-l/2/base -> origin/gh/tianyu-l/2/base 2025-12-04T08:57:03.9994395Z * [new branch] gh/tianyu-l/2/head -> origin/gh/tianyu-l/2/head 2025-12-04T08:57:03.9995934Z * [new branch] gh/tianyu-l/2/orig -> origin/gh/tianyu-l/2/orig 2025-12-04T08:57:03.9997984Z * [new branch] gh/tianyu-l/3/base -> origin/gh/tianyu-l/3/base 2025-12-04T08:57:03.9999595Z * [new branch] gh/tianyu-l/3/orig -> origin/gh/tianyu-l/3/orig 2025-12-04T08:57:04.0001669Z * [new branch] gh/tianyu-l/4/base -> origin/gh/tianyu-l/4/base 2025-12-04T08:57:04.0003214Z * [new branch] gh/tianyu-l/4/head -> origin/gh/tianyu-l/4/head 2025-12-04T08:57:04.0004816Z * [new branch] gh/tianyu-l/4/orig -> origin/gh/tianyu-l/4/orig 2025-12-04T08:57:04.0008126Z * [new branch] gh/tugsbayasgalan/10/base -> origin/gh/tugsbayasgalan/10/base 2025-12-04T08:57:04.0009807Z * [new branch] gh/tugsbayasgalan/10/head -> origin/gh/tugsbayasgalan/10/head 2025-12-04T08:57:04.0011679Z * [new branch] gh/tugsbayasgalan/10/orig -> origin/gh/tugsbayasgalan/10/orig 2025-12-04T08:57:04.0013707Z * [new branch] gh/tugsbayasgalan/13/base -> origin/gh/tugsbayasgalan/13/base 2025-12-04T08:57:04.0015259Z * [new branch] gh/tugsbayasgalan/13/head -> origin/gh/tugsbayasgalan/13/head 2025-12-04T08:57:04.0016821Z * [new branch] gh/tugsbayasgalan/13/orig -> origin/gh/tugsbayasgalan/13/orig 2025-12-04T08:57:04.0018958Z * [new branch] gh/tugsbayasgalan/17/base -> origin/gh/tugsbayasgalan/17/base 2025-12-04T08:57:04.0020454Z * [new branch] gh/tugsbayasgalan/17/head -> origin/gh/tugsbayasgalan/17/head 2025-12-04T08:57:04.0022075Z * [new branch] gh/tugsbayasgalan/17/orig -> origin/gh/tugsbayasgalan/17/orig 2025-12-04T08:57:04.0024337Z * [new branch] gh/tugsbayasgalan/2/base -> origin/gh/tugsbayasgalan/2/base 2025-12-04T08:57:04.0025886Z * [new branch] gh/tugsbayasgalan/2/head -> origin/gh/tugsbayasgalan/2/head 2025-12-04T08:57:04.0027444Z * [new branch] gh/tugsbayasgalan/2/orig -> origin/gh/tugsbayasgalan/2/orig 2025-12-04T08:57:04.0029881Z * [new branch] gh/tugsbayasgalan/28/base -> origin/gh/tugsbayasgalan/28/base 2025-12-04T08:57:04.0031476Z * [new branch] gh/tugsbayasgalan/28/head -> origin/gh/tugsbayasgalan/28/head 2025-12-04T08:57:04.0033012Z * [new branch] gh/tugsbayasgalan/28/orig -> origin/gh/tugsbayasgalan/28/orig 2025-12-04T08:57:04.0035077Z * [new branch] gh/tugsbayasgalan/32/base -> origin/gh/tugsbayasgalan/32/base 2025-12-04T08:57:04.0036730Z * [new branch] gh/tugsbayasgalan/32/head -> origin/gh/tugsbayasgalan/32/head 2025-12-04T08:57:04.0038256Z * [new branch] gh/tugsbayasgalan/32/orig -> origin/gh/tugsbayasgalan/32/orig 2025-12-04T08:57:04.0040444Z * [new branch] gh/tugsbayasgalan/35/base -> origin/gh/tugsbayasgalan/35/base 2025-12-04T08:57:04.0042076Z * [new branch] gh/tugsbayasgalan/35/head -> origin/gh/tugsbayasgalan/35/head 2025-12-04T08:57:04.0043678Z * [new branch] gh/tugsbayasgalan/35/orig -> origin/gh/tugsbayasgalan/35/orig 2025-12-04T08:57:04.0045818Z * [new branch] gh/tugsbayasgalan/36/base -> origin/gh/tugsbayasgalan/36/base 2025-12-04T08:57:04.0047425Z * [new branch] gh/tugsbayasgalan/36/head -> origin/gh/tugsbayasgalan/36/head 2025-12-04T08:57:04.0048965Z * [new branch] gh/tugsbayasgalan/36/orig -> origin/gh/tugsbayasgalan/36/orig 2025-12-04T08:57:04.0051162Z * [new branch] gh/tugsbayasgalan/37/base -> origin/gh/tugsbayasgalan/37/base 2025-12-04T08:57:04.0052693Z * [new branch] gh/tugsbayasgalan/37/head -> origin/gh/tugsbayasgalan/37/head 2025-12-04T08:57:04.0054397Z * [new branch] gh/tugsbayasgalan/37/orig -> origin/gh/tugsbayasgalan/37/orig 2025-12-04T08:57:04.0056438Z * [new branch] gh/tugsbayasgalan/43/base -> origin/gh/tugsbayasgalan/43/base 2025-12-04T08:57:04.0057970Z * [new branch] gh/tugsbayasgalan/43/head -> origin/gh/tugsbayasgalan/43/head 2025-12-04T08:57:04.0059645Z * [new branch] gh/tugsbayasgalan/43/orig -> origin/gh/tugsbayasgalan/43/orig 2025-12-04T08:57:04.0062061Z * [new branch] gh/tugsbayasgalan/48/base -> origin/gh/tugsbayasgalan/48/base 2025-12-04T08:57:04.0063671Z * [new branch] gh/tugsbayasgalan/48/head -> origin/gh/tugsbayasgalan/48/head 2025-12-04T08:57:04.0065223Z * [new branch] gh/tugsbayasgalan/48/orig -> origin/gh/tugsbayasgalan/48/orig 2025-12-04T08:57:04.0067321Z * [new branch] gh/tugsbayasgalan/51/base -> origin/gh/tugsbayasgalan/51/base 2025-12-04T08:57:04.0068878Z * [new branch] gh/tugsbayasgalan/51/head -> origin/gh/tugsbayasgalan/51/head 2025-12-04T08:57:04.0070568Z * [new branch] gh/tugsbayasgalan/51/orig -> origin/gh/tugsbayasgalan/51/orig 2025-12-04T08:57:04.0072530Z * [new branch] gh/tugsbayasgalan/52/base -> origin/gh/tugsbayasgalan/52/base 2025-12-04T08:57:04.0074096Z * [new branch] gh/tugsbayasgalan/52/head -> origin/gh/tugsbayasgalan/52/head 2025-12-04T08:57:04.0075663Z * [new branch] gh/tugsbayasgalan/52/orig -> origin/gh/tugsbayasgalan/52/orig 2025-12-04T08:57:04.0077832Z * [new branch] gh/tugsbayasgalan/53/base -> origin/gh/tugsbayasgalan/53/base 2025-12-04T08:57:04.0079390Z * [new branch] gh/tugsbayasgalan/53/head -> origin/gh/tugsbayasgalan/53/head 2025-12-04T08:57:04.0080965Z * [new branch] gh/tugsbayasgalan/53/orig -> origin/gh/tugsbayasgalan/53/orig 2025-12-04T08:57:04.0083186Z * [new branch] gh/tugsbayasgalan/55/base -> origin/gh/tugsbayasgalan/55/base 2025-12-04T08:57:04.0084838Z * [new branch] gh/tugsbayasgalan/55/head -> origin/gh/tugsbayasgalan/55/head 2025-12-04T08:57:04.0086495Z * [new branch] gh/tugsbayasgalan/55/orig -> origin/gh/tugsbayasgalan/55/orig 2025-12-04T08:57:04.0088630Z * [new branch] gh/tugsbayasgalan/59/base -> origin/gh/tugsbayasgalan/59/base 2025-12-04T08:57:04.0090233Z * [new branch] gh/tugsbayasgalan/59/head -> origin/gh/tugsbayasgalan/59/head 2025-12-04T08:57:04.0091833Z * [new branch] gh/tugsbayasgalan/59/orig -> origin/gh/tugsbayasgalan/59/orig 2025-12-04T08:57:04.0094109Z * [new branch] gh/tugsbayasgalan/6/base -> origin/gh/tugsbayasgalan/6/base 2025-12-04T08:57:04.0095602Z * [new branch] gh/tugsbayasgalan/6/head -> origin/gh/tugsbayasgalan/6/head 2025-12-04T08:57:04.0097148Z * [new branch] gh/tugsbayasgalan/6/orig -> origin/gh/tugsbayasgalan/6/orig 2025-12-04T08:57:04.0099173Z * [new branch] gh/tugsbayasgalan/60/base -> origin/gh/tugsbayasgalan/60/base 2025-12-04T08:57:04.0100724Z * [new branch] gh/tugsbayasgalan/60/head -> origin/gh/tugsbayasgalan/60/head 2025-12-04T08:57:04.0102262Z * [new branch] gh/tugsbayasgalan/60/orig -> origin/gh/tugsbayasgalan/60/orig 2025-12-04T08:57:04.0105067Z * [new branch] gh/tugsbayasgalan/61/base -> origin/gh/tugsbayasgalan/61/base 2025-12-04T08:57:04.0108240Z * [new branch] gh/tugsbayasgalan/61/head -> origin/gh/tugsbayasgalan/61/head 2025-12-04T08:57:04.0109737Z * [new branch] gh/tugsbayasgalan/61/orig -> origin/gh/tugsbayasgalan/61/orig 2025-12-04T08:57:04.0112099Z * [new branch] gh/tugsbayasgalan/63/base -> origin/gh/tugsbayasgalan/63/base 2025-12-04T08:57:04.0113671Z * [new branch] gh/tugsbayasgalan/63/head -> origin/gh/tugsbayasgalan/63/head 2025-12-04T08:57:04.0115263Z * [new branch] gh/tugsbayasgalan/63/orig -> origin/gh/tugsbayasgalan/63/orig 2025-12-04T08:57:04.0117971Z * [new branch] gh/tugsbayasgalan/67/base -> origin/gh/tugsbayasgalan/67/base 2025-12-04T08:57:04.0119698Z * [new branch] gh/tugsbayasgalan/67/head -> origin/gh/tugsbayasgalan/67/head 2025-12-04T08:57:04.0121254Z * [new branch] gh/tugsbayasgalan/67/orig -> origin/gh/tugsbayasgalan/67/orig 2025-12-04T08:57:04.0123591Z * [new branch] gh/tugsbayasgalan/68/base -> origin/gh/tugsbayasgalan/68/base 2025-12-04T08:57:04.0125153Z * [new branch] gh/tugsbayasgalan/68/head -> origin/gh/tugsbayasgalan/68/head 2025-12-04T08:57:04.0126692Z * [new branch] gh/tugsbayasgalan/68/orig -> origin/gh/tugsbayasgalan/68/orig 2025-12-04T08:57:04.0129123Z * [new branch] gh/tugsbayasgalan/7/base -> origin/gh/tugsbayasgalan/7/base 2025-12-04T08:57:04.0130450Z * [new branch] gh/tugsbayasgalan/7/head -> origin/gh/tugsbayasgalan/7/head 2025-12-04T08:57:04.0131934Z * [new branch] gh/tugsbayasgalan/7/orig -> origin/gh/tugsbayasgalan/7/orig 2025-12-04T08:57:04.0146471Z * [new branch] gh/tugsbayasgalan/70/base -> origin/gh/tugsbayasgalan/70/base 2025-12-04T08:57:04.0148601Z * [new branch] gh/tugsbayasgalan/70/head -> origin/gh/tugsbayasgalan/70/head 2025-12-04T08:57:04.0149485Z * [new branch] gh/tugsbayasgalan/70/orig -> origin/gh/tugsbayasgalan/70/orig 2025-12-04T08:57:04.0150079Z * [new branch] gh/tugsbayasgalan/71/base -> origin/gh/tugsbayasgalan/71/base 2025-12-04T08:57:04.0150606Z * [new branch] gh/tugsbayasgalan/71/head -> origin/gh/tugsbayasgalan/71/head 2025-12-04T08:57:04.0151123Z * [new branch] gh/tugsbayasgalan/71/orig -> origin/gh/tugsbayasgalan/71/orig 2025-12-04T08:57:04.0151988Z * [new branch] gh/tugsbayasgalan/72/base -> origin/gh/tugsbayasgalan/72/base 2025-12-04T08:57:04.0152524Z * [new branch] gh/tugsbayasgalan/72/head -> origin/gh/tugsbayasgalan/72/head 2025-12-04T08:57:04.0153239Z * [new branch] gh/tugsbayasgalan/72/orig -> origin/gh/tugsbayasgalan/72/orig 2025-12-04T08:57:04.0153870Z * [new branch] gh/tugsbayasgalan/73/base -> origin/gh/tugsbayasgalan/73/base 2025-12-04T08:57:04.0154326Z * [new branch] gh/tugsbayasgalan/73/head -> origin/gh/tugsbayasgalan/73/head 2025-12-04T08:57:04.0155171Z * [new branch] gh/tugsbayasgalan/73/orig -> origin/gh/tugsbayasgalan/73/orig 2025-12-04T08:57:04.0157814Z * [new branch] gh/tugsbayasgalan/74/base -> origin/gh/tugsbayasgalan/74/base 2025-12-04T08:57:04.0159058Z * [new branch] gh/tugsbayasgalan/74/head -> origin/gh/tugsbayasgalan/74/head 2025-12-04T08:57:04.0160716Z * [new branch] gh/tugsbayasgalan/74/orig -> origin/gh/tugsbayasgalan/74/orig 2025-12-04T08:57:04.0163401Z * [new branch] gh/tugsbayasgalan/75/base -> origin/gh/tugsbayasgalan/75/base 2025-12-04T08:57:04.0164772Z * [new branch] gh/tugsbayasgalan/75/head -> origin/gh/tugsbayasgalan/75/head 2025-12-04T08:57:04.0167069Z * [new branch] gh/tugsbayasgalan/75/orig -> origin/gh/tugsbayasgalan/75/orig 2025-12-04T08:57:04.0168817Z * [new branch] gh/tugsbayasgalan/76/base -> origin/gh/tugsbayasgalan/76/base 2025-12-04T08:57:04.0170443Z * [new branch] gh/tugsbayasgalan/76/head -> origin/gh/tugsbayasgalan/76/head 2025-12-04T08:57:04.0171981Z * [new branch] gh/tugsbayasgalan/76/orig -> origin/gh/tugsbayasgalan/76/orig 2025-12-04T08:57:04.0174677Z * [new branch] gh/tugsbayasgalan/77/base -> origin/gh/tugsbayasgalan/77/base 2025-12-04T08:57:04.0176133Z * [new branch] gh/tugsbayasgalan/77/head -> origin/gh/tugsbayasgalan/77/head 2025-12-04T08:57:04.0177675Z * [new branch] gh/tugsbayasgalan/77/orig -> origin/gh/tugsbayasgalan/77/orig 2025-12-04T08:57:04.0180161Z * [new branch] gh/tugsbayasgalan/78/base -> origin/gh/tugsbayasgalan/78/base 2025-12-04T08:57:04.0181635Z * [new branch] gh/tugsbayasgalan/78/head -> origin/gh/tugsbayasgalan/78/head 2025-12-04T08:57:04.0183552Z * [new branch] gh/tugsbayasgalan/78/orig -> origin/gh/tugsbayasgalan/78/orig 2025-12-04T08:57:04.0186086Z * [new branch] gh/tugsbayasgalan/79/base -> origin/gh/tugsbayasgalan/79/base 2025-12-04T08:57:04.0187453Z * [new branch] gh/tugsbayasgalan/79/head -> origin/gh/tugsbayasgalan/79/head 2025-12-04T08:57:04.0189056Z * [new branch] gh/tugsbayasgalan/79/orig -> origin/gh/tugsbayasgalan/79/orig 2025-12-04T08:57:04.0191921Z * [new branch] gh/tugsbayasgalan/8/base -> origin/gh/tugsbayasgalan/8/base 2025-12-04T08:57:04.0193124Z * [new branch] gh/tugsbayasgalan/8/head -> origin/gh/tugsbayasgalan/8/head 2025-12-04T08:57:04.0194735Z * [new branch] gh/tugsbayasgalan/8/orig -> origin/gh/tugsbayasgalan/8/orig 2025-12-04T08:57:04.0197300Z * [new branch] gh/tugsbayasgalan/80/base -> origin/gh/tugsbayasgalan/80/base 2025-12-04T08:57:04.0198386Z * [new branch] gh/tugsbayasgalan/80/head -> origin/gh/tugsbayasgalan/80/head 2025-12-04T08:57:04.0199872Z * [new branch] gh/tugsbayasgalan/80/orig -> origin/gh/tugsbayasgalan/80/orig 2025-12-04T08:57:04.0202303Z * [new branch] gh/tugsbayasgalan/81/base -> origin/gh/tugsbayasgalan/81/base 2025-12-04T08:57:04.0204036Z * [new branch] gh/tugsbayasgalan/81/head -> origin/gh/tugsbayasgalan/81/head 2025-12-04T08:57:04.0205906Z * [new branch] gh/tugsbayasgalan/81/orig -> origin/gh/tugsbayasgalan/81/orig 2025-12-04T08:57:04.0208784Z * [new branch] gh/tugsbayasgalan/82/base -> origin/gh/tugsbayasgalan/82/base 2025-12-04T08:57:04.0210306Z * [new branch] gh/tugsbayasgalan/82/head -> origin/gh/tugsbayasgalan/82/head 2025-12-04T08:57:04.0211955Z * [new branch] gh/tugsbayasgalan/82/orig -> origin/gh/tugsbayasgalan/82/orig 2025-12-04T08:57:04.0214326Z * [new branch] gh/tugsbayasgalan/83/base -> origin/gh/tugsbayasgalan/83/base 2025-12-04T08:57:04.0215678Z * [new branch] gh/tugsbayasgalan/83/head -> origin/gh/tugsbayasgalan/83/head 2025-12-04T08:57:04.0217258Z * [new branch] gh/tugsbayasgalan/83/orig -> origin/gh/tugsbayasgalan/83/orig 2025-12-04T08:57:04.0219532Z * [new branch] gh/tugsbayasgalan/84/base -> origin/gh/tugsbayasgalan/84/base 2025-12-04T08:57:04.0220870Z * [new branch] gh/tugsbayasgalan/84/head -> origin/gh/tugsbayasgalan/84/head 2025-12-04T08:57:04.0222424Z * [new branch] gh/tugsbayasgalan/84/orig -> origin/gh/tugsbayasgalan/84/orig 2025-12-04T08:57:04.0224669Z * [new branch] gh/tugsbayasgalan/85/base -> origin/gh/tugsbayasgalan/85/base 2025-12-04T08:57:04.0226015Z * [new branch] gh/tugsbayasgalan/85/head -> origin/gh/tugsbayasgalan/85/head 2025-12-04T08:57:04.0227570Z * [new branch] gh/tugsbayasgalan/85/orig -> origin/gh/tugsbayasgalan/85/orig 2025-12-04T08:57:04.0230018Z * [new branch] gh/tugsbayasgalan/86/base -> origin/gh/tugsbayasgalan/86/base 2025-12-04T08:57:04.0231419Z * [new branch] gh/tugsbayasgalan/86/head -> origin/gh/tugsbayasgalan/86/head 2025-12-04T08:57:04.0232962Z * [new branch] gh/tugsbayasgalan/86/orig -> origin/gh/tugsbayasgalan/86/orig 2025-12-04T08:57:04.0235534Z * [new branch] gh/tugsbayasgalan/87/base -> origin/gh/tugsbayasgalan/87/base 2025-12-04T08:57:04.0237004Z * [new branch] gh/tugsbayasgalan/87/head -> origin/gh/tugsbayasgalan/87/head 2025-12-04T08:57:04.0238520Z * [new branch] gh/tugsbayasgalan/87/orig -> origin/gh/tugsbayasgalan/87/orig 2025-12-04T08:57:04.0240951Z * [new branch] gh/tugsbayasgalan/88/base -> origin/gh/tugsbayasgalan/88/base 2025-12-04T08:57:04.0242294Z * [new branch] gh/tugsbayasgalan/88/head -> origin/gh/tugsbayasgalan/88/head 2025-12-04T08:57:04.0244082Z * [new branch] gh/tugsbayasgalan/88/orig -> origin/gh/tugsbayasgalan/88/orig 2025-12-04T08:57:04.0246841Z * [new branch] gh/tugsbayasgalan/89/base -> origin/gh/tugsbayasgalan/89/base 2025-12-04T08:57:04.0248198Z * [new branch] gh/tugsbayasgalan/89/head -> origin/gh/tugsbayasgalan/89/head 2025-12-04T08:57:04.0249788Z * [new branch] gh/tugsbayasgalan/89/orig -> origin/gh/tugsbayasgalan/89/orig 2025-12-04T08:57:04.0252168Z * [new branch] gh/tugsbayasgalan/9/base -> origin/gh/tugsbayasgalan/9/base 2025-12-04T08:57:04.0253550Z * [new branch] gh/tugsbayasgalan/9/head -> origin/gh/tugsbayasgalan/9/head 2025-12-04T08:57:04.0255158Z * [new branch] gh/tugsbayasgalan/9/orig -> origin/gh/tugsbayasgalan/9/orig 2025-12-04T08:57:04.0257511Z * [new branch] gh/tugsbayasgalan/90/base -> origin/gh/tugsbayasgalan/90/base 2025-12-04T08:57:04.0259043Z * [new branch] gh/tugsbayasgalan/90/head -> origin/gh/tugsbayasgalan/90/head 2025-12-04T08:57:04.0260555Z * [new branch] gh/tugsbayasgalan/90/orig -> origin/gh/tugsbayasgalan/90/orig 2025-12-04T08:57:04.0262879Z * [new branch] gh/tugsbayasgalan/91/base -> origin/gh/tugsbayasgalan/91/base 2025-12-04T08:57:04.0264246Z * [new branch] gh/tugsbayasgalan/91/head -> origin/gh/tugsbayasgalan/91/head 2025-12-04T08:57:04.0266413Z * [new branch] gh/tugsbayasgalan/91/orig -> origin/gh/tugsbayasgalan/91/orig 2025-12-04T08:57:04.0268679Z * [new branch] gh/tugsbayasgalan/92/base -> origin/gh/tugsbayasgalan/92/base 2025-12-04T08:57:04.0270140Z * [new branch] gh/tugsbayasgalan/92/head -> origin/gh/tugsbayasgalan/92/head 2025-12-04T08:57:04.0271827Z * [new branch] gh/tugsbayasgalan/92/orig -> origin/gh/tugsbayasgalan/92/orig 2025-12-04T08:57:04.0274157Z * [new branch] gh/tugsbayasgalan/93/base -> origin/gh/tugsbayasgalan/93/base 2025-12-04T08:57:04.0275652Z * [new branch] gh/tugsbayasgalan/93/head -> origin/gh/tugsbayasgalan/93/head 2025-12-04T08:57:04.0277186Z * [new branch] gh/tugsbayasgalan/93/orig -> origin/gh/tugsbayasgalan/93/orig 2025-12-04T08:57:04.0279981Z * [new branch] gh/v0i0/14/base -> origin/gh/v0i0/14/base 2025-12-04T08:57:04.0281857Z * [new branch] gh/v0i0/14/head -> origin/gh/v0i0/14/head 2025-12-04T08:57:04.0283336Z * [new branch] gh/v0i0/14/orig -> origin/gh/v0i0/14/orig 2025-12-04T08:57:04.0285524Z * [new branch] gh/v0i0/15/base -> origin/gh/v0i0/15/base 2025-12-04T08:57:04.0286800Z * [new branch] gh/v0i0/15/head -> origin/gh/v0i0/15/head 2025-12-04T08:57:04.0288375Z * [new branch] gh/v0i0/15/orig -> origin/gh/v0i0/15/orig 2025-12-04T08:57:04.0290767Z * [new branch] gh/v0i0/16/base -> origin/gh/v0i0/16/base 2025-12-04T08:57:04.0292051Z * [new branch] gh/v0i0/16/head -> origin/gh/v0i0/16/head 2025-12-04T08:57:04.0293718Z * [new branch] gh/v0i0/16/orig -> origin/gh/v0i0/16/orig 2025-12-04T08:57:04.0296182Z * [new branch] gh/v0i0/17/base -> origin/gh/v0i0/17/base 2025-12-04T08:57:04.0297507Z * [new branch] gh/v0i0/17/head -> origin/gh/v0i0/17/head 2025-12-04T08:57:04.0298976Z * [new branch] gh/v0i0/17/orig -> origin/gh/v0i0/17/orig 2025-12-04T08:57:04.0301457Z * [new branch] gh/v0i0/18/base -> origin/gh/v0i0/18/base 2025-12-04T08:57:04.0302810Z * [new branch] gh/v0i0/18/head -> origin/gh/v0i0/18/head 2025-12-04T08:57:04.0304273Z * [new branch] gh/v0i0/18/orig -> origin/gh/v0i0/18/orig 2025-12-04T08:57:04.0306997Z * [new branch] gh/v0i0/19/base -> origin/gh/v0i0/19/base 2025-12-04T08:57:04.0308326Z * [new branch] gh/v0i0/19/head -> origin/gh/v0i0/19/head 2025-12-04T08:57:04.0309843Z * [new branch] gh/v0i0/19/orig -> origin/gh/v0i0/19/orig 2025-12-04T08:57:04.0312826Z * [new branch] gh/vishal9-team/1/base -> origin/gh/vishal9-team/1/base 2025-12-04T08:57:04.0314226Z * [new branch] gh/vishal9-team/1/head -> origin/gh/vishal9-team/1/head 2025-12-04T08:57:04.0316431Z * [new branch] gh/vishal9-team/2/base -> origin/gh/vishal9-team/2/base 2025-12-04T08:57:04.0317815Z * [new branch] gh/vishal9-team/2/head -> origin/gh/vishal9-team/2/head 2025-12-04T08:57:04.0319308Z * [new branch] gh/vishal9-team/2/orig -> origin/gh/vishal9-team/2/orig 2025-12-04T08:57:04.0321649Z * [new branch] gh/vishal9-team/3/base -> origin/gh/vishal9-team/3/base 2025-12-04T08:57:04.0323303Z * [new branch] gh/vishal9-team/3/head -> origin/gh/vishal9-team/3/head 2025-12-04T08:57:04.0324648Z * [new branch] gh/vishal9-team/3/orig -> origin/gh/vishal9-team/3/orig 2025-12-04T08:57:04.0326973Z * [new branch] gh/vishal9-team/4/base -> origin/gh/vishal9-team/4/base 2025-12-04T08:57:04.0328242Z * [new branch] gh/vishal9-team/4/head -> origin/gh/vishal9-team/4/head 2025-12-04T08:57:04.0329777Z * [new branch] gh/vishal9-team/4/orig -> origin/gh/vishal9-team/4/orig 2025-12-04T08:57:04.0332508Z * [new branch] gh/vkuzo/1/next -> origin/gh/vkuzo/1/next 2025-12-04T08:57:04.0334819Z * [new branch] gh/vkuzo/2/next -> origin/gh/vkuzo/2/next 2025-12-04T08:57:04.0337322Z * [new branch] gh/vkuzo/3/next -> origin/gh/vkuzo/3/next 2025-12-04T08:57:04.0339874Z * [new branch] gh/wconstab/424/base -> origin/gh/wconstab/424/base 2025-12-04T08:57:04.0341785Z * [new branch] gh/wconstab/424/head -> origin/gh/wconstab/424/head 2025-12-04T08:57:04.0343297Z * [new branch] gh/wconstab/424/orig -> origin/gh/wconstab/424/orig 2025-12-04T08:57:04.0345656Z * [new branch] gh/wconstab/435/base -> origin/gh/wconstab/435/base 2025-12-04T08:57:04.0347063Z * [new branch] gh/wconstab/435/head -> origin/gh/wconstab/435/head 2025-12-04T08:57:04.0349307Z * [new branch] gh/wconstab/435/orig -> origin/gh/wconstab/435/orig 2025-12-04T08:57:04.0351055Z * [new branch] gh/wconstab/444/base -> origin/gh/wconstab/444/base 2025-12-04T08:57:04.0352542Z * [new branch] gh/wconstab/444/head -> origin/gh/wconstab/444/head 2025-12-04T08:57:04.0354552Z * [new branch] gh/wconstab/444/orig -> origin/gh/wconstab/444/orig 2025-12-04T08:57:04.0356816Z * [new branch] gh/wconstab/447/base -> origin/gh/wconstab/447/base 2025-12-04T08:57:04.0358320Z * [new branch] gh/wconstab/447/head -> origin/gh/wconstab/447/head 2025-12-04T08:57:04.0359869Z * [new branch] gh/wconstab/447/orig -> origin/gh/wconstab/447/orig 2025-12-04T08:57:04.0362042Z * [new branch] gh/wconstab/448/base -> origin/gh/wconstab/448/base 2025-12-04T08:57:04.0363647Z * [new branch] gh/wconstab/448/head -> origin/gh/wconstab/448/head 2025-12-04T08:57:04.0365205Z * [new branch] gh/wconstab/448/orig -> origin/gh/wconstab/448/orig 2025-12-04T08:57:04.0367348Z * [new branch] gh/wconstab/449/base -> origin/gh/wconstab/449/base 2025-12-04T08:57:04.0368915Z * [new branch] gh/wconstab/449/head -> origin/gh/wconstab/449/head 2025-12-04T08:57:04.0370542Z * [new branch] gh/wconstab/449/orig -> origin/gh/wconstab/449/orig 2025-12-04T08:57:04.0372511Z * [new branch] gh/wconstab/450/base -> origin/gh/wconstab/450/base 2025-12-04T08:57:04.0374257Z * [new branch] gh/wconstab/450/head -> origin/gh/wconstab/450/head 2025-12-04T08:57:04.0375740Z * [new branch] gh/wconstab/450/orig -> origin/gh/wconstab/450/orig 2025-12-04T08:57:04.0377785Z * [new branch] gh/wconstab/451/base -> origin/gh/wconstab/451/base 2025-12-04T08:57:04.0379418Z * [new branch] gh/wconstab/451/head -> origin/gh/wconstab/451/head 2025-12-04T08:57:04.0380979Z * [new branch] gh/wconstab/451/orig -> origin/gh/wconstab/451/orig 2025-12-04T08:57:04.0383141Z * [new branch] gh/wconstab/452/base -> origin/gh/wconstab/452/base 2025-12-04T08:57:04.0384963Z * [new branch] gh/wconstab/452/head -> origin/gh/wconstab/452/head 2025-12-04T08:57:04.0386484Z * [new branch] gh/wconstab/452/orig -> origin/gh/wconstab/452/orig 2025-12-04T08:57:04.0388546Z * [new branch] gh/wconstab/453/base -> origin/gh/wconstab/453/base 2025-12-04T08:57:04.0390067Z * [new branch] gh/wconstab/453/head -> origin/gh/wconstab/453/head 2025-12-04T08:57:04.0391646Z * [new branch] gh/wconstab/453/orig -> origin/gh/wconstab/453/orig 2025-12-04T08:57:04.0393618Z * [new branch] gh/wconstab/454/base -> origin/gh/wconstab/454/base 2025-12-04T08:57:04.0396519Z * [new branch] gh/wconstab/454/head -> origin/gh/wconstab/454/head 2025-12-04T08:57:04.0397429Z * [new branch] gh/wconstab/454/orig -> origin/gh/wconstab/454/orig 2025-12-04T08:57:04.0399170Z * [new branch] gh/wconstab/455/base -> origin/gh/wconstab/455/base 2025-12-04T08:57:04.0400540Z * [new branch] gh/wconstab/455/head -> origin/gh/wconstab/455/head 2025-12-04T08:57:04.0402216Z * [new branch] gh/wconstab/455/orig -> origin/gh/wconstab/455/orig 2025-12-04T08:57:04.0404685Z * [new branch] gh/wconstab/456/base -> origin/gh/wconstab/456/base 2025-12-04T08:57:04.0406629Z * [new branch] gh/wconstab/456/head -> origin/gh/wconstab/456/head 2025-12-04T08:57:04.0408386Z * [new branch] gh/wconstab/456/orig -> origin/gh/wconstab/456/orig 2025-12-04T08:57:04.0410428Z * [new branch] gh/wconstab/457/base -> origin/gh/wconstab/457/base 2025-12-04T08:57:04.0412046Z * [new branch] gh/wconstab/457/head -> origin/gh/wconstab/457/head 2025-12-04T08:57:04.0413719Z * [new branch] gh/wconstab/457/orig -> origin/gh/wconstab/457/orig 2025-12-04T08:57:04.0415966Z * [new branch] gh/wconstab/458/base -> origin/gh/wconstab/458/base 2025-12-04T08:57:04.0417507Z * [new branch] gh/wconstab/458/head -> origin/gh/wconstab/458/head 2025-12-04T08:57:04.0419107Z * [new branch] gh/wconstab/458/orig -> origin/gh/wconstab/458/orig 2025-12-04T08:57:04.0421149Z * [new branch] gh/wconstab/459/base -> origin/gh/wconstab/459/base 2025-12-04T08:57:04.0422854Z * [new branch] gh/wconstab/459/head -> origin/gh/wconstab/459/head 2025-12-04T08:57:04.0424169Z * [new branch] gh/wconstab/459/orig -> origin/gh/wconstab/459/orig 2025-12-04T08:57:04.0426951Z * [new branch] gh/wconstab/460/base -> origin/gh/wconstab/460/base 2025-12-04T08:57:04.0428662Z * [new branch] gh/wconstab/460/head -> origin/gh/wconstab/460/head 2025-12-04T08:57:04.0430787Z * [new branch] gh/wconstab/460/orig -> origin/gh/wconstab/460/orig 2025-12-04T08:57:04.0433016Z * [new branch] gh/wconstab/461/base -> origin/gh/wconstab/461/base 2025-12-04T08:57:04.0434587Z * [new branch] gh/wconstab/461/head -> origin/gh/wconstab/461/head 2025-12-04T08:57:04.0436120Z * [new branch] gh/wconstab/461/orig -> origin/gh/wconstab/461/orig 2025-12-04T08:57:04.0438140Z * [new branch] gh/wconstab/462/base -> origin/gh/wconstab/462/base 2025-12-04T08:57:04.0439742Z * [new branch] gh/wconstab/462/head -> origin/gh/wconstab/462/head 2025-12-04T08:57:04.0441320Z * [new branch] gh/wconstab/462/orig -> origin/gh/wconstab/462/orig 2025-12-04T08:57:04.0443611Z * [new branch] gh/wconstab/463/base -> origin/gh/wconstab/463/base 2025-12-04T08:57:04.0445244Z * [new branch] gh/wconstab/463/head -> origin/gh/wconstab/463/head 2025-12-04T08:57:04.0446824Z * [new branch] gh/wconstab/463/orig -> origin/gh/wconstab/463/orig 2025-12-04T08:57:04.0448927Z * [new branch] gh/wconstab/464/base -> origin/gh/wconstab/464/base 2025-12-04T08:57:04.0450578Z * [new branch] gh/wconstab/464/head -> origin/gh/wconstab/464/head 2025-12-04T08:57:04.0452320Z * [new branch] gh/wconstab/464/orig -> origin/gh/wconstab/464/orig 2025-12-04T08:57:04.0454407Z * [new branch] gh/wconstab/465/base -> origin/gh/wconstab/465/base 2025-12-04T08:57:04.0456466Z * [new branch] gh/wconstab/465/head -> origin/gh/wconstab/465/head 2025-12-04T08:57:04.0458010Z * [new branch] gh/wconstab/465/orig -> origin/gh/wconstab/465/orig 2025-12-04T08:57:04.0460261Z * [new branch] gh/wconstab/466/base -> origin/gh/wconstab/466/base 2025-12-04T08:57:04.0461811Z * [new branch] gh/wconstab/466/head -> origin/gh/wconstab/466/head 2025-12-04T08:57:04.0463309Z * [new branch] gh/wconstab/466/orig -> origin/gh/wconstab/466/orig 2025-12-04T08:57:04.0465849Z * [new branch] gh/wconstab/467/base -> origin/gh/wconstab/467/base 2025-12-04T08:57:04.0467499Z * [new branch] gh/wconstab/467/head -> origin/gh/wconstab/467/head 2025-12-04T08:57:04.0469006Z * [new branch] gh/wconstab/467/orig -> origin/gh/wconstab/467/orig 2025-12-04T08:57:04.0471021Z * [new branch] gh/wconstab/468/base -> origin/gh/wconstab/468/base 2025-12-04T08:57:04.0472621Z * [new branch] gh/wconstab/468/head -> origin/gh/wconstab/468/head 2025-12-04T08:57:04.0474153Z * [new branch] gh/wconstab/468/orig -> origin/gh/wconstab/468/orig 2025-12-04T08:57:04.0476675Z * [new branch] gh/weifengpy/39/base -> origin/gh/weifengpy/39/base 2025-12-04T08:57:04.0478248Z * [new branch] gh/weifengpy/39/head -> origin/gh/weifengpy/39/head 2025-12-04T08:57:04.0479828Z * [new branch] gh/weifengpy/39/orig -> origin/gh/weifengpy/39/orig 2025-12-04T08:57:04.0482406Z * [new branch] gh/weifengpy/40/base -> origin/gh/weifengpy/40/base 2025-12-04T08:57:04.0484011Z * [new branch] gh/weifengpy/40/head -> origin/gh/weifengpy/40/head 2025-12-04T08:57:04.0485664Z * [new branch] gh/weifengpy/40/orig -> origin/gh/weifengpy/40/orig 2025-12-04T08:57:04.0487734Z * [new branch] gh/weifengpy/41/base -> origin/gh/weifengpy/41/base 2025-12-04T08:57:04.0489343Z * [new branch] gh/weifengpy/41/head -> origin/gh/weifengpy/41/head 2025-12-04T08:57:04.0491023Z * [new branch] gh/weifengpy/41/orig -> origin/gh/weifengpy/41/orig 2025-12-04T08:57:04.0493624Z * [new branch] gh/williamwen42/250/base -> origin/gh/williamwen42/250/base 2025-12-04T08:57:04.0495349Z * [new branch] gh/williamwen42/250/head -> origin/gh/williamwen42/250/head 2025-12-04T08:57:04.0496819Z * [new branch] gh/williamwen42/250/orig -> origin/gh/williamwen42/250/orig 2025-12-04T08:57:04.0499055Z * [new branch] gh/williamwen42/279/base -> origin/gh/williamwen42/279/base 2025-12-04T08:57:04.0500756Z * [new branch] gh/williamwen42/279/head -> origin/gh/williamwen42/279/head 2025-12-04T08:57:04.0502285Z * [new branch] gh/williamwen42/279/orig -> origin/gh/williamwen42/279/orig 2025-12-04T08:57:04.0504970Z * [new branch] gh/williamwen42/282/base -> origin/gh/williamwen42/282/base 2025-12-04T08:57:04.0506803Z * [new branch] gh/williamwen42/282/head -> origin/gh/williamwen42/282/head 2025-12-04T08:57:04.0508378Z * [new branch] gh/williamwen42/282/orig -> origin/gh/williamwen42/282/orig 2025-12-04T08:57:04.0510499Z * [new branch] gh/williamwen42/287/base -> origin/gh/williamwen42/287/base 2025-12-04T08:57:04.0512142Z * [new branch] gh/williamwen42/287/head -> origin/gh/williamwen42/287/head 2025-12-04T08:57:04.0513719Z * [new branch] gh/williamwen42/287/orig -> origin/gh/williamwen42/287/orig 2025-12-04T08:57:04.0515776Z * [new branch] gh/williamwen42/288/base -> origin/gh/williamwen42/288/base 2025-12-04T08:57:04.0517640Z * [new branch] gh/williamwen42/288/head -> origin/gh/williamwen42/288/head 2025-12-04T08:57:04.0518906Z * [new branch] gh/williamwen42/288/orig -> origin/gh/williamwen42/288/orig 2025-12-04T08:57:04.0521205Z * [new branch] gh/williamwen42/296/base -> origin/gh/williamwen42/296/base 2025-12-04T08:57:04.0522900Z * [new branch] gh/williamwen42/296/head -> origin/gh/williamwen42/296/head 2025-12-04T08:57:04.0524595Z * [new branch] gh/williamwen42/296/orig -> origin/gh/williamwen42/296/orig 2025-12-04T08:57:04.0526508Z * [new branch] gh/williamwen42/297/base -> origin/gh/williamwen42/297/base 2025-12-04T08:57:04.0528179Z * [new branch] gh/williamwen42/297/head -> origin/gh/williamwen42/297/head 2025-12-04T08:57:04.0529734Z * [new branch] gh/williamwen42/297/orig -> origin/gh/williamwen42/297/orig 2025-12-04T08:57:04.0531857Z * [new branch] gh/williamwen42/306/base -> origin/gh/williamwen42/306/base 2025-12-04T08:57:04.0533606Z * [new branch] gh/williamwen42/306/head -> origin/gh/williamwen42/306/head 2025-12-04T08:57:04.0535284Z * [new branch] gh/williamwen42/306/orig -> origin/gh/williamwen42/306/orig 2025-12-04T08:57:04.0537388Z * [new branch] gh/williamwen42/309/base -> origin/gh/williamwen42/309/base 2025-12-04T08:57:04.0539067Z * [new branch] gh/williamwen42/309/head -> origin/gh/williamwen42/309/head 2025-12-04T08:57:04.0540413Z * [new branch] gh/williamwen42/309/orig -> origin/gh/williamwen42/309/orig 2025-12-04T08:57:04.0542603Z * [new branch] gh/williamwen42/310/base -> origin/gh/williamwen42/310/base 2025-12-04T08:57:04.0544251Z * [new branch] gh/williamwen42/310/head -> origin/gh/williamwen42/310/head 2025-12-04T08:57:04.0545711Z * [new branch] gh/williamwen42/310/orig -> origin/gh/williamwen42/310/orig 2025-12-04T08:57:04.0548847Z * [new branch] gh/williamwen42/311/base -> origin/gh/williamwen42/311/base 2025-12-04T08:57:04.0550496Z * [new branch] gh/williamwen42/311/head -> origin/gh/williamwen42/311/head 2025-12-04T08:57:04.0552051Z * [new branch] gh/williamwen42/311/orig -> origin/gh/williamwen42/311/orig 2025-12-04T08:57:04.0554142Z * [new branch] gh/williamwen42/319/base -> origin/gh/williamwen42/319/base 2025-12-04T08:57:04.0555812Z * [new branch] gh/williamwen42/319/head -> origin/gh/williamwen42/319/head 2025-12-04T08:57:04.0557375Z * [new branch] gh/williamwen42/319/orig -> origin/gh/williamwen42/319/orig 2025-12-04T08:57:04.0559461Z * [new branch] gh/williamwen42/325/base -> origin/gh/williamwen42/325/base 2025-12-04T08:57:04.0561142Z * [new branch] gh/williamwen42/325/head -> origin/gh/williamwen42/325/head 2025-12-04T08:57:04.0563071Z * [new branch] gh/williamwen42/325/orig -> origin/gh/williamwen42/325/orig 2025-12-04T08:57:04.0565136Z * [new branch] gh/williamwen42/326/base -> origin/gh/williamwen42/326/base 2025-12-04T08:57:04.0566832Z * [new branch] gh/williamwen42/326/head -> origin/gh/williamwen42/326/head 2025-12-04T08:57:04.0568420Z * [new branch] gh/williamwen42/326/orig -> origin/gh/williamwen42/326/orig 2025-12-04T08:57:04.0570510Z * [new branch] gh/williamwen42/327/base -> origin/gh/williamwen42/327/base 2025-12-04T08:57:04.0572181Z * [new branch] gh/williamwen42/327/head -> origin/gh/williamwen42/327/head 2025-12-04T08:57:04.0573835Z * [new branch] gh/williamwen42/327/orig -> origin/gh/williamwen42/327/orig 2025-12-04T08:57:04.0575940Z * [new branch] gh/williamwen42/328/base -> origin/gh/williamwen42/328/base 2025-12-04T08:57:04.0577565Z * [new branch] gh/williamwen42/328/head -> origin/gh/williamwen42/328/head 2025-12-04T08:57:04.0579221Z * [new branch] gh/williamwen42/328/orig -> origin/gh/williamwen42/328/orig 2025-12-04T08:57:04.0581789Z * [new branch] gh/williamwen42/329/base -> origin/gh/williamwen42/329/base 2025-12-04T08:57:04.0583532Z * [new branch] gh/williamwen42/329/head -> origin/gh/williamwen42/329/head 2025-12-04T08:57:04.0585090Z * [new branch] gh/williamwen42/329/orig -> origin/gh/williamwen42/329/orig 2025-12-04T08:57:04.0587666Z * [new branch] gh/williamwen42/330/base -> origin/gh/williamwen42/330/base 2025-12-04T08:57:04.0589343Z * [new branch] gh/williamwen42/330/head -> origin/gh/williamwen42/330/head 2025-12-04T08:57:04.0590874Z * [new branch] gh/williamwen42/330/orig -> origin/gh/williamwen42/330/orig 2025-12-04T08:57:04.0592989Z * [new branch] gh/williamwen42/331/base -> origin/gh/williamwen42/331/base 2025-12-04T08:57:04.0594723Z * [new branch] gh/williamwen42/331/head -> origin/gh/williamwen42/331/head 2025-12-04T08:57:04.0596374Z * [new branch] gh/williamwen42/331/orig -> origin/gh/williamwen42/331/orig 2025-12-04T08:57:04.0598368Z * [new branch] gh/williamwen42/332/base -> origin/gh/williamwen42/332/base 2025-12-04T08:57:04.0600000Z * [new branch] gh/williamwen42/332/head -> origin/gh/williamwen42/332/head 2025-12-04T08:57:04.0601551Z * [new branch] gh/williamwen42/332/orig -> origin/gh/williamwen42/332/orig 2025-12-04T08:57:04.0604321Z * [new branch] gh/williamwen42/333/base -> origin/gh/williamwen42/333/base 2025-12-04T08:57:04.0608018Z * [new branch] gh/williamwen42/333/head -> origin/gh/williamwen42/333/head 2025-12-04T08:57:04.0609743Z * [new branch] gh/williamwen42/333/orig -> origin/gh/williamwen42/333/orig 2025-12-04T08:57:04.0611934Z * [new branch] gh/williamwen42/334/base -> origin/gh/williamwen42/334/base 2025-12-04T08:57:04.0613695Z * [new branch] gh/williamwen42/334/head -> origin/gh/williamwen42/334/head 2025-12-04T08:57:04.0615297Z * [new branch] gh/williamwen42/334/orig -> origin/gh/williamwen42/334/orig 2025-12-04T08:57:04.0617486Z * [new branch] gh/williamwen42/335/base -> origin/gh/williamwen42/335/base 2025-12-04T08:57:04.0622178Z * [new branch] gh/williamwen42/335/head -> origin/gh/williamwen42/335/head 2025-12-04T08:57:04.0623788Z * [new branch] gh/williamwen42/335/orig -> origin/gh/williamwen42/335/orig 2025-12-04T08:57:04.0626082Z * [new branch] gh/williamwen42/336/base -> origin/gh/williamwen42/336/base 2025-12-04T08:57:04.0627555Z * [new branch] gh/williamwen42/336/head -> origin/gh/williamwen42/336/head 2025-12-04T08:57:04.0629104Z * [new branch] gh/williamwen42/336/orig -> origin/gh/williamwen42/336/orig 2025-12-04T08:57:04.0631379Z * [new branch] gh/williamwen42/337/base -> origin/gh/williamwen42/337/base 2025-12-04T08:57:04.0632828Z * [new branch] gh/williamwen42/337/head -> origin/gh/williamwen42/337/head 2025-12-04T08:57:04.0634259Z * [new branch] gh/williamwen42/337/orig -> origin/gh/williamwen42/337/orig 2025-12-04T08:57:04.0636606Z * [new branch] gh/williamwen42/338/base -> origin/gh/williamwen42/338/base 2025-12-04T08:57:04.0638223Z * [new branch] gh/williamwen42/338/head -> origin/gh/williamwen42/338/head 2025-12-04T08:57:04.0639764Z * [new branch] gh/williamwen42/338/orig -> origin/gh/williamwen42/338/orig 2025-12-04T08:57:04.0642278Z * [new branch] gh/williamwen42/339/base -> origin/gh/williamwen42/339/base 2025-12-04T08:57:04.0644007Z * [new branch] gh/williamwen42/339/head -> origin/gh/williamwen42/339/head 2025-12-04T08:57:04.0645574Z * [new branch] gh/williamwen42/339/orig -> origin/gh/williamwen42/339/orig 2025-12-04T08:57:04.0647979Z * [new branch] gh/williamwen42/340/base -> origin/gh/williamwen42/340/base 2025-12-04T08:57:04.0649285Z * [new branch] gh/williamwen42/340/head -> origin/gh/williamwen42/340/head 2025-12-04T08:57:04.0650953Z * [new branch] gh/williamwen42/340/orig -> origin/gh/williamwen42/340/orig 2025-12-04T08:57:04.0653136Z * [new branch] gh/williamwen42/341/base -> origin/gh/williamwen42/341/base 2025-12-04T08:57:04.0654968Z * [new branch] gh/williamwen42/341/head -> origin/gh/williamwen42/341/head 2025-12-04T08:57:04.0656415Z * [new branch] gh/williamwen42/341/orig -> origin/gh/williamwen42/341/orig 2025-12-04T08:57:04.0658593Z * [new branch] gh/williamwen42/342/base -> origin/gh/williamwen42/342/base 2025-12-04T08:57:04.0660173Z * [new branch] gh/williamwen42/342/head -> origin/gh/williamwen42/342/head 2025-12-04T08:57:04.0661675Z * [new branch] gh/williamwen42/342/orig -> origin/gh/williamwen42/342/orig 2025-12-04T08:57:04.0663864Z * [new branch] gh/williamwen42/343/base -> origin/gh/williamwen42/343/base 2025-12-04T08:57:04.0665662Z * [new branch] gh/williamwen42/343/head -> origin/gh/williamwen42/343/head 2025-12-04T08:57:04.0667039Z * [new branch] gh/williamwen42/343/orig -> origin/gh/williamwen42/343/orig 2025-12-04T08:57:04.0669344Z * [new branch] gh/williamwen42/344/base -> origin/gh/williamwen42/344/base 2025-12-04T08:57:04.0670946Z * [new branch] gh/williamwen42/344/head -> origin/gh/williamwen42/344/head 2025-12-04T08:57:04.0672542Z * [new branch] gh/williamwen42/344/orig -> origin/gh/williamwen42/344/orig 2025-12-04T08:57:04.0674679Z * [new branch] gh/williamwen42/345/base -> origin/gh/williamwen42/345/base 2025-12-04T08:57:04.0676358Z * [new branch] gh/williamwen42/345/head -> origin/gh/williamwen42/345/head 2025-12-04T08:57:04.0677930Z * [new branch] gh/williamwen42/345/orig -> origin/gh/williamwen42/345/orig 2025-12-04T08:57:04.0680069Z * [new branch] gh/williamwen42/346/base -> origin/gh/williamwen42/346/base 2025-12-04T08:57:04.0681670Z * [new branch] gh/williamwen42/346/head -> origin/gh/williamwen42/346/head 2025-12-04T08:57:04.0683134Z * [new branch] gh/williamwen42/346/orig -> origin/gh/williamwen42/346/orig 2025-12-04T08:57:04.0685411Z * [new branch] gh/williamwen42/347/base -> origin/gh/williamwen42/347/base 2025-12-04T08:57:04.0687076Z * [new branch] gh/williamwen42/347/head -> origin/gh/williamwen42/347/head 2025-12-04T08:57:04.0688536Z * [new branch] gh/williamwen42/347/orig -> origin/gh/williamwen42/347/orig 2025-12-04T08:57:04.0690668Z * [new branch] gh/williamwen42/348/base -> origin/gh/williamwen42/348/base 2025-12-04T08:57:04.0692248Z * [new branch] gh/williamwen42/348/head -> origin/gh/williamwen42/348/head 2025-12-04T08:57:04.0693809Z * [new branch] gh/williamwen42/348/orig -> origin/gh/williamwen42/348/orig 2025-12-04T08:57:04.0695850Z * [new branch] gh/williamwen42/349/base -> origin/gh/williamwen42/349/base 2025-12-04T08:57:04.0697428Z * [new branch] gh/williamwen42/349/head -> origin/gh/williamwen42/349/head 2025-12-04T08:57:04.0698907Z * [new branch] gh/williamwen42/349/orig -> origin/gh/williamwen42/349/orig 2025-12-04T08:57:04.0701142Z * [new branch] gh/williamwen42/350/base -> origin/gh/williamwen42/350/base 2025-12-04T08:57:04.0702814Z * [new branch] gh/williamwen42/350/head -> origin/gh/williamwen42/350/head 2025-12-04T08:57:04.0704397Z * [new branch] gh/williamwen42/350/orig -> origin/gh/williamwen42/350/orig 2025-12-04T08:57:04.0707082Z * [new branch] gh/williamwen42/351/base -> origin/gh/williamwen42/351/base 2025-12-04T08:57:04.0708716Z * [new branch] gh/williamwen42/351/head -> origin/gh/williamwen42/351/head 2025-12-04T08:57:04.0710143Z * [new branch] gh/williamwen42/351/orig -> origin/gh/williamwen42/351/orig 2025-12-04T08:57:04.0712339Z * [new branch] gh/williamwen42/352/base -> origin/gh/williamwen42/352/base 2025-12-04T08:57:04.0714106Z * [new branch] gh/williamwen42/352/head -> origin/gh/williamwen42/352/head 2025-12-04T08:57:04.0715664Z * [new branch] gh/williamwen42/352/orig -> origin/gh/williamwen42/352/orig 2025-12-04T08:57:04.0717862Z * [new branch] gh/williamwen42/353/base -> origin/gh/williamwen42/353/base 2025-12-04T08:57:04.0719527Z * [new branch] gh/williamwen42/353/head -> origin/gh/williamwen42/353/head 2025-12-04T08:57:04.0721101Z * [new branch] gh/williamwen42/353/orig -> origin/gh/williamwen42/353/orig 2025-12-04T08:57:04.0723621Z * [new branch] gh/williamwen42/354/base -> origin/gh/williamwen42/354/base 2025-12-04T08:57:04.0725777Z * [new branch] gh/williamwen42/354/head -> origin/gh/williamwen42/354/head 2025-12-04T08:57:04.0727400Z * [new branch] gh/williamwen42/354/orig -> origin/gh/williamwen42/354/orig 2025-12-04T08:57:04.0729604Z * [new branch] gh/williamwen42/355/base -> origin/gh/williamwen42/355/base 2025-12-04T08:57:04.0731286Z * [new branch] gh/williamwen42/355/head -> origin/gh/williamwen42/355/head 2025-12-04T08:57:04.0732846Z * [new branch] gh/williamwen42/355/orig -> origin/gh/williamwen42/355/orig 2025-12-04T08:57:04.0735144Z * [new branch] gh/williamwen42/356/base -> origin/gh/williamwen42/356/base 2025-12-04T08:57:04.0736721Z * [new branch] gh/williamwen42/356/head -> origin/gh/williamwen42/356/head 2025-12-04T08:57:04.0738291Z * [new branch] gh/williamwen42/356/orig -> origin/gh/williamwen42/356/orig 2025-12-04T08:57:04.0740365Z * [new branch] gh/williamwen42/357/base -> origin/gh/williamwen42/357/base 2025-12-04T08:57:04.0741997Z * [new branch] gh/williamwen42/357/head -> origin/gh/williamwen42/357/head 2025-12-04T08:57:04.0743606Z * [new branch] gh/williamwen42/357/orig -> origin/gh/williamwen42/357/orig 2025-12-04T08:57:04.0745661Z * [new branch] gh/williamwen42/358/base -> origin/gh/williamwen42/358/base 2025-12-04T08:57:04.0747358Z * [new branch] gh/williamwen42/358/head -> origin/gh/williamwen42/358/head 2025-12-04T08:57:04.0749002Z * [new branch] gh/williamwen42/358/orig -> origin/gh/williamwen42/358/orig 2025-12-04T08:57:04.0751608Z * [new branch] gh/xmfan/169/base -> origin/gh/xmfan/169/base 2025-12-04T08:57:04.0753215Z * [new branch] gh/xmfan/169/head -> origin/gh/xmfan/169/head 2025-12-04T08:57:04.0755226Z * [new branch] gh/xmfan/170/base -> origin/gh/xmfan/170/base 2025-12-04T08:57:04.0756755Z * [new branch] gh/xmfan/170/head -> origin/gh/xmfan/170/head 2025-12-04T08:57:04.0758755Z * [new branch] gh/xmfan/274/base -> origin/gh/xmfan/274/base 2025-12-04T08:57:04.0760310Z * [new branch] gh/xmfan/274/head -> origin/gh/xmfan/274/head 2025-12-04T08:57:04.0761845Z * [new branch] gh/xmfan/274/orig -> origin/gh/xmfan/274/orig 2025-12-04T08:57:04.0763899Z * [new branch] gh/xmfan/277/base -> origin/gh/xmfan/277/base 2025-12-04T08:57:04.0765557Z * [new branch] gh/xmfan/277/head -> origin/gh/xmfan/277/head 2025-12-04T08:57:04.0767076Z * [new branch] gh/xmfan/277/orig -> origin/gh/xmfan/277/orig 2025-12-04T08:57:04.0769036Z * [new branch] gh/xmfan/301/base -> origin/gh/xmfan/301/base 2025-12-04T08:57:04.0770725Z * [new branch] gh/xmfan/301/head -> origin/gh/xmfan/301/head 2025-12-04T08:57:04.0772225Z * [new branch] gh/xmfan/301/orig -> origin/gh/xmfan/301/orig 2025-12-04T08:57:04.0774369Z * [new branch] gh/xmfan/304/base -> origin/gh/xmfan/304/base 2025-12-04T08:57:04.0775936Z * [new branch] gh/xmfan/304/head -> origin/gh/xmfan/304/head 2025-12-04T08:57:04.0777545Z * [new branch] gh/xmfan/304/orig -> origin/gh/xmfan/304/orig 2025-12-04T08:57:04.0779601Z * [new branch] gh/xmfan/309/base -> origin/gh/xmfan/309/base 2025-12-04T08:57:04.0781114Z * [new branch] gh/xmfan/309/head -> origin/gh/xmfan/309/head 2025-12-04T08:57:04.0782665Z * [new branch] gh/xmfan/309/orig -> origin/gh/xmfan/309/orig 2025-12-04T08:57:04.0784664Z * [new branch] gh/xmfan/310/base -> origin/gh/xmfan/310/base 2025-12-04T08:57:04.0786291Z * [new branch] gh/xmfan/310/head -> origin/gh/xmfan/310/head 2025-12-04T08:57:04.0787730Z * [new branch] gh/xmfan/310/orig -> origin/gh/xmfan/310/orig 2025-12-04T08:57:04.0789739Z * [new branch] gh/xmfan/311/base -> origin/gh/xmfan/311/base 2025-12-04T08:57:04.0791334Z * [new branch] gh/xmfan/311/head -> origin/gh/xmfan/311/head 2025-12-04T08:57:04.0792955Z * [new branch] gh/xmfan/311/orig -> origin/gh/xmfan/311/orig 2025-12-04T08:57:04.0794996Z * [new branch] gh/xmfan/312/base -> origin/gh/xmfan/312/base 2025-12-04T08:57:04.0797004Z * [new branch] gh/xmfan/312/head -> origin/gh/xmfan/312/head 2025-12-04T08:57:04.0798636Z * [new branch] gh/xmfan/312/orig -> origin/gh/xmfan/312/orig 2025-12-04T08:57:04.0800660Z * [new branch] gh/xmfan/313/base -> origin/gh/xmfan/313/base 2025-12-04T08:57:04.0802265Z * [new branch] gh/xmfan/313/head -> origin/gh/xmfan/313/head 2025-12-04T08:57:04.0803943Z * [new branch] gh/xmfan/313/orig -> origin/gh/xmfan/313/orig 2025-12-04T08:57:04.0807206Z * [new branch] gh/xuanzhang816/27/base -> origin/gh/xuanzhang816/27/base 2025-12-04T08:57:04.0808814Z * [new branch] gh/xuanzhang816/27/head -> origin/gh/xuanzhang816/27/head 2025-12-04T08:57:04.0810244Z * [new branch] gh/xuanzhang816/27/orig -> origin/gh/xuanzhang816/27/orig 2025-12-04T08:57:04.0812454Z * [new branch] gh/xuanzhang816/32/base -> origin/gh/xuanzhang816/32/base 2025-12-04T08:57:04.0814233Z * [new branch] gh/xuanzhang816/32/head -> origin/gh/xuanzhang816/32/head 2025-12-04T08:57:04.0815703Z * [new branch] gh/xuanzhang816/32/orig -> origin/gh/xuanzhang816/32/orig 2025-12-04T08:57:04.0817738Z * [new branch] gh/xuanzhang816/33/base -> origin/gh/xuanzhang816/33/base 2025-12-04T08:57:04.0819403Z * [new branch] gh/xuanzhang816/33/head -> origin/gh/xuanzhang816/33/head 2025-12-04T08:57:04.0820997Z * [new branch] gh/xuanzhang816/33/orig -> origin/gh/xuanzhang816/33/orig 2025-12-04T08:57:04.0823242Z * [new branch] gh/xuanzhang816/34/base -> origin/gh/xuanzhang816/34/base 2025-12-04T08:57:04.0824828Z * [new branch] gh/xuanzhang816/34/head -> origin/gh/xuanzhang816/34/head 2025-12-04T08:57:04.0826404Z * [new branch] gh/xuanzhang816/34/orig -> origin/gh/xuanzhang816/34/orig 2025-12-04T08:57:04.0828756Z * [new branch] gh/xuanzhang816/35/base -> origin/gh/xuanzhang816/35/base 2025-12-04T08:57:04.0830245Z * [new branch] gh/xuanzhang816/35/head -> origin/gh/xuanzhang816/35/head 2025-12-04T08:57:04.0831856Z * [new branch] gh/xuanzhang816/35/orig -> origin/gh/xuanzhang816/35/orig 2025-12-04T08:57:04.0834538Z * [new branch] gh/yanbing-j/11/base -> origin/gh/yanbing-j/11/base 2025-12-04T08:57:04.0835936Z * [new branch] gh/yanbing-j/11/head -> origin/gh/yanbing-j/11/head 2025-12-04T08:57:04.0837896Z * [new branch] gh/yanbing-j/11/orig -> origin/gh/yanbing-j/11/orig 2025-12-04T08:57:04.0839967Z * [new branch] gh/yanbing-j/12/base -> origin/gh/yanbing-j/12/base 2025-12-04T08:57:04.0841579Z * [new branch] gh/yanbing-j/12/head -> origin/gh/yanbing-j/12/head 2025-12-04T08:57:04.0843149Z * [new branch] gh/yanbing-j/12/orig -> origin/gh/yanbing-j/12/orig 2025-12-04T08:57:04.0845193Z * [new branch] gh/yanbing-j/13/base -> origin/gh/yanbing-j/13/base 2025-12-04T08:57:04.0846813Z * [new branch] gh/yanbing-j/13/head -> origin/gh/yanbing-j/13/head 2025-12-04T08:57:04.0848368Z * [new branch] gh/yanbing-j/13/orig -> origin/gh/yanbing-j/13/orig 2025-12-04T08:57:04.0850438Z * [new branch] gh/yanbing-j/14/base -> origin/gh/yanbing-j/14/base 2025-12-04T08:57:04.0851950Z * [new branch] gh/yanbing-j/14/head -> origin/gh/yanbing-j/14/head 2025-12-04T08:57:04.0853615Z * [new branch] gh/yanbing-j/14/orig -> origin/gh/yanbing-j/14/orig 2025-12-04T08:57:04.0855699Z * [new branch] gh/yanbing-j/15/base -> origin/gh/yanbing-j/15/base 2025-12-04T08:57:04.0857319Z * [new branch] gh/yanbing-j/15/head -> origin/gh/yanbing-j/15/head 2025-12-04T08:57:04.0858846Z * [new branch] gh/yanbing-j/15/orig -> origin/gh/yanbing-j/15/orig 2025-12-04T08:57:04.0860792Z * [new branch] gh/yanbing-j/18/base -> origin/gh/yanbing-j/18/base 2025-12-04T08:57:04.0862404Z * [new branch] gh/yanbing-j/18/head -> origin/gh/yanbing-j/18/head 2025-12-04T08:57:04.0863944Z * [new branch] gh/yanbing-j/18/orig -> origin/gh/yanbing-j/18/orig 2025-12-04T08:57:04.0865981Z * [new branch] gh/yanbing-j/19/base -> origin/gh/yanbing-j/19/base 2025-12-04T08:57:04.0867585Z * [new branch] gh/yanbing-j/19/head -> origin/gh/yanbing-j/19/head 2025-12-04T08:57:04.0869116Z * [new branch] gh/yanbing-j/19/orig -> origin/gh/yanbing-j/19/orig 2025-12-04T08:57:04.0871164Z * [new branch] gh/yanbing-j/20/base -> origin/gh/yanbing-j/20/base 2025-12-04T08:57:04.0872722Z * [new branch] gh/yanbing-j/20/head -> origin/gh/yanbing-j/20/head 2025-12-04T08:57:04.0874278Z * [new branch] gh/yanbing-j/20/orig -> origin/gh/yanbing-j/20/orig 2025-12-04T08:57:04.0876409Z * [new branch] gh/yanbing-j/21/base -> origin/gh/yanbing-j/21/base 2025-12-04T08:57:04.0878436Z * [new branch] gh/yanbing-j/21/head -> origin/gh/yanbing-j/21/head 2025-12-04T08:57:04.0880464Z * [new branch] gh/yanbing-j/22/base -> origin/gh/yanbing-j/22/base 2025-12-04T08:57:04.0882113Z * [new branch] gh/yanbing-j/22/head -> origin/gh/yanbing-j/22/head 2025-12-04T08:57:04.0883570Z * [new branch] gh/yanbing-j/22/orig -> origin/gh/yanbing-j/22/orig 2025-12-04T08:57:04.0885810Z * [new branch] gh/yanbing-j/23/base -> origin/gh/yanbing-j/23/base 2025-12-04T08:57:04.0887424Z * [new branch] gh/yanbing-j/23/head -> origin/gh/yanbing-j/23/head 2025-12-04T08:57:04.0888970Z * [new branch] gh/yanbing-j/23/orig -> origin/gh/yanbing-j/23/orig 2025-12-04T08:57:04.0891035Z * [new branch] gh/yanbing-j/24/base -> origin/gh/yanbing-j/24/base 2025-12-04T08:57:04.0892618Z * [new branch] gh/yanbing-j/24/head -> origin/gh/yanbing-j/24/head 2025-12-04T08:57:04.0894296Z * [new branch] gh/yanbing-j/24/orig -> origin/gh/yanbing-j/24/orig 2025-12-04T08:57:04.0896392Z * [new branch] gh/yanbing-j/25/base -> origin/gh/yanbing-j/25/base 2025-12-04T08:57:04.0898022Z * [new branch] gh/yanbing-j/25/head -> origin/gh/yanbing-j/25/head 2025-12-04T08:57:04.0899416Z * [new branch] gh/yanbing-j/25/orig -> origin/gh/yanbing-j/25/orig 2025-12-04T08:57:04.0901502Z * [new branch] gh/yanbing-j/26/base -> origin/gh/yanbing-j/26/base 2025-12-04T08:57:04.0903221Z * [new branch] gh/yanbing-j/26/head -> origin/gh/yanbing-j/26/head 2025-12-04T08:57:04.0905255Z * [new branch] gh/yanbing-j/26/orig -> origin/gh/yanbing-j/26/orig 2025-12-04T08:57:04.0907922Z * [new branch] gh/yang-yu-hang/1/base -> origin/gh/yang-yu-hang/1/base 2025-12-04T08:57:04.0909643Z * [new branch] gh/yang-yu-hang/1/head -> origin/gh/yang-yu-hang/1/head 2025-12-04T08:57:04.0911363Z * [new branch] gh/yang-yu-hang/1/orig -> origin/gh/yang-yu-hang/1/orig 2025-12-04T08:57:04.0913596Z * [new branch] gh/yang-yu-hang/2/base -> origin/gh/yang-yu-hang/2/base 2025-12-04T08:57:04.0915550Z * [new branch] gh/yang-yu-hang/2/head -> origin/gh/yang-yu-hang/2/head 2025-12-04T08:57:04.0917302Z * [new branch] gh/yang-yu-hang/2/orig -> origin/gh/yang-yu-hang/2/orig 2025-12-04T08:57:04.0919393Z * [new branch] gh/yang-yu-hang/3/base -> origin/gh/yang-yu-hang/3/base 2025-12-04T08:57:04.0921054Z * [new branch] gh/yang-yu-hang/3/head -> origin/gh/yang-yu-hang/3/head 2025-12-04T08:57:04.0922545Z * [new branch] gh/yang-yu-hang/3/orig -> origin/gh/yang-yu-hang/3/orig 2025-12-04T08:57:04.0924991Z * [new branch] gh/yangw-dev/12/base -> origin/gh/yangw-dev/12/base 2025-12-04T08:57:04.0926621Z * [new branch] gh/yangw-dev/12/head -> origin/gh/yangw-dev/12/head 2025-12-04T08:57:04.0928195Z * [new branch] gh/yangw-dev/12/orig -> origin/gh/yangw-dev/12/orig 2025-12-04T08:57:04.0930157Z * [new branch] gh/yangw-dev/13/base -> origin/gh/yangw-dev/13/base 2025-12-04T08:57:04.0931871Z * [new branch] gh/yangw-dev/13/head -> origin/gh/yangw-dev/13/head 2025-12-04T08:57:04.0933612Z * [new branch] gh/yangw-dev/13/orig -> origin/gh/yangw-dev/13/orig 2025-12-04T08:57:04.0935738Z * [new branch] gh/yangw-dev/14/base -> origin/gh/yangw-dev/14/base 2025-12-04T08:57:04.0937284Z * [new branch] gh/yangw-dev/14/head -> origin/gh/yangw-dev/14/head 2025-12-04T08:57:04.0938825Z * [new branch] gh/yangw-dev/14/orig -> origin/gh/yangw-dev/14/orig 2025-12-04T08:57:04.0940935Z * [new branch] gh/yangw-dev/15/base -> origin/gh/yangw-dev/15/base 2025-12-04T08:57:04.0942647Z * [new branch] gh/yangw-dev/15/head -> origin/gh/yangw-dev/15/head 2025-12-04T08:57:04.0944139Z * [new branch] gh/yangw-dev/15/orig -> origin/gh/yangw-dev/15/orig 2025-12-04T08:57:04.0946101Z * [new branch] gh/yangw-dev/19/base -> origin/gh/yangw-dev/19/base 2025-12-04T08:57:04.0947726Z * [new branch] gh/yangw-dev/19/head -> origin/gh/yangw-dev/19/head 2025-12-04T08:57:04.0949354Z * [new branch] gh/yangw-dev/19/orig -> origin/gh/yangw-dev/19/orig 2025-12-04T08:57:04.0951325Z * [new branch] gh/yangw-dev/26/base -> origin/gh/yangw-dev/26/base 2025-12-04T08:57:04.0953342Z * [new branch] gh/yangw-dev/26/head -> origin/gh/yangw-dev/26/head 2025-12-04T08:57:04.0954937Z * [new branch] gh/yangw-dev/26/orig -> origin/gh/yangw-dev/26/orig 2025-12-04T08:57:04.0956994Z * [new branch] gh/yangw-dev/27/base -> origin/gh/yangw-dev/27/base 2025-12-04T08:57:04.0958546Z * [new branch] gh/yangw-dev/27/head -> origin/gh/yangw-dev/27/head 2025-12-04T08:57:04.0960231Z * [new branch] gh/yangw-dev/27/orig -> origin/gh/yangw-dev/27/orig 2025-12-04T08:57:04.0962779Z * [new branch] gh/ydwu4/292/base -> origin/gh/ydwu4/292/base 2025-12-04T08:57:04.0964189Z * [new branch] gh/ydwu4/292/head -> origin/gh/ydwu4/292/head 2025-12-04T08:57:04.0965779Z * [new branch] gh/ydwu4/292/orig -> origin/gh/ydwu4/292/orig 2025-12-04T08:57:04.0967882Z * [new branch] gh/ydwu4/294/base -> origin/gh/ydwu4/294/base 2025-12-04T08:57:04.0969492Z * [new branch] gh/ydwu4/294/head -> origin/gh/ydwu4/294/head 2025-12-04T08:57:04.0971000Z * [new branch] gh/ydwu4/294/orig -> origin/gh/ydwu4/294/orig 2025-12-04T08:57:04.0973268Z * [new branch] gh/ydwu4/295/base -> origin/gh/ydwu4/295/base 2025-12-04T08:57:04.0975388Z * [new branch] gh/ydwu4/295/head -> origin/gh/ydwu4/295/head 2025-12-04T08:57:04.0977051Z * [new branch] gh/ydwu4/295/orig -> origin/gh/ydwu4/295/orig 2025-12-04T08:57:04.0978951Z * [new branch] gh/ydwu4/296/base -> origin/gh/ydwu4/296/base 2025-12-04T08:57:04.0980507Z * [new branch] gh/ydwu4/296/head -> origin/gh/ydwu4/296/head 2025-12-04T08:57:04.0982158Z * [new branch] gh/ydwu4/296/orig -> origin/gh/ydwu4/296/orig 2025-12-04T08:57:04.0984268Z * [new branch] gh/ydwu4/306/base -> origin/gh/ydwu4/306/base 2025-12-04T08:57:04.0986015Z * [new branch] gh/ydwu4/306/head -> origin/gh/ydwu4/306/head 2025-12-04T08:57:04.0987623Z * [new branch] gh/ydwu4/306/orig -> origin/gh/ydwu4/306/orig 2025-12-04T08:57:04.0989678Z * [new branch] gh/ydwu4/312/base -> origin/gh/ydwu4/312/base 2025-12-04T08:57:04.0991264Z * [new branch] gh/ydwu4/312/head -> origin/gh/ydwu4/312/head 2025-12-04T08:57:04.0992810Z * [new branch] gh/ydwu4/312/orig -> origin/gh/ydwu4/312/orig 2025-12-04T08:57:04.0994817Z * [new branch] gh/ydwu4/322/base -> origin/gh/ydwu4/322/base 2025-12-04T08:57:04.0996395Z * [new branch] gh/ydwu4/322/head -> origin/gh/ydwu4/322/head 2025-12-04T08:57:04.0997966Z * [new branch] gh/ydwu4/322/orig -> origin/gh/ydwu4/322/orig 2025-12-04T08:57:04.0999976Z * [new branch] gh/ydwu4/327/base -> origin/gh/ydwu4/327/base 2025-12-04T08:57:04.1001616Z * [new branch] gh/ydwu4/327/head -> origin/gh/ydwu4/327/head 2025-12-04T08:57:04.1003238Z * [new branch] gh/ydwu4/327/orig -> origin/gh/ydwu4/327/orig 2025-12-04T08:57:04.1007614Z * [new branch] gh/ydwu4/328/base -> origin/gh/ydwu4/328/base 2025-12-04T08:57:04.1009223Z * [new branch] gh/ydwu4/328/head -> origin/gh/ydwu4/328/head 2025-12-04T08:57:04.1010669Z * [new branch] gh/ydwu4/328/orig -> origin/gh/ydwu4/328/orig 2025-12-04T08:57:04.1012608Z * [new branch] gh/ydwu4/329/base -> origin/gh/ydwu4/329/base 2025-12-04T08:57:04.1014351Z * [new branch] gh/ydwu4/329/head -> origin/gh/ydwu4/329/head 2025-12-04T08:57:04.1015876Z * [new branch] gh/ydwu4/329/orig -> origin/gh/ydwu4/329/orig 2025-12-04T08:57:04.1017946Z * [new branch] gh/ydwu4/330/base -> origin/gh/ydwu4/330/base 2025-12-04T08:57:04.1019490Z * [new branch] gh/ydwu4/330/head -> origin/gh/ydwu4/330/head 2025-12-04T08:57:04.1021027Z * [new branch] gh/ydwu4/330/orig -> origin/gh/ydwu4/330/orig 2025-12-04T08:57:04.1022985Z * [new branch] gh/ydwu4/331/base -> origin/gh/ydwu4/331/base 2025-12-04T08:57:04.1024630Z * [new branch] gh/ydwu4/331/head -> origin/gh/ydwu4/331/head 2025-12-04T08:57:04.1026429Z * [new branch] gh/ydwu4/331/orig -> origin/gh/ydwu4/331/orig 2025-12-04T08:57:04.1028267Z * [new branch] gh/ydwu4/332/base -> origin/gh/ydwu4/332/base 2025-12-04T08:57:04.1029990Z * [new branch] gh/ydwu4/332/head -> origin/gh/ydwu4/332/head 2025-12-04T08:57:04.1031088Z * [new branch] gh/ydwu4/332/orig -> origin/gh/ydwu4/332/orig 2025-12-04T08:57:04.1033125Z * [new branch] gh/ydwu4/333/base -> origin/gh/ydwu4/333/base 2025-12-04T08:57:04.1034724Z * [new branch] gh/ydwu4/333/head -> origin/gh/ydwu4/333/head 2025-12-04T08:57:04.1036113Z * [new branch] gh/ydwu4/333/orig -> origin/gh/ydwu4/333/orig 2025-12-04T08:57:04.1038217Z * [new branch] gh/ydwu4/334/base -> origin/gh/ydwu4/334/base 2025-12-04T08:57:04.1039813Z * [new branch] gh/ydwu4/334/head -> origin/gh/ydwu4/334/head 2025-12-04T08:57:04.1041408Z * [new branch] gh/ydwu4/334/orig -> origin/gh/ydwu4/334/orig 2025-12-04T08:57:04.1043274Z * [new branch] gh/ydwu4/335/base -> origin/gh/ydwu4/335/base 2025-12-04T08:57:04.1044727Z * [new branch] gh/ydwu4/335/head -> origin/gh/ydwu4/335/head 2025-12-04T08:57:04.1046335Z * [new branch] gh/ydwu4/335/orig -> origin/gh/ydwu4/335/orig 2025-12-04T08:57:04.1048899Z * [new branch] gh/ydwu4/337/base -> origin/gh/ydwu4/337/base 2025-12-04T08:57:04.1050384Z * [new branch] gh/ydwu4/337/head -> origin/gh/ydwu4/337/head 2025-12-04T08:57:04.1051952Z * [new branch] gh/ydwu4/337/orig -> origin/gh/ydwu4/337/orig 2025-12-04T08:57:04.1054801Z * [new branch] gh/ydwu4/339/base -> origin/gh/ydwu4/339/base 2025-12-04T08:57:04.1056334Z * [new branch] gh/ydwu4/339/head -> origin/gh/ydwu4/339/head 2025-12-04T08:57:04.1057819Z * [new branch] gh/ydwu4/339/orig -> origin/gh/ydwu4/339/orig 2025-12-04T08:57:04.1060379Z * [new branch] gh/yf225/133/base -> origin/gh/yf225/133/base 2025-12-04T08:57:04.1061894Z * [new branch] gh/yf225/133/head -> origin/gh/yf225/133/head 2025-12-04T08:57:04.1064066Z * [new branch] gh/yf225/93/base -> origin/gh/yf225/93/base 2025-12-04T08:57:04.1065979Z * [new branch] gh/yf225/93/head -> origin/gh/yf225/93/head 2025-12-04T08:57:04.1069065Z * [new branch] gh/yifuwang/152/base -> origin/gh/yifuwang/152/base 2025-12-04T08:57:04.1070936Z * [new branch] gh/yifuwang/152/head -> origin/gh/yifuwang/152/head 2025-12-04T08:57:04.1072589Z * [new branch] gh/yifuwang/152/orig -> origin/gh/yifuwang/152/orig 2025-12-04T08:57:04.1074552Z * [new branch] gh/yifuwang/195/base -> origin/gh/yifuwang/195/base 2025-12-04T08:57:04.1076182Z * [new branch] gh/yifuwang/195/head -> origin/gh/yifuwang/195/head 2025-12-04T08:57:04.1077723Z * [new branch] gh/yifuwang/195/orig -> origin/gh/yifuwang/195/orig 2025-12-04T08:57:04.1080287Z * [new branch] gh/yiming0416/1/base -> origin/gh/yiming0416/1/base 2025-12-04T08:57:04.1081842Z * [new branch] gh/yiming0416/1/head -> origin/gh/yiming0416/1/head 2025-12-04T08:57:04.1083789Z * [new branch] gh/yiming0416/2/base -> origin/gh/yiming0416/2/base 2025-12-04T08:57:04.1085266Z * [new branch] gh/yiming0416/2/head -> origin/gh/yiming0416/2/head 2025-12-04T08:57:04.1088301Z * [new branch] gh/yushangdi/1/base -> origin/gh/yushangdi/1/base 2025-12-04T08:57:04.1089970Z * [new branch] gh/yushangdi/1/head -> origin/gh/yushangdi/1/head 2025-12-04T08:57:04.1092195Z * [new branch] gh/yushangdi/10/base -> origin/gh/yushangdi/10/base 2025-12-04T08:57:04.1093812Z * [new branch] gh/yushangdi/10/head -> origin/gh/yushangdi/10/head 2025-12-04T08:57:04.1095619Z * [new branch] gh/yushangdi/10/orig -> origin/gh/yushangdi/10/orig 2025-12-04T08:57:04.1097706Z * [new branch] gh/yushangdi/11/base -> origin/gh/yushangdi/11/base 2025-12-04T08:57:04.1099220Z * [new branch] gh/yushangdi/11/head -> origin/gh/yushangdi/11/head 2025-12-04T08:57:04.1100891Z * [new branch] gh/yushangdi/11/orig -> origin/gh/yushangdi/11/orig 2025-12-04T08:57:04.1102770Z * [new branch] gh/yushangdi/2/base -> origin/gh/yushangdi/2/base 2025-12-04T08:57:04.1104307Z * [new branch] gh/yushangdi/2/head -> origin/gh/yushangdi/2/head 2025-12-04T08:57:04.1106664Z * [new branch] gh/yushangdi/7/base -> origin/gh/yushangdi/7/base 2025-12-04T08:57:04.1108222Z * [new branch] gh/yushangdi/7/head -> origin/gh/yushangdi/7/head 2025-12-04T08:57:04.1109745Z * [new branch] gh/yushangdi/7/orig -> origin/gh/yushangdi/7/orig 2025-12-04T08:57:04.1112052Z * [new branch] gh/yushangdi/8/base -> origin/gh/yushangdi/8/base 2025-12-04T08:57:04.1113795Z * [new branch] gh/yushangdi/8/head -> origin/gh/yushangdi/8/head 2025-12-04T08:57:04.1115411Z * [new branch] gh/yushangdi/8/orig -> origin/gh/yushangdi/8/orig 2025-12-04T08:57:04.1117300Z * [new branch] gh/yushangdi/9/base -> origin/gh/yushangdi/9/base 2025-12-04T08:57:04.1118853Z * [new branch] gh/yushangdi/9/head -> origin/gh/yushangdi/9/head 2025-12-04T08:57:04.1120590Z * [new branch] gh/yushangdi/9/orig -> origin/gh/yushangdi/9/orig 2025-12-04T08:57:04.1123121Z * [new branch] gh/zklaus/19/base -> origin/gh/zklaus/19/base 2025-12-04T08:57:04.1124662Z * [new branch] gh/zklaus/19/head -> origin/gh/zklaus/19/head 2025-12-04T08:57:04.1126185Z * [new branch] gh/zklaus/19/orig -> origin/gh/zklaus/19/orig 2025-12-04T08:57:04.1128322Z * [new branch] gh/zklaus/20/base -> origin/gh/zklaus/20/base 2025-12-04T08:57:04.1129886Z * [new branch] gh/zklaus/20/head -> origin/gh/zklaus/20/head 2025-12-04T08:57:04.1131441Z * [new branch] gh/zklaus/20/orig -> origin/gh/zklaus/20/orig 2025-12-04T08:57:04.1133714Z * [new branch] gh/zklaus/21/base -> origin/gh/zklaus/21/base 2025-12-04T08:57:04.1135258Z * [new branch] gh/zklaus/21/head -> origin/gh/zklaus/21/head 2025-12-04T08:57:04.1136798Z * [new branch] gh/zklaus/21/orig -> origin/gh/zklaus/21/orig 2025-12-04T08:57:04.1138827Z * [new branch] gh/zklaus/22/base -> origin/gh/zklaus/22/base 2025-12-04T08:57:04.1140463Z * [new branch] gh/zklaus/22/head -> origin/gh/zklaus/22/head 2025-12-04T08:57:04.1142019Z * [new branch] gh/zklaus/22/orig -> origin/gh/zklaus/22/orig 2025-12-04T08:57:04.1144211Z * [new branch] gh/zklaus/23/base -> origin/gh/zklaus/23/base 2025-12-04T08:57:04.1145989Z * [new branch] gh/zklaus/23/head -> origin/gh/zklaus/23/head 2025-12-04T08:57:04.1147630Z * [new branch] gh/zklaus/23/orig -> origin/gh/zklaus/23/orig 2025-12-04T08:57:04.1149666Z * [new branch] gh/zklaus/24/base -> origin/gh/zklaus/24/base 2025-12-04T08:57:04.1151341Z * [new branch] gh/zklaus/24/head -> origin/gh/zklaus/24/head 2025-12-04T08:57:04.1152717Z * [new branch] gh/zklaus/24/orig -> origin/gh/zklaus/24/orig 2025-12-04T08:57:04.1155284Z * [new branch] gh/zou3519/1197/base -> origin/gh/zou3519/1197/base 2025-12-04T08:57:04.1157002Z * [new branch] gh/zou3519/1197/head -> origin/gh/zou3519/1197/head 2025-12-04T08:57:04.1158515Z * [new branch] gh/zou3519/1197/orig -> origin/gh/zou3519/1197/orig 2025-12-04T08:57:04.1160685Z * [new branch] gh/zou3519/1199/base -> origin/gh/zou3519/1199/base 2025-12-04T08:57:04.1162359Z * [new branch] gh/zou3519/1199/head -> origin/gh/zou3519/1199/head 2025-12-04T08:57:04.1163925Z * [new branch] gh/zou3519/1199/orig -> origin/gh/zou3519/1199/orig 2025-12-04T08:57:04.1166105Z * [new branch] gh/zou3519/1200/base -> origin/gh/zou3519/1200/base 2025-12-04T08:57:04.1167587Z * [new branch] gh/zou3519/1200/head -> origin/gh/zou3519/1200/head 2025-12-04T08:57:04.1169211Z * [new branch] gh/zou3519/1200/orig -> origin/gh/zou3519/1200/orig 2025-12-04T08:57:04.1171302Z * [new branch] gh/zou3519/1201/base -> origin/gh/zou3519/1201/base 2025-12-04T08:57:04.1172877Z * [new branch] gh/zou3519/1201/head -> origin/gh/zou3519/1201/head 2025-12-04T08:57:04.1174479Z * [new branch] gh/zou3519/1201/orig -> origin/gh/zou3519/1201/orig 2025-12-04T08:57:04.1176404Z * [new branch] gh/zou3519/1202/base -> origin/gh/zou3519/1202/base 2025-12-04T08:57:04.1177950Z * [new branch] gh/zou3519/1202/head -> origin/gh/zou3519/1202/head 2025-12-04T08:57:04.1179633Z * [new branch] gh/zou3519/1202/orig -> origin/gh/zou3519/1202/orig 2025-12-04T08:57:04.1182114Z * [new branch] gh/zpcore/1/base -> origin/gh/zpcore/1/base 2025-12-04T08:57:04.1183624Z * [new branch] gh/zpcore/1/head -> origin/gh/zpcore/1/head 2025-12-04T08:57:04.1185778Z * [new branch] gh/zpcore/11/base -> origin/gh/zpcore/11/base 2025-12-04T08:57:04.1187406Z * [new branch] gh/zpcore/11/head -> origin/gh/zpcore/11/head 2025-12-04T08:57:04.1189009Z * [new branch] gh/zpcore/11/orig -> origin/gh/zpcore/11/orig 2025-12-04T08:57:04.1191499Z * [new branch] gh/zpcore/12/base -> origin/gh/zpcore/12/base 2025-12-04T08:57:04.1193052Z * [new branch] gh/zpcore/12/head -> origin/gh/zpcore/12/head 2025-12-04T08:57:04.1194651Z * [new branch] gh/zpcore/12/orig -> origin/gh/zpcore/12/orig 2025-12-04T08:57:04.1196820Z * [new branch] gh/zpcore/13/base -> origin/gh/zpcore/13/base 2025-12-04T08:57:04.1198317Z * [new branch] gh/zpcore/13/head -> origin/gh/zpcore/13/head 2025-12-04T08:57:04.1199853Z * [new branch] gh/zpcore/13/orig -> origin/gh/zpcore/13/orig 2025-12-04T08:57:04.1202415Z * [new branch] gh/zpcore/14/base -> origin/gh/zpcore/14/base 2025-12-04T08:57:04.1204053Z * [new branch] gh/zpcore/14/head -> origin/gh/zpcore/14/head 2025-12-04T08:57:04.1205926Z * [new branch] gh/zpcore/14/orig -> origin/gh/zpcore/14/orig 2025-12-04T08:57:04.1208236Z * [new branch] gh/zpcore/15/base -> origin/gh/zpcore/15/base 2025-12-04T08:57:04.1209764Z * [new branch] gh/zpcore/15/head -> origin/gh/zpcore/15/head 2025-12-04T08:57:04.1211397Z * [new branch] gh/zpcore/15/orig -> origin/gh/zpcore/15/orig 2025-12-04T08:57:04.1213584Z * [new branch] gh/zpcore/2/base -> origin/gh/zpcore/2/base 2025-12-04T08:57:04.1215167Z * [new branch] gh/zpcore/2/head -> origin/gh/zpcore/2/head 2025-12-04T08:57:04.1217682Z * [new branch] gh/zpcore/21/base -> origin/gh/zpcore/21/base 2025-12-04T08:57:04.1219351Z * [new branch] gh/zpcore/21/head -> origin/gh/zpcore/21/head 2025-12-04T08:57:04.1220925Z * [new branch] gh/zpcore/21/orig -> origin/gh/zpcore/21/orig 2025-12-04T08:57:04.1223371Z * [new branch] gh/zpcore/22/base -> origin/gh/zpcore/22/base 2025-12-04T08:57:04.1224814Z * [new branch] gh/zpcore/22/head -> origin/gh/zpcore/22/head 2025-12-04T08:57:04.1226765Z * [new branch] gh/zpcore/22/orig -> origin/gh/zpcore/22/orig 2025-12-04T08:57:04.1228938Z * [new branch] gh/zpcore/23/base -> origin/gh/zpcore/23/base 2025-12-04T08:57:04.1230502Z * [new branch] gh/zpcore/23/head -> origin/gh/zpcore/23/head 2025-12-04T08:57:04.1232108Z * [new branch] gh/zpcore/23/orig -> origin/gh/zpcore/23/orig 2025-12-04T08:57:04.1234275Z * [new branch] gh/zpcore/24/base -> origin/gh/zpcore/24/base 2025-12-04T08:57:04.1235955Z * [new branch] gh/zpcore/24/head -> origin/gh/zpcore/24/head 2025-12-04T08:57:04.1237520Z * [new branch] gh/zpcore/24/orig -> origin/gh/zpcore/24/orig 2025-12-04T08:57:04.1239814Z * [new branch] gh/zpcore/25/base -> origin/gh/zpcore/25/base 2025-12-04T08:57:04.1241821Z * [new branch] gh/zpcore/25/head -> origin/gh/zpcore/25/head 2025-12-04T08:57:04.1243396Z * [new branch] gh/zpcore/25/orig -> origin/gh/zpcore/25/orig 2025-12-04T08:57:04.1245525Z * [new branch] gh/zpcore/26/base -> origin/gh/zpcore/26/base 2025-12-04T08:57:04.1247534Z * [new branch] gh/zpcore/26/head -> origin/gh/zpcore/26/head 2025-12-04T08:57:04.1249087Z * [new branch] gh/zpcore/26/orig -> origin/gh/zpcore/26/orig 2025-12-04T08:57:04.1251260Z * [new branch] gh/zpcore/27/base -> origin/gh/zpcore/27/base 2025-12-04T08:57:04.1252917Z * [new branch] gh/zpcore/27/head -> origin/gh/zpcore/27/head 2025-12-04T08:57:04.1254566Z * [new branch] gh/zpcore/27/orig -> origin/gh/zpcore/27/orig 2025-12-04T08:57:04.1257191Z * [new branch] gh/zpcore/28/base -> origin/gh/zpcore/28/base 2025-12-04T08:57:04.1259047Z * [new branch] gh/zpcore/28/head -> origin/gh/zpcore/28/head 2025-12-04T08:57:04.1260631Z * [new branch] gh/zpcore/28/orig -> origin/gh/zpcore/28/orig 2025-12-04T08:57:04.1262622Z * [new branch] gh/zpcore/3/base -> origin/gh/zpcore/3/base 2025-12-04T08:57:04.1264130Z * [new branch] gh/zpcore/3/head -> origin/gh/zpcore/3/head 2025-12-04T08:57:04.1266200Z * [new branch] gh/zpcore/4/base -> origin/gh/zpcore/4/base 2025-12-04T08:57:04.1267705Z * [new branch] gh/zpcore/4/head -> origin/gh/zpcore/4/head 2025-12-04T08:57:04.1269612Z * [new branch] gh/zpcore/5/base -> origin/gh/zpcore/5/base 2025-12-04T08:57:04.1271146Z * [new branch] gh/zpcore/5/head -> origin/gh/zpcore/5/head 2025-12-04T08:57:04.1273120Z * [new branch] gh/zpcore/6/base -> origin/gh/zpcore/6/base 2025-12-04T08:57:04.1274666Z * [new branch] gh/zpcore/6/head -> origin/gh/zpcore/6/head 2025-12-04T08:57:04.1277147Z * [new branch] gh/zpcore/7/base -> origin/gh/zpcore/7/base 2025-12-04T08:57:04.1278655Z * [new branch] gh/zpcore/7/head -> origin/gh/zpcore/7/head 2025-12-04T08:57:04.1280711Z * [new branch] gh/zpcore/8/base -> origin/gh/zpcore/8/base 2025-12-04T08:57:04.1282256Z * [new branch] gh/zpcore/8/head -> origin/gh/zpcore/8/head 2025-12-04T08:57:04.1283975Z * [new branch] google-main -> origin/google-main 2025-12-04T08:57:04.1286041Z * [new branch] guangyey/external_stream -> origin/guangyey/external_stream 2025-12-04T08:57:04.1287645Z * [new branch] guangyey/test_2025 -> origin/guangyey/test_2025 2025-12-04T08:57:04.1289986Z * [new branch] guilhermeleobas/cherry-pick-55d87d9dfd9 -> origin/guilhermeleobas/cherry-pick-55d87d9dfd9 2025-12-04T08:57:04.1291954Z * [new branch] hameerabbasi/complex_tensor_subclass -> origin/hameerabbasi/complex_tensor_subclass 2025-12-04T08:57:04.1293648Z * [new branch] hameerabbasi/fix-ctensor-gradcheck-tests -> origin/hameerabbasi/fix-ctensor-gradcheck-tests 2025-12-04T08:57:04.1295145Z * [new branch] hameerabbasi/gradcheck-allclose -> origin/hameerabbasi/gradcheck-allclose 2025-12-04T08:57:04.1296720Z * [new branch] hc_baseline -> origin/hc_baseline 2025-12-04T08:57:04.1298347Z * [new branch] hhh_rand -> origin/hhh_rand 2025-12-04T08:57:04.1300413Z * [new branch] huba/f1 -> origin/huba/f1 2025-12-04T08:57:04.1302689Z * [new branch] increase-timeout-linux-jammy-cuda12_8-py3_10-gcc11-test -> origin/increase-timeout-linux-jammy-cuda12_8-py3_10-gcc11-test 2025-12-04T08:57:04.1304014Z * [new branch] inlining -> origin/inlining 2025-12-04T08:57:04.1306219Z * [new branch] inlining-ezyang -> origin/inlining-ezyang 2025-12-04T08:57:04.1307978Z * [new branch] install-torchao-0.13.0 -> origin/install-torchao-0.13.0 2025-12-04T08:57:04.1309976Z * [new branch] instrument-trunk-pull-linux-with-job-test-filters -> origin/instrument-trunk-pull-linux-with-job-test-filters 2025-12-04T08:57:04.1311219Z * [new branch] invoke-subgraph -> origin/invoke-subgraph 2025-12-04T08:57:04.1313000Z * [new branch] issue#58739 -> origin/issue#58739 2025-12-04T08:57:04.1314588Z * [new branch] jainapurva-patch-1 -> origin/jainapurva-patch-1 2025-12-04T08:57:04.1317051Z * [new branch] jathu/o3 -> origin/jathu/o3 2025-12-04T08:57:04.1318538Z * [new branch] jathu/sve -> origin/jathu/sve 2025-12-04T08:57:04.1320953Z * [new branch] jcaip/test-cusparselt-version-0.6.2 -> origin/jcaip/test-cusparselt-version-0.6.2 2025-12-04T08:57:04.1322854Z * [new branch] jcaip/update-cusparselt-0.6.2 -> origin/jcaip/update-cusparselt-0.6.2 2025-12-04T08:57:04.1324995Z * [new branch] jiannanWang/memorysnapshot_filter -> origin/jiannanWang/memorysnapshot_filter 2025-12-04T08:57:04.1326543Z * [new branch] jiannanWang/profilerstepwarning -> origin/jiannanWang/profilerstepwarning 2025-12-04T08:57:04.1328099Z * [new branch] jithunnair-amd-patch-1 -> origin/jithunnair-amd-patch-1 2025-12-04T08:57:04.1329807Z * [new branch] jithunnair-amd-patch-10 -> origin/jithunnair-amd-patch-10 2025-12-04T08:57:04.1331516Z * [new branch] jithunnair-amd-patch-2 -> origin/jithunnair-amd-patch-2 2025-12-04T08:57:04.1333125Z * [new branch] jithunnair-amd-patch-3 -> origin/jithunnair-amd-patch-3 2025-12-04T08:57:04.1335017Z * [new branch] jithunnair-amd-patch-4 -> origin/jithunnair-amd-patch-4 2025-12-04T08:57:04.1336579Z * [new branch] jithunnair-amd-patch-5 -> origin/jithunnair-amd-patch-5 2025-12-04T08:57:04.1338353Z * [new branch] jithunnair-amd-patch-6 -> origin/jithunnair-amd-patch-6 2025-12-04T08:57:04.1339928Z * [new branch] jithunnair-amd-patch-7 -> origin/jithunnair-amd-patch-7 2025-12-04T08:57:04.1341616Z * [new branch] jithunnair-amd-patch-8 -> origin/jithunnair-amd-patch-8 2025-12-04T08:57:04.1343303Z * [new branch] jithunnair-amd-patch-9 -> origin/jithunnair-amd-patch-9 2025-12-04T08:57:04.1345634Z * [new branch] justinchu/native-qdq -> origin/justinchu/native-qdq 2025-12-04T08:57:04.1347672Z * [new branch] kainan666/xlf_debug -> origin/kainan666/xlf_debug 2025-12-04T08:57:04.1349398Z * [new branch] kainan_test -> origin/kainan_test 2025-12-04T08:57:04.1350905Z * [new branch] larryliu0820-patch-1 -> origin/larryliu0820-patch-1 2025-12-04T08:57:04.1353101Z * [new branch] leslie/test_group_gemm_epilogues -> origin/leslie/test_group_gemm_epilogues 2025-12-04T08:57:04.1355111Z * [new branch] lessw2020/fix_cutlass_cache_error -> origin/lessw2020/fix_cutlass_cache_error 2025-12-04T08:57:04.1357136Z * [new branch] liaoxuan/shm_all_reduce -> origin/liaoxuan/shm_all_reduce 2025-12-04T08:57:04.1358679Z * [new branch] liaoxuan/test_fa_disable_softmax -> origin/liaoxuan/test_fa_disable_softmax 2025-12-04T08:57:04.1360092Z * [new branch] liaoxuan/test_int8_sdpa -> origin/liaoxuan/test_int8_sdpa 2025-12-04T08:57:04.1361649Z * [new branch] llama4-stable -> origin/llama4-stable 2025-12-04T08:57:04.1364384Z * [new branch] lts/release/1.8 -> origin/lts/release/1.8 2025-12-04T08:57:04.1366551Z * [new branch] lucaskabela/#94773 -> origin/lucaskabela/#94773 2025-12-04T08:57:04.1368334Z * [new branch] lucaskabela/fix_164876 -> origin/lucaskabela/fix_164876 2025-12-04T08:57:04.1369563Z * [new branch] lucaskabela/flop_counter -> origin/lucaskabela/flop_counter 2025-12-04T08:57:04.1371355Z * [new branch] lucaskabela/func_under_decomp -> origin/lucaskabela/func_under_decomp 2025-12-04T08:57:04.1372468Z * [new branch] lucaskabela/functional_in_dynamo -> origin/lucaskabela/functional_in_dynamo 2025-12-04T08:57:04.1374219Z * [new branch] lucaskabela/install_params_as_graph_attr -> origin/lucaskabela/install_params_as_graph_attr 2025-12-04T08:57:04.1376022Z * [new branch] lucaskabela/parameters_as_graph_attr -> origin/lucaskabela/parameters_as_graph_attr 2025-12-04T08:57:04.1377973Z * [new branch] lucaskabela/remove_aot_dispatcher_metadata -> origin/lucaskabela/remove_aot_dispatcher_metadata 2025-12-04T08:57:04.1379448Z * [new branch] lucaskabela/rnn_decomp -> origin/lucaskabela/rnn_decomp 2025-12-04T08:57:04.1381028Z * [new branch] lucaskabela/typing_backends -> origin/lucaskabela/typing_backends 2025-12-04T08:57:04.1382672Z * [new branch] lucaskabela/typing_ctx_manager -> origin/lucaskabela/typing_ctx_manager 2025-12-04T08:57:04.1384222Z * [new branch] lucaskabela/typing_nn_module -> origin/lucaskabela/typing_nn_module 2025-12-04T08:57:04.1385849Z * [new branch] lucaskabela/typing_user_defined -> origin/lucaskabela/typing_user_defined 2025-12-04T08:57:04.1387364Z * [new branch] lucaskabela/typing_variables -> origin/lucaskabela/typing_variables 2025-12-04T08:57:04.1389101Z * [new branch] lucaskabela/typing_variables_dicts -> origin/lucaskabela/typing_variables_dicts 2025-12-04T08:57:04.1390713Z * [new branch] lucaskabela/typing_variables_functions -> origin/lucaskabela/typing_variables_functions 2025-12-04T08:57:04.1392256Z * [new branch] lucaskabela/typing_variables_lists -> origin/lucaskabela/typing_variables_lists 2025-12-04T08:57:04.1394320Z * [new branch] lw/torch_box_by_ref -> origin/lw/torch_box_by_ref 2025-12-04T08:57:04.1395973Z * [new branch] main -> origin/main 2025-12-04T08:57:04.1397767Z * [new branch] malfet-patch-1 -> origin/malfet-patch-1 2025-12-04T08:57:04.1399555Z * [new branch] malfet-patch-2 -> origin/malfet-patch-2 2025-12-04T08:57:04.1401204Z * [new branch] malfet-patch-3 -> origin/malfet-patch-3 2025-12-04T08:57:04.1402907Z * [new branch] malfet-patch-4 -> origin/malfet-patch-4 2025-12-04T08:57:04.1404794Z * [new branch] malfet-patch-5 -> origin/malfet-patch-5 2025-12-04T08:57:04.1408840Z * [new branch] malfet-patch-6 -> origin/malfet-patch-6 2025-12-04T08:57:04.1410388Z * [new branch] malfet-patch-7 -> origin/malfet-patch-7 2025-12-04T08:57:04.1412041Z * [new branch] malfet-patch-8 -> origin/malfet-patch-8 2025-12-04T08:57:04.1414283Z * [new branch] malfet/add-3.14-ci -> origin/malfet/add-3.14-ci 2025-12-04T08:57:04.1416091Z * [new branch] malfet/be-do-not-make-typos-in-build-artifacts -> origin/malfet/be-do-not-make-typos-in-build-artifacts 2025-12-04T08:57:04.1417569Z * [new branch] malfet/be-move-more-settings-to-checkout-pytorch -> origin/malfet/be-move-more-settings-to-checkout-pytorch 2025-12-04T08:57:04.1419319Z * [new branch] malfet/be-remove-misisng-neon-headers -> origin/malfet/be-remove-misisng-neon-headers 2025-12-04T08:57:04.1421033Z * [new branch] malfet/mps-implement-col2im -> origin/malfet/mps-implement-col2im 2025-12-04T08:57:04.1423182Z * [new branch] manuel/aoti_metal_shimify-thread_safe -> origin/manuel/aoti_metal_shimify-thread_safe 2025-12-04T08:57:04.1425052Z * [new branch] manuel/inductor_link_openmp -> origin/manuel/inductor_link_openmp 2025-12-04T08:57:04.1427056Z * [new branch] masnesral/metaconda -> origin/masnesral/metaconda 2025-12-04T08:57:04.1428691Z * [new branch] mem_profiler_flaky_fix -> origin/mem_profiler_flaky_fix 2025-12-04T08:57:04.1430351Z * [new branch] mem_profiler_stack_trace -> origin/mem_profiler_stack_trace 2025-12-04T08:57:04.1432003Z * [new branch] memory_profiler_stack -> origin/memory_profiler_stack 2025-12-04T08:57:04.1433735Z * [new branch] metascroy-patch-1 -> origin/metascroy-patch-1 2025-12-04T08:57:04.1435389Z * [new branch] mingw_posix -> origin/mingw_posix 2025-12-04T08:57:04.1437574Z * [new branch] mlazos/S429861-debug -> origin/mlazos/S429861-debug 2025-12-04T08:57:04.1439125Z * [new branch] mlazos/aa -> origin/mlazos/aa 2025-12-04T08:57:04.1440634Z * [new branch] mlazos/acts -> origin/mlazos/acts 2025-12-04T08:57:04.1442205Z * [new branch] mlazos/arg-renames -> origin/mlazos/arg-renames 2025-12-04T08:57:04.1443643Z * [new branch] mlazos/bad-cudagraphs -> origin/mlazos/bad-cudagraphs 2025-12-04T08:57:04.1445076Z * [new branch] mlazos/baseline-graph-breaks -> origin/mlazos/baseline-graph-breaks 2025-12-04T08:57:04.1446533Z * [new branch] mlazos/beta-tensor -> origin/mlazos/beta-tensor 2025-12-04T08:57:04.1448016Z * [new branch] mlazos/buffers -> origin/mlazos/buffers 2025-12-04T08:57:04.1449411Z * [new branch] mlazos/buffers2 -> origin/mlazos/buffers2 2025-12-04T08:57:04.1451194Z * [new branch] mlazos/buffers3 -> origin/mlazos/buffers3 2025-12-04T08:57:04.1453039Z * [new branch] mlazos/bwd -> origin/mlazos/bwd 2025-12-04T08:57:04.1454690Z * [new branch] mlazos/combo-test -> origin/mlazos/combo-test 2025-12-04T08:57:04.1456267Z * [new branch] mlazos/ctx-cleanup -> origin/mlazos/ctx-cleanup 2025-12-04T08:57:04.1457842Z * [new branch] mlazos/cuda-cmd-log -> origin/mlazos/cuda-cmd-log 2025-12-04T08:57:04.1459578Z * [new branch] mlazos/cudagraph-tests -> origin/mlazos/cudagraph-tests 2025-12-04T08:57:04.1461222Z * [new branch] mlazos/cudagraphs-measurement -> origin/mlazos/cudagraphs-measurement 2025-12-04T08:57:04.1462780Z * [new branch] mlazos/cutlass-test -> origin/mlazos/cutlass-test 2025-12-04T08:57:04.1464456Z * [new branch] mlazos/cutlass-topo-bug -> origin/mlazos/cutlass-topo-bug 2025-12-04T08:57:04.1466172Z * [new branch] mlazos/dataclass-proxy -> origin/mlazos/dataclass-proxy 2025-12-04T08:57:04.1467657Z * [new branch] mlazos/dc-attrs -> origin/mlazos/dc-attrs 2025-12-04T08:57:04.1469176Z * [new branch] mlazos/dc-helion -> origin/mlazos/dc-helion 2025-12-04T08:57:04.1470698Z * [new branch] mlazos/dict-fix -> origin/mlazos/dict-fix 2025-12-04T08:57:04.1472358Z * [new branch] mlazos/disable-tf -> origin/mlazos/disable-tf 2025-12-04T08:57:04.1473948Z * [new branch] mlazos/dupe-fix -> origin/mlazos/dupe-fix 2025-12-04T08:57:04.1476022Z * [new branch] mlazos/dyn-batch -> origin/mlazos/dyn-batch 2025-12-04T08:57:04.1477595Z * [new branch] mlazos/evt -> origin/mlazos/evt 2025-12-04T08:57:04.1479190Z * [new branch] mlazos/extract-examples -> origin/mlazos/extract-examples 2025-12-04T08:57:04.1481187Z * [new branch] mlazos/foreach-op -> origin/mlazos/foreach-op 2025-12-04T08:57:04.1482720Z * [new branch] mlazos/fp8 -> origin/mlazos/fp8 2025-12-04T08:57:04.1484301Z * [new branch] mlazos/fp8-bias -> origin/mlazos/fp8-bias 2025-12-04T08:57:04.1485897Z * [new branch] mlazos/fp8-bias-fusion -> origin/mlazos/fp8-bias-fusion 2025-12-04T08:57:04.1487512Z * [new branch] mlazos/fp8-fixes -> origin/mlazos/fp8-fixes 2025-12-04T08:57:04.1489215Z * [new branch] mlazos/freezing -> origin/mlazos/freezing 2025-12-04T08:57:04.1490916Z * [new branch] mlazos/h-comp -> origin/mlazos/h-comp 2025-12-04T08:57:04.1492549Z * [new branch] mlazos/h-comp2 -> origin/mlazos/h-comp2 2025-12-04T08:57:04.1494387Z * [new branch] mlazos/hash-hop -> origin/mlazos/hash-hop 2025-12-04T08:57:04.1495954Z * [new branch] mlazos/hc -> origin/mlazos/hc 2025-12-04T08:57:04.1497588Z * [new branch] mlazos/hc-cycles -> origin/mlazos/hc-cycles 2025-12-04T08:57:04.1499159Z * [new branch] mlazos/hc-fixes -> origin/mlazos/hc-fixes 2025-12-04T08:57:04.1500720Z * [new branch] mlazos/hc-fixes3 -> origin/mlazos/hc-fixes3 2025-12-04T08:57:04.1502285Z * [new branch] mlazos/hc-fixes4 -> origin/mlazos/hc-fixes4 2025-12-04T08:57:04.1503853Z * [new branch] mlazos/hc-hf -> origin/mlazos/hc-hf 2025-12-04T08:57:04.1505746Z * [new branch] mlazos/hc-mut -> origin/mlazos/hc-mut 2025-12-04T08:57:04.1507406Z * [new branch] mlazos/hc10 -> origin/mlazos/hc10 2025-12-04T08:57:04.1508937Z * [new branch] mlazos/hc11 -> origin/mlazos/hc11 2025-12-04T08:57:04.1510505Z * [new branch] mlazos/hc12 -> origin/mlazos/hc12 2025-12-04T08:57:04.1512095Z * [new branch] mlazos/hc13 -> origin/mlazos/hc13 2025-12-04T08:57:04.1513641Z * [new branch] mlazos/hc14 -> origin/mlazos/hc14 2025-12-04T08:57:04.1515202Z * [new branch] mlazos/hc15 -> origin/mlazos/hc15 2025-12-04T08:57:04.1516919Z * [new branch] mlazos/hc2 -> origin/mlazos/hc2 2025-12-04T08:57:04.1518498Z * [new branch] mlazos/hc4 -> origin/mlazos/hc4 2025-12-04T08:57:04.1520090Z * [new branch] mlazos/hc5 -> origin/mlazos/hc5 2025-12-04T08:57:04.1521610Z * [new branch] mlazos/hc6 -> origin/mlazos/hc6 2025-12-04T08:57:04.1523179Z * [new branch] mlazos/hc7 -> origin/mlazos/hc7 2025-12-04T08:57:04.1524633Z * [new branch] mlazos/hc8 -> origin/mlazos/hc8 2025-12-04T08:57:04.1526358Z * [new branch] mlazos/hc9 -> origin/mlazos/hc9 2025-12-04T08:57:04.1527881Z * [new branch] mlazos/hc_baseline2 -> origin/mlazos/hc_baseline2 2025-12-04T08:57:04.1529341Z * [new branch] mlazos/inductor-streams -> origin/mlazos/inductor-streams 2025-12-04T08:57:04.1530732Z * [new branch] mlazos/main -> origin/mlazos/main 2025-12-04T08:57:04.1532309Z * [new branch] mlazos/mcg2 -> origin/mlazos/mcg2 2025-12-04T08:57:04.1534101Z * [new branch] mlazos/meta-guards -> origin/mlazos/meta-guards 2025-12-04T08:57:04.1536247Z * [new branch] mlazos/mlazos/foreach-map-adam -> origin/mlazos/mlazos/foreach-map-adam 2025-12-04T08:57:04.1537801Z * [new branch] mlazos/mlazos/tf-mode-backup -> origin/mlazos/mlazos/tf-mode-backup 2025-12-04T08:57:04.1539389Z * [new branch] mlazos/mod-fix -> origin/mlazos/mod-fix 2025-12-04T08:57:04.1541165Z * [new branch] mlazos/mode-fix -> origin/mlazos/mode-fix 2025-12-04T08:57:04.1542720Z * [new branch] mlazos/offsets -> origin/mlazos/offsets 2025-12-04T08:57:04.1544210Z * [new branch] mlazos/overguarding -> origin/mlazos/overguarding 2025-12-04T08:57:04.1545857Z * [new branch] mlazos/proxy-ctors -> origin/mlazos/proxy-ctors 2025-12-04T08:57:04.1547390Z * [new branch] mlazos/quant-fix -> origin/mlazos/quant-fix 2025-12-04T08:57:04.1548977Z * [new branch] mlazos/resnet-fix -> origin/mlazos/resnet-fix 2025-12-04T08:57:04.1550552Z * [new branch] mlazos/rm-buf-names -> origin/mlazos/rm-buf-names 2025-12-04T08:57:04.1552142Z * [new branch] mlazos/rm-code -> origin/mlazos/rm-code 2025-12-04T08:57:04.1553717Z * [new branch] mlazos/rm-spam -> origin/mlazos/rm-spam 2025-12-04T08:57:04.1555317Z * [new branch] mlazos/rtp -> origin/mlazos/rtp 2025-12-04T08:57:04.1556993Z * [new branch] mlazos/static-idx-dbg -> origin/mlazos/static-idx-dbg 2025-12-04T08:57:04.1558566Z * [new branch] mlazos/static-inputs-log -> origin/mlazos/static-inputs-log 2025-12-04T08:57:04.1559992Z * [new branch] mlazos/stests -> origin/mlazos/stests 2025-12-04T08:57:04.1561633Z * [new branch] mlazos/stream-ops -> origin/mlazos/stream-ops 2025-12-04T08:57:04.1563181Z * [new branch] mlazos/td-fix2 -> origin/mlazos/td-fix2 2025-12-04T08:57:04.1564756Z * [new branch] mlazos/tensor-hasattr2 -> origin/mlazos/tensor-hasattr2 2025-12-04T08:57:04.1566348Z * [new branch] mlazos/test -> origin/mlazos/test 2025-12-04T08:57:04.1567911Z * [new branch] mlazos/tf-mode -> origin/mlazos/tf-mode 2025-12-04T08:57:04.1569515Z * [new branch] mlazos/tf-mode-backup2 -> origin/mlazos/tf-mode-backup2 2025-12-04T08:57:04.1571085Z * [new branch] mlazos/tf-mode-reland -> origin/mlazos/tf-mode-reland 2025-12-04T08:57:04.1572739Z * [new branch] mlazos/tf-mode-reland2 -> origin/mlazos/tf-mode-reland2 2025-12-04T08:57:04.1574467Z * [new branch] mlazos/tf-mode-reland3 -> origin/mlazos/tf-mode-reland3 2025-12-04T08:57:04.1576093Z * [new branch] mlazos/triton-no-epi -> origin/mlazos/triton-no-epi 2025-12-04T08:57:04.1577709Z * [new branch] mlazos/tune-proto -> origin/mlazos/tune-proto 2025-12-04T08:57:04.1579273Z * [new branch] mlazos/tuple-fixes -> origin/mlazos/tuple-fixes 2025-12-04T08:57:04.1580833Z * [new branch] mlazos/tuple-fixes2 -> origin/mlazos/tuple-fixes2 2025-12-04T08:57:04.1582983Z * [new branch] mlazos/tuple-handling -> origin/mlazos/tuple-handling 2025-12-04T08:57:04.1584501Z * [new branch] mlazos/user-stream-base -> origin/mlazos/user-stream-base 2025-12-04T08:57:04.1586158Z * [new branch] mlazos/user-streams -> origin/mlazos/user-streams 2025-12-04T08:57:04.1588204Z * [new branch] mlazos/user-streams-backup -> origin/mlazos/user-streams-backup 2025-12-04T08:57:04.1589801Z * [new branch] mlazos/user-streams-backup2 -> origin/mlazos/user-streams-backup2 2025-12-04T08:57:04.1591325Z * [new branch] mlazos/vary-beta -> origin/mlazos/vary-beta 2025-12-04T08:57:04.1593016Z * [new branch] mlazos/vary-beta2 -> origin/mlazos/vary-beta2 2025-12-04T08:57:04.1594616Z * [new branch] mlazos/weird-perf1 -> origin/mlazos/weird-perf1 2025-12-04T08:57:04.1596333Z * [new branch] mm_out_dtype_compile -> origin/mm_out_dtype_compile 2025-12-04T08:57:04.1597931Z * [new branch] module-shim -> origin/module-shim 2025-12-04T08:57:04.1599577Z * [new branch] move_config -> origin/move_config 2025-12-04T08:57:04.1601665Z * [new branch] msaroufim/reduce -> origin/msaroufim/reduce 2025-12-04T08:57:04.1603769Z * [new branch] mtia/basic-cmake -> origin/mtia/basic-cmake 2025-12-04T08:57:04.1606375Z * [new branch] mwizak/fix-triton-block-shape -> origin/mwizak/fix-triton-block-shape 2025-12-04T08:57:04.1608531Z * [new branch] my_varlen_backup -> origin/my_varlen_backup 2025-12-04T08:57:04.1610194Z * [new branch] nativert_num_outputs -> origin/nativert_num_outputs 2025-12-04T08:57:04.1611798Z * [new branch] new-codegen -> origin/new-codegen 2025-12-04T08:57:04.1613513Z * [new branch] newtest-base -> origin/newtest-base 2025-12-04T08:57:04.1615783Z * [new branch] ngimel/addmm_dtype -> origin/ngimel/addmm_dtype 2025-12-04T08:57:04.1617238Z * [new branch] ngimel/div_inv -> origin/ngimel/div_inv 2025-12-04T08:57:04.1618704Z * [new branch] ngimel/error_index_list -> origin/ngimel/error_index_list 2025-12-04T08:57:04.1620154Z * [new branch] ngimel/gather_grid -> origin/ngimel/gather_grid 2025-12-04T08:57:04.1621714Z * [new branch] ngimel/gather_grid_release -> origin/ngimel/gather_grid_release 2025-12-04T08:57:04.1623179Z * [new branch] ngimel/gg_new -> origin/ngimel/gg_new 2025-12-04T08:57:04.1624661Z * [new branch] ngimel/hostalloc -> origin/ngimel/hostalloc 2025-12-04T08:57:04.1626106Z * [new branch] ngimel/storage_id -> origin/ngimel/storage_id 2025-12-04T08:57:04.1627721Z * [new branch] nightly -> origin/nightly 2025-12-04T08:57:04.1629921Z * [new branch] nikitaved/addmm_1_rowcol_lt_path_check -> origin/nikitaved/addmm_1_rowcol_lt_path_check 2025-12-04T08:57:04.1631433Z * [new branch] nikitaved/addmm_epilogue_fusions_2d_bias -> origin/nikitaved/addmm_epilogue_fusions_2d_bias 2025-12-04T08:57:04.1632894Z * [new branch] nikitaved/addmm_epilogue_fusions_inductor -> origin/nikitaved/addmm_epilogue_fusions_inductor 2025-12-04T08:57:04.1634722Z * [new branch] nikitaved/addmm_epilogue_fusions_scratch -> origin/nikitaved/addmm_epilogue_fusions_scratch 2025-12-04T08:57:04.1636538Z * [new branch] nikitaved/grad_addmm_epilogue_fusions -> origin/nikitaved/grad_addmm_epilogue_fusions 2025-12-04T08:57:04.1638505Z * [new branch] nikitaved/simpler_can_use_32bit_index -> origin/nikitaved/simpler_can_use_32bit_index 2025-12-04T08:57:04.1640047Z * [new branch] nikitaved/test -> origin/nikitaved/test 2025-12-04T08:57:04.1642201Z * [new branch] nmacchioni-perf-test-async-autotune -> origin/nmacchioni-perf-test-async-autotune 2025-12-04T08:57:04.1643705Z * [new branch] no_distributed_log_spew -> origin/no_distributed_log_spew 2025-12-04T08:57:04.1645735Z * [new branch] nofun-hack -> origin/nofun-hack 2025-12-04T08:57:04.1647560Z * [new branch] norm_bench -> origin/norm_bench 2025-12-04T08:57:04.1649722Z * [new branch] nullplay/fuse_matmul -> origin/nullplay/fuse_matmul 2025-12-04T08:57:04.1651366Z * [new branch] nullplay_fuse_matmul -> origin/nullplay_fuse_matmul 2025-12-04T08:57:04.1652963Z * [new branch] optimizer_test -> origin/optimizer_test 2025-12-04T08:57:04.1655692Z * [new branch] orig/release/1.10 -> origin/orig/release/1.10 2025-12-04T08:57:04.1657320Z * [new branch] orig/release/1.11 -> origin/orig/release/1.11 2025-12-04T08:57:04.1658992Z * [new branch] orig/release/1.12 -> origin/orig/release/1.12 2025-12-04T08:57:04.1660895Z * [new branch] orig/release/1.13 -> origin/orig/release/1.13 2025-12-04T08:57:04.1662539Z * [new branch] orig/release/1.6 -> origin/orig/release/1.6 2025-12-04T08:57:04.1664331Z * [new branch] orig/release/1.7 -> origin/orig/release/1.7 2025-12-04T08:57:04.1665979Z * [new branch] orig/release/1.8 -> origin/orig/release/1.8 2025-12-04T08:57:04.1667562Z * [new branch] orig/release/1.9 -> origin/orig/release/1.9 2025-12-04T08:57:04.1669159Z * [new branch] orig/release/2.0 -> origin/orig/release/2.0 2025-12-04T08:57:04.1671206Z * [new branch] orig/release/2.1 -> origin/orig/release/2.1 2025-12-04T08:57:04.1672939Z * [new branch] orig/release/2.2 -> origin/orig/release/2.2 2025-12-04T08:57:04.1674465Z * [new branch] orig/release/2.3 -> origin/orig/release/2.3 2025-12-04T08:57:04.1675982Z * [new branch] orig/release/2.4 -> origin/orig/release/2.4 2025-12-04T08:57:04.1677530Z * [new branch] orig/release/2.5 -> origin/orig/release/2.5 2025-12-04T08:57:04.1679063Z * [new branch] orig/release/2.6 -> origin/orig/release/2.6 2025-12-04T08:57:04.1681060Z * [new branch] orig/release/2.7 -> origin/orig/release/2.7 2025-12-04T08:57:04.1683297Z * [new branch] orig/release/2.8 -> origin/orig/release/2.8 2025-12-04T08:57:04.1684857Z * [new branch] orig/release/2.9 -> origin/orig/release/2.9 2025-12-04T08:57:04.1688366Z * [new branch] origin/gh/fxdawnn/1/base -> origin/origin/gh/fxdawnn/1/base 2025-12-04T08:57:04.1689909Z * [new branch] origin/gh/fxdawnn/1/orig -> origin/origin/gh/fxdawnn/1/orig 2025-12-04T08:57:04.1692427Z * [new branch] origin/gh/zpcore/14/orig -> origin/origin/gh/zpcore/14/orig 2025-12-04T08:57:04.1694336Z * [new branch] oulgen-patch-1 -> origin/oulgen-patch-1 2025-12-04T08:57:04.1696252Z * [new branch] oulgen-patch-2 -> origin/oulgen-patch-2 2025-12-04T08:57:04.1697973Z * [new branch] oulgen-patch-3 -> origin/oulgen-patch-3 2025-12-04T08:57:04.1700020Z * [new branch] oulgen-patch-4 -> origin/oulgen-patch-4 2025-12-04T08:57:04.1701788Z * [new branch] padded-tensor -> origin/padded-tensor 2025-12-04T08:57:04.1703407Z * [new branch] pca2 -> origin/pca2 2025-12-04T08:57:04.1705017Z * [new branch] per_channel_backup -> origin/per_channel_backup 2025-12-04T08:57:04.1706875Z * [new branch] perf_ops -> origin/perf_ops 2025-12-04T08:57:04.1708576Z * [new branch] perf_ops_2_9 -> origin/perf_ops_2_9 2025-12-04T08:57:04.1710463Z * [new branch] pianpwk-patch-1 -> origin/pianpwk-patch-1 2025-12-04T08:57:04.1712481Z * [new branch] pianpwk/__draft_debug_mode -> origin/pianpwk/__draft_debug_mode 2025-12-04T08:57:04.1714031Z * [new branch] pianpwk/_debug_mode_for_triton_draft -> origin/pianpwk/_debug_mode_for_triton_draft 2025-12-04T08:57:04.1715492Z * [new branch] pianpwk/_debug_nn_module_compile -> origin/pianpwk/_debug_nn_module_compile 2025-12-04T08:57:04.1716909Z * [new branch] pianpwk/_draft_triton_11_3 -> origin/pianpwk/_draft_triton_11_3 2025-12-04T08:57:04.1718358Z * [new branch] pianpwk/_manual_bucket_draft -> origin/pianpwk/_manual_bucket_draft 2025-12-04T08:57:04.1720078Z * [new branch] pianpwk/_profile_w_dispatch_keys -> origin/pianpwk/_profile_w_dispatch_keys 2025-12-04T08:57:04.1722129Z * [new branch] pianpwk/_super_draft_debug_mode -> origin/pianpwk/_super_draft_debug_mode 2025-12-04T08:57:04.1723767Z * [new branch] pianpwk/_unbacked_local_shard_size -> origin/pianpwk/_unbacked_local_shard_size 2025-12-04T08:57:04.1725287Z * [new branch] pianpwk/anomaly_tb -> origin/pianpwk/anomaly_tb 2025-12-04T08:57:04.1726835Z * [new branch] pianpwk/auto_fx_annotate -> origin/pianpwk/auto_fx_annotate 2025-12-04T08:57:04.1728537Z * [new branch] pianpwk/backed_size_oblivious_export -> origin/pianpwk/backed_size_oblivious_export 2025-12-04T08:57:04.1729998Z * [new branch] pianpwk/bert_dynamic_perf -> origin/pianpwk/bert_dynamic_perf 2025-12-04T08:57:04.1731645Z * [new branch] pianpwk/debug_fwd_stack_traces -> origin/pianpwk/debug_fwd_stack_traces 2025-12-04T08:57:04.1733335Z * [new branch] pianpwk/debug_hash_tensor -> origin/pianpwk/debug_hash_tensor 2025-12-04T08:57:04.1735015Z * [new branch] pianpwk/debug_mode_annotate -> origin/pianpwk/debug_mode_annotate 2025-12-04T08:57:04.1736601Z * [new branch] pianpwk/debug_mode_defaults -> origin/pianpwk/debug_mode_defaults 2025-12-04T08:57:04.1738127Z * [new branch] pianpwk/debug_mode_hacks -> origin/pianpwk/debug_mode_hacks 2025-12-04T08:57:04.1739831Z * [new branch] pianpwk/debug_mode_opcall_refactor -> origin/pianpwk/debug_mode_opcall_refactor 2025-12-04T08:57:04.1741365Z * [new branch] pianpwk/debug_mode_show_ids -> origin/pianpwk/debug_mode_show_ids 2025-12-04T08:57:04.1742904Z * [new branch] pianpwk/debug_mode_triton -> origin/pianpwk/debug_mode_triton 2025-12-04T08:57:04.1744695Z * [new branch] pianpwk/debug_show_stack_trace -> origin/pianpwk/debug_show_stack_trace 2025-12-04T08:57:04.1746305Z * [new branch] pianpwk/debug_wait_on_collective -> origin/pianpwk/debug_wait_on_collective 2025-12-04T08:57:04.1747854Z * [new branch] pianpwk/debugmode_compile_tf -> origin/pianpwk/debugmode_compile_tf 2025-12-04T08:57:04.1749581Z * [new branch] pianpwk/dispatch_key_debugging_for_debug -> origin/pianpwk/dispatch_key_debugging_for_debug 2025-12-04T08:57:04.1751128Z * [new branch] pianpwk/draft_debug_mode_tfcompile -> origin/pianpwk/draft_debug_mode_tfcompile 2025-12-04T08:57:04.1753131Z * [new branch] pianpwk/draft_multikernel_nn -> origin/pianpwk/draft_multikernel_nn 2025-12-04T08:57:04.1754780Z * [new branch] pianpwk/draft_multikernel_status_10_5 -> origin/pianpwk/draft_multikernel_status_10_5 2025-12-04T08:57:04.1756389Z * [new branch] pianpwk/dtensor_custom_chunk -> origin/pianpwk/dtensor_custom_chunk 2025-12-04T08:57:04.1758554Z * [new branch] pianpwk/dtensor_unbacked_keypath -> origin/pianpwk/dtensor_unbacked_keypath 2025-12-04T08:57:04.1760215Z * [new branch] pianpwk/event_list_tree -> origin/pianpwk/event_list_tree 2025-12-04T08:57:04.1761944Z * [new branch] pianpwk/false_numel_refs -> origin/pianpwk/false_numel_refs 2025-12-04T08:57:04.1763578Z * [new branch] pianpwk/maybe_guard_rel -> origin/pianpwk/maybe_guard_rel 2025-12-04T08:57:04.1765156Z * [new branch] pianpwk/multikernel_hints_draft -> origin/pianpwk/multikernel_hints_draft 2025-12-04T08:57:04.1766724Z * [new branch] pianpwk/no_size_oblivious_slice_scat -> origin/pianpwk/no_size_oblivious_slice_scat 2025-12-04T08:57:04.1768292Z * [new branch] pianpwk/oblivious_reshape_view_better -> origin/pianpwk/oblivious_reshape_view_better 2025-12-04T08:57:04.1769851Z * [new branch] pianpwk/pre_forward_hook -> origin/pianpwk/pre_forward_hook 2025-12-04T08:57:04.1771468Z * [new branch] pianpwk/skip_python_keys_alternate -> origin/pianpwk/skip_python_keys_alternate 2025-12-04T08:57:04.1773080Z * [new branch] pianpwk/skip_python_keys_in_guards -> origin/pianpwk/skip_python_keys_in_guards 2025-12-04T08:57:04.1774700Z * [new branch] pianpwk/sym_tokens_draft -> origin/pianpwk/sym_tokens_draft 2025-12-04T08:57:04.1776318Z * [new branch] pianpwk/symint_one_hot -> origin/pianpwk/symint_one_hot 2025-12-04T08:57:04.1778007Z * [new branch] pianpwk/test_pointwise_guard_or_false -> origin/pianpwk/test_pointwise_guard_or_false 2025-12-04T08:57:04.1779534Z * [new branch] pianpwk/totally_draft_sym_wrap -> origin/pianpwk/totally_draft_sym_wrap 2025-12-04T08:57:04.1782751Z * [new branch] pianpwk/try_dumb_stuff -> origin/pianpwk/try_dumb_stuff 2025-12-04T08:57:04.1783015Z * [new branch] pianpwk/try_dumb_stuff_2 -> origin/pianpwk/try_dumb_stuff_2 2025-12-04T08:57:04.1797663Z * [new branch] pianpwk/unbacked_dtensor_mm -> origin/pianpwk/unbacked_dtensor_mm 2025-12-04T08:57:04.1798059Z * [new branch] pianpwk/unbacked_tracing_12_2 -> origin/pianpwk/unbacked_tracing_12_2 2025-12-04T08:57:04.1798377Z * [new branch] pianpwk/user_symints -> origin/pianpwk/user_symints 2025-12-04T08:57:04.1798690Z * [new branch] pianpwk/wan21_reshape -> origin/pianpwk/wan21_reshape 2025-12-04T08:57:04.1799031Z * [new branch] piz/fix_partial_backward_1112 -> origin/piz/fix_partial_backward_1112 2025-12-04T08:57:04.1799304Z * [new branch] piz/prop_cache_clean -> origin/piz/prop_cache_clean 2025-12-04T08:57:04.1799561Z * [new branch] pool-separate -> origin/pool-separate 2025-12-04T08:57:04.1799825Z * [new branch] pr-156087 -> origin/pr-156087 2025-12-04T08:57:04.1800045Z * [new branch] pr/131860 -> origin/pr/131860 2025-12-04T08:57:04.1800307Z * [new branch] predispatch_to -> origin/predispatch_to 2025-12-04T08:57:04.1801334Z * [new branch] protect-c17 -> origin/protect-c17 2025-12-04T08:57:04.1803260Z * [new branch] pt-opt-cuda3 -> origin/pt-opt-cuda3 2025-12-04T08:57:04.1805233Z * [new branch] python_compiled_autograd -> origin/python_compiled_autograd 2025-12-04T08:57:04.1809207Z * [new branch] q1l1/fix_device_moved_constant_type_unknown -> origin/q1l1/fix_device_moved_constant_type_unknown 2025-12-04T08:57:04.1810825Z * [new branch] q1l1/fix_wrong_default_type_for_kernel_call_args -> origin/q1l1/fix_wrong_default_type_for_kernel_call_args 2025-12-04T08:57:04.1813314Z * [new branch] qchip/export-D54134695 -> origin/qchip/export-D54134695 2025-12-04T08:57:04.1815177Z * [new branch] quote-pytest_cache -> origin/quote-pytest_cache 2025-12-04T08:57:04.1817283Z * [new branch] reland-accgrad-stream-warn -> origin/reland-accgrad-stream-warn 2025-12-04T08:57:04.1819439Z * [new branch] release/1.10 -> origin/release/1.10 2025-12-04T08:57:04.1821210Z * [new branch] release/1.11 -> origin/release/1.11 2025-12-04T08:57:04.1822665Z * [new branch] release/1.12 -> origin/release/1.12 2025-12-04T08:57:04.1824181Z * [new branch] release/1.13 -> origin/release/1.13 2025-12-04T08:57:04.1825663Z * [new branch] release/1.4 -> origin/release/1.4 2025-12-04T08:57:04.1827001Z * [new branch] release/1.4.1 -> origin/release/1.4.1 2025-12-04T08:57:04.1828563Z * [new branch] release/1.5 -> origin/release/1.5 2025-12-04T08:57:04.1830162Z * [new branch] release/1.6 -> origin/release/1.6 2025-12-04T08:57:04.1831762Z * [new branch] release/1.7 -> origin/release/1.7 2025-12-04T08:57:04.1833407Z * [new branch] release/1.8 -> origin/release/1.8 2025-12-04T08:57:04.1834934Z * [new branch] release/1.9 -> origin/release/1.9 2025-12-04T08:57:04.1836613Z * [new branch] release/2.0 -> origin/release/2.0 2025-12-04T08:57:04.1838333Z * [new branch] release/2.1 -> origin/release/2.1 2025-12-04T08:57:04.1840156Z * [new branch] release/2.2 -> origin/release/2.2 2025-12-04T08:57:04.1842061Z * [new branch] release/2.3 -> origin/release/2.3 2025-12-04T08:57:04.1844141Z * [new branch] release/2.4 -> origin/release/2.4 2025-12-04T08:57:04.1846303Z * [new branch] release/2.5 -> origin/release/2.5 2025-12-04T08:57:04.1847941Z * [new branch] release/2.6 -> origin/release/2.6 2025-12-04T08:57:04.1849594Z * [new branch] release/2.7 -> origin/release/2.7 2025-12-04T08:57:04.1851370Z * [new branch] release/2.8 -> origin/release/2.8 2025-12-04T08:57:04.1853046Z * [new branch] release/2.9 -> origin/release/2.9 2025-12-04T08:57:04.1854808Z * [new branch] release_notes -> origin/release_notes 2025-12-04T08:57:04.1856435Z * [new branch] remove_pyinterpreter -> origin/remove_pyinterpreter 2025-12-04T08:57:04.1858367Z * [new branch] replace-pytorch-labs-20250812-195836 -> origin/replace-pytorch-labs-20250812-195836 2025-12-04T08:57:04.1859864Z * [new branch] replace-pytorch-labs-20250812-200248 -> origin/replace-pytorch-labs-20250812-200248 2025-12-04T08:57:04.1861335Z * [new branch] replace-pytorch-labs-20250812-200324 -> origin/replace-pytorch-labs-20250812-200324 2025-12-04T08:57:04.1862980Z * [new branch] replace-pytorch-labs-20250812-204020 -> origin/replace-pytorch-labs-20250812-204020 2025-12-04T08:57:04.1866090Z * [new branch] revert-131069-gh/krzysztofjordan/1/head -> origin/revert-131069-gh/krzysztofjordan/1/head 2025-12-04T08:57:04.1869127Z * [new branch] revert-131469-gh/andrewor14/51/head -> origin/revert-131469-gh/andrewor14/51/head 2025-12-04T08:57:04.1872115Z * [new branch] revert-152361-gh/fadara01/1/head -> origin/revert-152361-gh/fadara01/1/head 2025-12-04T08:57:04.1875142Z * [new branch] revert-156870-gh/skarjala/3/head -> origin/revert-156870-gh/skarjala/3/head 2025-12-04T08:57:04.1876968Z * [new branch] revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ -> origin/revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ 2025-12-04T08:57:04.1878535Z * [new branch] revert-hoo-invoke-subgraph -> origin/revert-hoo-invoke-subgraph 2025-12-04T08:57:04.1880273Z * [new branch] revert_always_build_distributed -> origin/revert_always_build_distributed 2025-12-04T08:57:04.1881918Z * [new branch] rms_norm_patch -> origin/rms_norm_patch 2025-12-04T08:57:04.1884129Z * [new branch] ruisi/fix_all_to_all_estimation -> origin/ruisi/fix_all_to_all_estimation 2025-12-04T08:57:04.1885725Z * [new branch] ruisi/fix_comm_estimation -> origin/ruisi/fix_comm_estimation 2025-12-04T08:57:04.1887090Z * [new branch] ruisi/fix_dynamic_shape_estimation -> origin/ruisi/fix_dynamic_shape_estimation 2025-12-04T08:57:04.1889103Z * [new branch] ruisi/fix_llama3_autobucketing -> origin/ruisi/fix_llama3_autobucketing 2025-12-04T08:57:04.1890845Z * [new branch] ruisi/fix_manual_bucketing_ep_pass -> origin/ruisi/fix_manual_bucketing_ep_pass 2025-12-04T08:57:04.1892721Z * [new branch] ruisi/manual_bucket_pass -> origin/ruisi/manual_bucket_pass 2025-12-04T08:57:04.1895158Z * [new branch] ryanguo99/cleanup-dynamo-expected-failures -> origin/ryanguo99/cleanup-dynamo-expected-failures 2025-12-04T08:57:04.1896508Z * [new branch] ryanguo99/fix-closure-var -> origin/ryanguo99/fix-closure-var 2025-12-04T08:57:04.1898631Z * [new branch] rzou/faketensor_bench -> origin/rzou/faketensor_bench 2025-12-04T08:57:04.1900036Z * [new branch] rzou/njt -> origin/rzou/njt 2025-12-04T08:57:04.1901637Z * [new branch] rzou/pca -> origin/rzou/pca 2025-12-04T08:57:04.1903125Z * [new branch] rzou/realprop -> origin/rzou/realprop 2025-12-04T08:57:04.1904884Z * [new branch] samplevllm -> origin/samplevllm 2025-12-04T08:57:04.1907574Z * [new branch] sanchitintel/weird_thing_with_test_cpu_select_algorithm -> origin/sanchitintel/weird_thing_with_test_cpu_select_algorithm 2025-12-04T08:57:04.1909183Z * [new branch] sapling-pr-archive-SS-JIA -> origin/sapling-pr-archive-SS-JIA 2025-12-04T08:57:04.1910915Z * [new branch] sapling-pr-archive-tushar00jain -> origin/sapling-pr-archive-tushar00jain 2025-12-04T08:57:04.1912437Z * [new branch] save -> origin/save 2025-12-04T08:57:04.1914316Z * [new branch] scaled_mm -> origin/scaled_mm 2025-12-04T08:57:04.1916072Z * [new branch] scan_attempt -> origin/scan_attempt 2025-12-04T08:57:04.1918146Z * [new branch] sdym/2.5.1 -> origin/sdym/2.5.1 2025-12-04T08:57:04.1919823Z * [new branch] sekyondaMeta-dynamoconfig-fix -> origin/sekyondaMeta-dynamoconfig-fix 2025-12-04T08:57:04.1921855Z * [new branch] shengf/fx-xform-perf -> origin/shengf/fx-xform-perf 2025-12-04T08:57:04.1923734Z * [new branch] shoumikhin-patch-1 -> origin/shoumikhin-patch-1 2025-12-04T08:57:04.1925554Z * [new branch] solve-accuracy-fix -> origin/solve-accuracy-fix 2025-12-04T08:57:04.1927171Z * [new branch] some_rocm_inductor_skips -> origin/some_rocm_inductor_skips 2025-12-04T08:57:04.1929260Z * [new branch] soulitzer/stash-tls-ac -> origin/soulitzer/stash-tls-ac 2025-12-04T08:57:04.1930981Z * [new branch] sparse-mm-bf16-support -> origin/sparse-mm-bf16-support 2025-12-04T08:57:04.1932594Z * [new branch] starterTaskUpdate -> origin/starterTaskUpdate 2025-12-04T08:57:04.1934467Z * [new branch] suo -> origin/suo 2025-12-04T08:57:04.1936094Z * [new branch] sve-poc -> origin/sve-poc 2025-12-04T08:57:04.1937740Z * [new branch] switch-bn -> origin/switch-bn 2025-12-04T08:57:04.1939859Z * [new branch] sy_annotation_in_autograd_hop -> origin/sy_annotation_in_autograd_hop 2025-12-04T08:57:04.1941565Z * [new branch] sy_aot_eager_record -> origin/sy_aot_eager_record 2025-12-04T08:57:04.1943203Z * [new branch] sy_custom_bucketing -> origin/sy_custom_bucketing 2025-12-04T08:57:04.1944824Z * [new branch] sy_debug_mode_test -> origin/sy_debug_mode_test 2025-12-04T08:57:04.1946716Z * [new branch] sy_deserialize -> origin/sy_deserialize 2025-12-04T08:57:04.1948270Z * [new branch] sy_dump_gm_code -> origin/sy_dump_gm_code 2025-12-04T08:57:04.1949826Z * [new branch] sy_exp -> origin/sy_exp 2025-12-04T08:57:04.1951569Z * [new branch] sy_export_annotation -> origin/sy_export_annotation 2025-12-04T08:57:04.1953181Z * [new branch] sy_invoke_subgraph -> origin/sy_invoke_subgraph 2025-12-04T08:57:04.1954874Z * [new branch] sy_kernel_bw_name -> origin/sy_kernel_bw_name 2025-12-04T08:57:04.1956509Z * [new branch] sy_multi_arch -> origin/sy_multi_arch 2025-12-04T08:57:04.1958248Z * [new branch] sy_nn_module_stack -> origin/sy_nn_module_stack 2025-12-04T08:57:04.1959890Z * [new branch] sy_original_dtensor -> origin/sy_original_dtensor 2025-12-04T08:57:04.1961498Z * [new branch] sy_profiler_cia -> origin/sy_profiler_cia 2025-12-04T08:57:04.1963128Z * [new branch] symm_mem_sync -> origin/symm_mem_sync 2025-12-04T08:57:04.1964805Z * [new branch] sympy-bottleneck-repro -> origin/sympy-bottleneck-repro 2025-12-04T08:57:04.1966900Z * [new branch] tensordict_integration -> origin/tensordict_integration 2025-12-04T08:57:04.1968618Z * [new branch] test-move-conda-builds -> origin/test-move-conda-builds 2025-12-04T08:57:04.1970243Z * [new branch] test-old -> origin/test-old 2025-12-04T08:57:04.1972392Z * [new branch] test/bmm_heur -> origin/test/bmm_heur 2025-12-04T08:57:04.1974695Z * [new branch] tianren/customOp_autotune_fix -> origin/tianren/customOp_autotune_fix 2025-12-04T08:57:04.1976433Z * [new branch] tianren/customOp_enable_max_autotune -> origin/tianren/customOp_enable_max_autotune 2025-12-04T08:57:04.1977764Z * [new branch] tianren/customOp_fusion -> origin/tianren/customOp_fusion 2025-12-04T08:57:04.1979433Z * [new branch] tianren/customop_collectiveop_benchmark -> origin/tianren/customop_collectiveop_benchmark 2025-12-04T08:57:04.1981210Z * [new branch] tianren/customop_collectiveop_benchmark_fix -> origin/tianren/customop_collectiveop_benchmark_fix 2025-12-04T08:57:04.1983058Z * [new branch] tianren/customop_dynamic_config -> origin/tianren/customop_dynamic_config 2025-12-04T08:57:04.1984522Z * [new branch] tianren/dynamic_range_input -> origin/tianren/dynamic_range_input 2025-12-04T08:57:04.1986126Z * [new branch] tianren/dynamic_range_input_fix -> origin/tianren/dynamic_range_input_fix 2025-12-04T08:57:04.1987677Z * [new branch] tianren/dynamic_range_input_merge -> origin/tianren/dynamic_range_input_merge 2025-12-04T08:57:04.1989228Z * [new branch] tianren/flex_paged_attn_fix_temp -> origin/tianren/flex_paged_attn_fix_temp 2025-12-04T08:57:04.1990898Z * [new branch] tianren/fx_codegen_dump -> origin/tianren/fx_codegen_dump 2025-12-04T08:57:04.1992458Z * [new branch] tianren/symmetric_memory -> origin/tianren/symmetric_memory 2025-12-04T08:57:04.1994031Z * [new branch] tianren/test -> origin/tianren/test 2025-12-04T08:57:04.1995693Z * [new branch] tidy_performance_cyy -> origin/tidy_performance_cyy 2025-12-04T08:57:04.1997289Z * [new branch] tmp -> origin/tmp 2025-12-04T08:57:04.1998986Z * [new branch] torchtitan_ep -> origin/torchtitan_ep 2025-12-04T08:57:04.2000685Z * [new branch] torchtitan_integration -> origin/torchtitan_integration 2025-12-04T08:57:04.2002382Z * [new branch] trace_fsdp_torchtune_lora -> origin/trace_fsdp_torchtune_lora 2025-12-04T08:57:04.2004236Z * [new branch] traceable_fsdp_unit_tests -> origin/traceable_fsdp_unit_tests 2025-12-04T08:57:04.2006123Z * [new branch] tree_loop_vec_base -> origin/tree_loop_vec_base 2025-12-04T08:57:04.2007697Z * [new branch] triton_kernel -> origin/triton_kernel 2025-12-04T08:57:04.2009403Z * [new branch] tt_pkg_1908 -> origin/tt_pkg_1908 2025-12-04T08:57:04.2011057Z * [new branch] type_dec -> origin/type_dec 2025-12-04T08:57:04.2012748Z * [new branch] udate-sphinx-dependancies -> origin/udate-sphinx-dependancies 2025-12-04T08:57:04.2015136Z * [new branch] update-audio-commit-hash/17630256502-1803-1 -> origin/update-audio-commit-hash/17630256502-1803-1 2025-12-04T08:57:04.2016587Z * [new branch] update-audio-commit-hash/19087141161-1916-1 -> origin/update-audio-commit-hash/19087141161-1916-1 2025-12-04T08:57:04.2018068Z * [new branch] update-audio-commit-hash/19250643381-1929-1 -> origin/update-audio-commit-hash/19250643381-1929-1 2025-12-04T08:57:04.2019553Z * [new branch] update-audio-commit-hash/19397724337-1935-1 -> origin/update-audio-commit-hash/19397724337-1935-1 2025-12-04T08:57:04.2020999Z * [new branch] update-audio-commit-hash/19555670148-1941-1 -> origin/update-audio-commit-hash/19555670148-1941-1 2025-12-04T08:57:04.2022934Z * [new branch] update-audio-commit-hash/19750627930-1946-1 -> origin/update-audio-commit-hash/19750627930-1946-1 2025-12-04T08:57:04.2025188Z * [new branch] update-triton-commit-hash/13663274526-1487-2 -> origin/update-triton-commit-hash/13663274526-1487-2 2025-12-04T08:57:04.2027222Z * [new branch] update-vision-commit-hash/19087141161-1916-1 -> origin/update-vision-commit-hash/19087141161-1916-1 2025-12-04T08:57:04.2028684Z * [new branch] update-vision-commit-hash/19184897099-1925-1 -> origin/update-vision-commit-hash/19184897099-1925-1 2025-12-04T08:57:04.2030148Z * [new branch] update-vision-commit-hash/19250643381-1929-1 -> origin/update-vision-commit-hash/19250643381-1929-1 2025-12-04T08:57:04.2031695Z * [new branch] update-vision-commit-hash/19381328640-1934-1 -> origin/update-vision-commit-hash/19381328640-1934-1 2025-12-04T08:57:04.2033182Z * [new branch] update-vision-commit-hash/19485237164-1938-1 -> origin/update-vision-commit-hash/19485237164-1938-1 2025-12-04T08:57:04.2035280Z * [new branch] update-vllm-commit-hash/18451675449-1879-1 -> origin/update-vllm-commit-hash/18451675449-1879-1 2025-12-04T08:57:04.2036895Z * [new branch] update-vllm-dockerfile -> origin/update-vllm-dockerfile 2025-12-04T08:57:04.2039265Z * [new branch] update-xla-commit-hash/19224287370-211-1 -> origin/update-xla-commit-hash/19224287370-211-1 2025-12-04T08:57:04.2040779Z * [new branch] update-xla-commit-hash/19422028566-212-1 -> origin/update-xla-commit-hash/19422028566-212-1 2025-12-04T08:57:04.2042309Z * [new branch] update-xla-commit-hash/19626841311-213-1 -> origin/update-xla-commit-hash/19626841311-213-1 2025-12-04T08:57:04.2043985Z * [new branch] update_docs_torch_multinomial_issue#125388 -> origin/update_docs_torch_multinomial_issue#125388 2025-12-04T08:57:04.2045514Z * [new branch] update_operator_readme -> origin/update_operator_readme 2025-12-04T08:57:04.2047212Z * [new branch] update_slow_tests_1722488736 -> origin/update_slow_tests_1722488736 2025-12-04T08:57:04.2048897Z * [new branch] update_slow_tests_1722879173 -> origin/update_slow_tests_1722879173 2025-12-04T08:57:04.2050564Z * [new branch] update_slow_tests_1762155677 -> origin/update_slow_tests_1762155677 2025-12-04T08:57:04.2052267Z * [new branch] update_slow_tests_1763365283 -> origin/update_slow_tests_1763365283 2025-12-04T08:57:04.2054184Z * [new branch] update_submodule_FBGEMM -> origin/update_submodule_FBGEMM 2025-12-04T08:57:04.2055647Z * [new branch] update_submodule_kineto -> origin/update_submodule_kineto 2025-12-04T08:57:04.2057395Z * [new branch] update_submodule_tensorpipe -> origin/update_submodule_tensorpipe 2025-12-04T08:57:04.2059054Z * [new branch] upload-tests-for-autorevert -> origin/upload-tests-for-autorevert 2025-12-04T08:57:04.2060685Z * [new branch] v0.1.2 -> origin/v0.1.2 2025-12-04T08:57:04.2062442Z * [new branch] v1.0.1 -> origin/v1.0.1 2025-12-04T08:57:04.2064425Z * [new branch] v1.0.3 -> origin/v1.0.3 2025-12-04T08:57:04.2066227Z * [new branch] v1.1.0 -> origin/v1.1.0 2025-12-04T08:57:04.2068148Z * [new branch] v1.2.0 -> origin/v1.2.0 2025-12-04T08:57:04.2069877Z * [new branch] v1.3.0 -> origin/v1.3.0 2025-12-04T08:57:04.2071643Z * [new branch] v1.3.1 -> origin/v1.3.1 2025-12-04T08:57:04.2073295Z * [new branch] validate_fn -> origin/validate_fn 2025-12-04T08:57:04.2075019Z * [new branch] validations_2.6 -> origin/validations_2.6 2025-12-04T08:57:04.2076931Z * [new branch] validations_2.8 -> origin/validations_2.8 2025-12-04T08:57:04.2078618Z * [new branch] varlen-api -> origin/varlen-api 2025-12-04T08:57:04.2080241Z * [new branch] varlen-api-backup -> origin/varlen-api-backup 2025-12-04T08:57:04.2081835Z * [new branch] varlen_batch_invariance -> origin/varlen_batch_invariance 2025-12-04T08:57:04.2083748Z * [new branch] viable/strict -> origin/viable/strict 2025-12-04T08:57:04.2086111Z * [new branch] vishal9-team/dtensor_parallelism_toy -> origin/vishal9-team/dtensor_parallelism_toy 2025-12-04T08:57:04.2087618Z * [new branch] vllmbuildci -> origin/vllmbuildci 2025-12-04T08:57:04.2089329Z * [new branch] vllmpin -> origin/vllmpin 2025-12-04T08:57:04.2091172Z * [new branch] vscode-recommend-pyrefly -> origin/vscode-recommend-pyrefly 2025-12-04T08:57:04.2092893Z * [new branch] wdvr-patch-1 -> origin/wdvr-patch-1 2025-12-04T08:57:04.2095092Z * [new branch] wdvr/iss_145259 -> origin/wdvr/iss_145259 2025-12-04T08:57:04.2097133Z * [new branch] whc/pei -> origin/whc/pei 2025-12-04T08:57:04.2098738Z * [new branch] whc/pp_fix -> origin/whc/pp_fix 2025-12-04T08:57:04.2100356Z * [new branch] whc/sharding -> origin/whc/sharding 2025-12-04T08:57:04.2101815Z * [new branch] whc/sharding2 -> origin/whc/sharding2 2025-12-04T08:57:04.2103240Z * [new branch] whc/uneven -> origin/whc/uneven 2025-12-04T08:57:04.2105244Z * [new branch] whc/uneven-merge -> origin/whc/uneven-merge 2025-12-04T08:57:04.2107077Z * [new branch] win_warnings -> origin/win_warnings 2025-12-04T08:57:04.2109029Z * [new branch] windows_libtorch_free -> origin/windows_libtorch_free 2025-12-04T08:57:04.2110670Z * [new branch] xmfan-war -> origin/xmfan-war 2025-12-04T08:57:04.2112796Z * [new branch] xmfan/ca_0516 -> origin/xmfan/ca_0516 2025-12-04T08:57:04.2114331Z * [new branch] xmfan/ca_1051b93192 -> origin/xmfan/ca_1051b93192 2025-12-04T08:57:04.2116008Z * [new branch] xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 -> origin/xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 2025-12-04T08:57:04.2117102Z * [new branch] xmfan/ca_5a2be192d1 -> origin/xmfan/ca_5a2be192d1 2025-12-04T08:57:04.2118912Z * [new branch] xmfan/ca_9d59b516e9 -> origin/xmfan/ca_9d59b516e9 2025-12-04T08:57:04.2120072Z * [new branch] xmfan/ca_apr8 -> origin/xmfan/ca_apr8 2025-12-04T08:57:04.2121695Z * [new branch] xmfan/ca_base -> origin/xmfan/ca_base 2025-12-04T08:57:04.2123527Z * [new branch] xmfan/ca_dynamic -> origin/xmfan/ca_dynamic 2025-12-04T08:57:04.2125353Z * [new branch] xmfan/ca_fix_dyn -> origin/xmfan/ca_fix_dyn 2025-12-04T08:57:04.2126995Z * [new branch] xmfan/ca_fix_lowering -> origin/xmfan/ca_fix_lowering 2025-12-04T08:57:04.2128987Z * [new branch] xmfan/ca_fix_polyfills -> origin/xmfan/ca_fix_polyfills 2025-12-04T08:57:04.2130466Z * [new branch] xmfan/ca_jan3 -> origin/xmfan/ca_jan3 2025-12-04T08:57:04.2132036Z * [new branch] xmfan/ca_jun18 -> origin/xmfan/ca_jun18 2025-12-04T08:57:04.2133740Z * [new branch] xmfan/ca_jun24 -> origin/xmfan/ca_jun24 2025-12-04T08:57:04.2135411Z * [new branch] xmfan/ca_nested -> origin/xmfan/ca_nested 2025-12-04T08:57:04.2137011Z * [new branch] xmfan/ca_overhead -> origin/xmfan/ca_overhead 2025-12-04T08:57:04.2138755Z * [new branch] xmfan/ca_overhead_0eba7e5451 -> origin/xmfan/ca_overhead_0eba7e5451 2025-12-04T08:57:04.2140303Z * [new branch] xmfan/cacu_jun18 -> origin/xmfan/cacu_jun18 2025-12-04T08:57:04.2141818Z * [new branch] xmfan/cacu_jun19 -> origin/xmfan/cacu_jun19 2025-12-04T08:57:04.2143375Z * [new branch] xmfan/cacu_jun4 -> origin/xmfan/cacu_jun4 2025-12-04T08:57:04.2144916Z * [new branch] xmfan/disable_duck_shape -> origin/xmfan/disable_duck_shape 2025-12-04T08:57:04.2146533Z * [new branch] xmfan/fca_cpp_node_passthrough -> origin/xmfan/fca_cpp_node_passthrough 2025-12-04T08:57:04.2148231Z * [new branch] xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 2025-12-04T08:57:04.2149878Z * [new branch] xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 2025-12-04T08:57:04.2151281Z * [new branch] xmfan/single_step -> origin/xmfan/single_step 2025-12-04T08:57:04.2152807Z * [new branch] xmfan/sth_0829 -> origin/xmfan/sth_0829 2025-12-04T08:57:04.2154445Z * [new branch] xmfan/test -> origin/xmfan/test 2025-12-04T08:57:04.2156625Z * [new branch] yguo/debug-0226-constexpr -> origin/yguo/debug-0226-constexpr 2025-12-04T08:57:04.2158137Z * [new branch] yguo/new_latest_changes -> origin/yguo/new_latest_changes 2025-12-04T08:57:04.2159687Z * [new branch] yguo/patch_constexpr_changes -> origin/yguo/patch_constexpr_changes 2025-12-04T08:57:04.2161733Z * [new branch] yiming/bootcamp -> origin/yiming/bootcamp 2025-12-04T08:57:04.2163277Z * [new branch] yiming/run_with_start_end_rng_hop -> origin/yiming/run_with_start_end_rng_hop 2025-12-04T08:57:04.2164853Z * [new branch] yolo-llama3 -> origin/yolo-llama3 2025-12-04T08:57:04.2166929Z * [new branch] zainr/canary-test -> origin/zainr/canary-test 2025-12-04T08:57:04.2168532Z * [new branch] zainr/cleanup-gh-runners -> origin/zainr/cleanup-gh-runners 2025-12-04T08:57:04.2170013Z * [new branch] zainr/pull-migration-c -> origin/zainr/pull-migration-c 2025-12-04T08:57:04.2171514Z * [new branch] zainr/test2 -> origin/zainr/test2 2025-12-04T08:57:04.2173289Z * [new branch] zasdfgbnm-patch-3 -> origin/zasdfgbnm-patch-3 2025-12-04T08:57:04.2175084Z * [new branch] zb2p -> origin/zb2p 2025-12-04T08:57:04.2177196Z * [new branch] zeros-and-scatter-part2 -> origin/zeros-and-scatter-part2 2025-12-04T08:57:04.2179807Z * [new branch] zhxchen17/ci/vllm_lora_oom -> origin/zhxchen17/ci/vllm_lora_oom 2025-12-04T08:57:04.2181300Z * [new branch] zhxchen17/ci/vllm_multimodal_oom -> origin/zhxchen17/ci/vllm_multimodal_oom 2025-12-04T08:57:04.2182790Z * [new branch] zhxchen17/ci/vllm_pin -> origin/zhxchen17/ci/vllm_pin 2025-12-04T08:57:04.2184985Z * [new branch] zhxchen17/dynamo/unsafe_drop_all_guards -> origin/zhxchen17/dynamo/unsafe_drop_all_guards 2025-12-04T08:57:04.2187009Z * [new branch] zhxchen17/export/call_override -> origin/zhxchen17/export/call_override 2025-12-04T08:57:04.2188526Z * [new branch] zhxchen17/export/codemod1 -> origin/zhxchen17/export/codemod1 2025-12-04T08:57:04.2190085Z * [new branch] zhxchen17/export/ctx_return -> origin/zhxchen17/export/ctx_return 2025-12-04T08:57:04.2191699Z * [new branch] zhxchen17/export/disable_side_effect_warn -> origin/zhxchen17/export/disable_side_effect_warn 2025-12-04T08:57:04.2193767Z * [new branch] zhxchen17/export/pytree_check -> origin/zhxchen17/export/pytree_check 2025-12-04T08:57:04.2195797Z * [new branch] zhxchen17/precompile/aoti -> origin/zhxchen17/precompile/aoti 2025-12-04T08:57:04.2197377Z * [new branch] zhxchen17/precompile/globals -> origin/zhxchen17/precompile/globals 2025-12-04T08:57:04.2198930Z * [new branch] zhxchen17/precompile/inductor_guards -> origin/zhxchen17/precompile/inductor_guards 2025-12-04T08:57:04.2201434Z * [new branch] zhxchen17/scratch/0 -> origin/zhxchen17/scratch/0 2025-12-04T08:57:04.2202993Z * [new branch] zhxchen17/torch_export_api_update -> origin/zhxchen17/torch_export_api_update 2025-12-04T08:57:04.2205221Z * [new branch] zhxhcen17/moodycamel -> origin/zhxhcen17/moodycamel 2025-12-04T08:57:04.2209138Z * [new branch] zxiiro/build-times -> origin/zxiiro/build-times 2025-12-04T08:57:04.2210674Z * [new branch] zxiiro/c7i.2xlarge -> origin/zxiiro/c7i.2xlarge 2025-12-04T08:57:04.2212228Z * [new branch] zxiiro/c7i.2xlarge.h100 -> origin/zxiiro/c7i.2xlarge.h100 2025-12-04T08:57:04.2213848Z * [new branch] zxiiro/main -> origin/zxiiro/main 2025-12-04T08:57:04.2215492Z * [new branch] zxiiro/risc64 -> origin/zxiiro/risc64 2025-12-04T08:57:04.2217125Z * [new branch] zxiiro/test-multicloud-arc -> origin/zxiiro/test-multicloud-arc 2025-12-04T08:57:04.2218407Z * [new tag] bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug -> bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug 2025-12-04T08:57:04.2219768Z * [new tag] ci/binaries/77164 -> ci/binaries/77164 2025-12-04T08:57:04.2221221Z * [new tag] ciflow/b200/115316 -> ciflow/b200/115316 2025-12-04T08:57:04.2222147Z * [new tag] ciflow/b200/160685 -> ciflow/b200/160685 2025-12-04T08:57:04.2223177Z * [new tag] ciflow/b200/161607 -> ciflow/b200/161607 2025-12-04T08:57:04.2224306Z * [new tag] ciflow/b200/161938 -> ciflow/b200/161938 2025-12-04T08:57:04.2225553Z * [new tag] ciflow/b200/167207 -> ciflow/b200/167207 2025-12-04T08:57:04.2226584Z * [new tag] ciflow/b200/167989 -> ciflow/b200/167989 2025-12-04T08:57:04.2227731Z * [new tag] ciflow/b200/168096 -> ciflow/b200/168096 2025-12-04T08:57:04.2228906Z * [new tag] ciflow/b200/168175 -> ciflow/b200/168175 2025-12-04T08:57:04.2229987Z * [new tag] ciflow/b200/168195 -> ciflow/b200/168195 2025-12-04T08:57:04.2231328Z * [new tag] ciflow/b200/169200 -> ciflow/b200/169200 2025-12-04T08:57:04.2232210Z * [new tag] ciflow/b200/169216 -> ciflow/b200/169216 2025-12-04T08:57:04.2233876Z * [new tag] ciflow/b200/169380 -> ciflow/b200/169380 2025-12-04T08:57:04.2235506Z * [new tag] ciflow/b200/169412 -> ciflow/b200/169412 2025-12-04T08:57:04.2236902Z * [new tag] ciflow/b200/169470 -> ciflow/b200/169470 2025-12-04T08:57:04.2238040Z * [new tag] ciflow/b200/169471 -> ciflow/b200/169471 2025-12-04T08:57:04.2239178Z * [new tag] ciflow/b200/169472 -> ciflow/b200/169472 2025-12-04T08:57:04.2241085Z * [new tag] ciflow/b200/169514 -> ciflow/b200/169514 2025-12-04T08:57:04.2242213Z * [new tag] ciflow/b200/169517 -> ciflow/b200/169517 2025-12-04T08:57:04.2243638Z * [new tag] ciflow/binaries/165922 -> ciflow/binaries/165922 2025-12-04T08:57:04.2244713Z * [new tag] ciflow/binaries/169510 -> ciflow/binaries/169510 2025-12-04T08:57:04.2246102Z * [new tag] ciflow/binaries_wheel/157994 -> ciflow/binaries_wheel/157994 2025-12-04T08:57:04.2247242Z * [new tag] ciflow/binaries_wheel/166829 -> ciflow/binaries_wheel/166829 2025-12-04T08:57:04.2248350Z * [new tag] ciflow/binaries_wheel/167972 -> ciflow/binaries_wheel/167972 2025-12-04T08:57:04.2249501Z * [new tag] ciflow/binaries_wheel/167981 -> ciflow/binaries_wheel/167981 2025-12-04T08:57:04.2250788Z * [new tag] ciflow/dynamo/167695 -> ciflow/dynamo/167695 2025-12-04T08:57:04.2251900Z * [new tag] ciflow/dynamo/168096 -> ciflow/dynamo/168096 2025-12-04T08:57:04.2253061Z * [new tag] ciflow/dynamo/169525 -> ciflow/dynamo/169525 2025-12-04T08:57:04.2254512Z * [new tag] ciflow/h100-cutlass-backend/161938 -> ciflow/h100-cutlass-backend/161938 2025-12-04T08:57:04.2255595Z * [new tag] ciflow/h100-cutlass-backend/161940 -> ciflow/h100-cutlass-backend/161940 2025-12-04T08:57:04.2256894Z * [new tag] ciflow/h100-distributed/168923 -> ciflow/h100-distributed/168923 2025-12-04T08:57:04.2258096Z * [new tag] ciflow/h100-symm-mem/167552 -> ciflow/h100-symm-mem/167552 2025-12-04T08:57:04.2259131Z * [new tag] ciflow/h100-symm-mem/168129 -> ciflow/h100-symm-mem/168129 2025-12-04T08:57:04.2260172Z * [new tag] ciflow/h100-symm-mem/168917 -> ciflow/h100-symm-mem/168917 2025-12-04T08:57:04.2261507Z * [new tag] ciflow/h100-symm-mem/169156 -> ciflow/h100-symm-mem/169156 2025-12-04T08:57:04.2262535Z * [new tag] ciflow/h100-symm-mem/169200 -> ciflow/h100-symm-mem/169200 2025-12-04T08:57:04.2263587Z * [new tag] ciflow/h100-symm-mem/169216 -> ciflow/h100-symm-mem/169216 2025-12-04T08:57:04.2264611Z * [new tag] ciflow/h100-symm-mem/169338 -> ciflow/h100-symm-mem/169338 2025-12-04T08:57:04.2265746Z * [new tag] ciflow/h100-symm-mem/169355 -> ciflow/h100-symm-mem/169355 2025-12-04T08:57:04.2266774Z * [new tag] ciflow/h100-symm-mem/169543 -> ciflow/h100-symm-mem/169543 2025-12-04T08:57:04.2267965Z * [new tag] ciflow/h100/115316 -> ciflow/h100/115316 2025-12-04T08:57:04.2268993Z * [new tag] ciflow/h100/160685 -> ciflow/h100/160685 2025-12-04T08:57:04.2270012Z * [new tag] ciflow/h100/160729 -> ciflow/h100/160729 2025-12-04T08:57:04.2271067Z * [new tag] ciflow/h100/161607 -> ciflow/h100/161607 2025-12-04T08:57:04.2272114Z * [new tag] ciflow/h100/161938 -> ciflow/h100/161938 2025-12-04T08:57:04.2273138Z * [new tag] ciflow/h100/167207 -> ciflow/h100/167207 2025-12-04T08:57:04.2274333Z * [new tag] ciflow/h100/167989 -> ciflow/h100/167989 2025-12-04T08:57:04.2275188Z * [new tag] ciflow/h100/168096 -> ciflow/h100/168096 2025-12-04T08:57:04.2276288Z * [new tag] ciflow/h100/168175 -> ciflow/h100/168175 2025-12-04T08:57:04.2277475Z * [new tag] ciflow/h100/168195 -> ciflow/h100/168195 2025-12-04T08:57:04.2278560Z * [new tag] ciflow/h100/168980 -> ciflow/h100/168980 2025-12-04T08:57:04.2279905Z * [new tag] ciflow/h100/169200 -> ciflow/h100/169200 2025-12-04T08:57:04.2281337Z * [new tag] ciflow/h100/169216 -> ciflow/h100/169216 2025-12-04T08:57:04.2282656Z * [new tag] ciflow/h100/169380 -> ciflow/h100/169380 2025-12-04T08:57:04.2283743Z * [new tag] ciflow/h100/169412 -> ciflow/h100/169412 2025-12-04T08:57:04.2284856Z * [new tag] ciflow/h100/169470 -> ciflow/h100/169470 2025-12-04T08:57:04.2285956Z * [new tag] ciflow/h100/169471 -> ciflow/h100/169471 2025-12-04T08:57:04.2287045Z * [new tag] ciflow/h100/169472 -> ciflow/h100/169472 2025-12-04T08:57:04.2288166Z * [new tag] ciflow/h100/169514 -> ciflow/h100/169514 2025-12-04T08:57:04.2289451Z * [new tag] ciflow/inductor-cu126/168096 -> ciflow/inductor-cu126/168096 2025-12-04T08:57:04.2291063Z * [new tag] ciflow/inductor-micro-benchmark-cpu-x86/168096 -> ciflow/inductor-micro-benchmark-cpu-x86/168096 2025-12-04T08:57:04.2292258Z * [new tag] ciflow/inductor-micro-benchmark/166165 -> ciflow/inductor-micro-benchmark/166165 2025-12-04T08:57:04.2293171Z * [new tag] ciflow/inductor-micro-benchmark/168096 -> ciflow/inductor-micro-benchmark/168096 2025-12-04T08:57:04.2294802Z * [new tag] ciflow/inductor-perf-compare/168096 -> ciflow/inductor-perf-compare/168096 2025-12-04T08:57:04.2296395Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/168073 -> ciflow/inductor-perf-test-nightly-rocm-mi300/168073 2025-12-04T08:57:04.2297285Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/168096 -> ciflow/inductor-perf-test-nightly-rocm-mi300/168096 2025-12-04T08:57:04.2298548Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/169024 -> ciflow/inductor-perf-test-nightly-rocm-mi300/169024 2025-12-04T08:57:04.2299803Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi355/169024 -> ciflow/inductor-perf-test-nightly-rocm-mi355/169024 2025-12-04T08:57:04.2301002Z * [new tag] ciflow/inductor-perf-test-nightly/168096 -> ciflow/inductor-perf-test-nightly/168096 2025-12-04T08:57:04.2302232Z * [new tag] ciflow/inductor-periodic/168096 -> ciflow/inductor-periodic/168096 2025-12-04T08:57:04.2303129Z * [new tag] ciflow/inductor-periodic/169024 -> ciflow/inductor-periodic/169024 2025-12-04T08:57:04.2304416Z * [new tag] ciflow/inductor-periodic/169425 -> ciflow/inductor-periodic/169425 2025-12-04T08:57:04.2306195Z * [new tag] ciflow/inductor-rocm-mi200/165545 -> ciflow/inductor-rocm-mi200/165545 2025-12-04T08:57:04.2307399Z * [new tag] ciflow/inductor-rocm-mi200/165997 -> ciflow/inductor-rocm-mi200/165997 2025-12-04T08:57:04.2308300Z * [new tag] ciflow/inductor-rocm-mi200/168096 -> ciflow/inductor-rocm-mi200/168096 2025-12-04T08:57:04.2309635Z * [new tag] ciflow/inductor-rocm-mi200/169063 -> ciflow/inductor-rocm-mi200/169063 2025-12-04T08:57:04.2310544Z * [new tag] ciflow/inductor-rocm-mi200/169425 -> ciflow/inductor-rocm-mi200/169425 2025-12-04T08:57:04.2311923Z * [new tag] ciflow/inductor-rocm-mi300/165545 -> ciflow/inductor-rocm-mi300/165545 2025-12-04T08:57:04.2312981Z * [new tag] ciflow/inductor-rocm-mi300/168096 -> ciflow/inductor-rocm-mi300/168096 2025-12-04T08:57:04.2314038Z * [new tag] ciflow/inductor-rocm-mi300/169063 -> ciflow/inductor-rocm-mi300/169063 2025-12-04T08:57:04.2315193Z * [new tag] ciflow/inductor-rocm-mi300/169425 -> ciflow/inductor-rocm-mi300/169425 2025-12-04T08:57:04.2316655Z * [new tag] ciflow/inductor-rocm/162052 -> ciflow/inductor-rocm/162052 2025-12-04T08:57:04.2317540Z * [new tag] ciflow/inductor-rocm/168971 -> ciflow/inductor-rocm/168971 2025-12-04T08:57:04.2318945Z * [new tag] ciflow/inductor-windows/168096 -> ciflow/inductor-windows/168096 2025-12-04T08:57:04.2320233Z * [new tag] ciflow/inductor/144542 -> ciflow/inductor/144542 2025-12-04T08:57:04.2321221Z * [new tag] ciflow/inductor/146506 -> ciflow/inductor/146506 2025-12-04T08:57:04.2322340Z * [new tag] ciflow/inductor/147990 -> ciflow/inductor/147990 2025-12-04T08:57:04.2323467Z * [new tag] ciflow/inductor/148294 -> ciflow/inductor/148294 2025-12-04T08:57:04.2324576Z * [new tag] ciflow/inductor/148492 -> ciflow/inductor/148492 2025-12-04T08:57:04.2325440Z * [new tag] ciflow/inductor/157149 -> ciflow/inductor/157149 2025-12-04T08:57:04.2326609Z * [new tag] ciflow/inductor/157994 -> ciflow/inductor/157994 2025-12-04T08:57:04.2327583Z * [new tag] ciflow/inductor/160685 -> ciflow/inductor/160685 2025-12-04T08:57:04.2328691Z * [new tag] ciflow/inductor/160686 -> ciflow/inductor/160686 2025-12-04T08:57:04.2329752Z * [new tag] ciflow/inductor/160687 -> ciflow/inductor/160687 2025-12-04T08:57:04.2330741Z * [new tag] ciflow/inductor/160688 -> ciflow/inductor/160688 2025-12-04T08:57:04.2332178Z * [new tag] ciflow/inductor/160706 -> ciflow/inductor/160706 2025-12-04T08:57:04.2333833Z * [new tag] ciflow/inductor/160729 -> ciflow/inductor/160729 2025-12-04T08:57:04.2335462Z * [new tag] ciflow/inductor/161938 -> ciflow/inductor/161938 2025-12-04T08:57:04.2336572Z * [new tag] ciflow/inductor/161939 -> ciflow/inductor/161939 2025-12-04T08:57:04.2338243Z * [new tag] ciflow/inductor/161940 -> ciflow/inductor/161940 2025-12-04T08:57:04.2339365Z * [new tag] ciflow/inductor/162052 -> ciflow/inductor/162052 2025-12-04T08:57:04.2340521Z * [new tag] ciflow/inductor/162275 -> ciflow/inductor/162275 2025-12-04T08:57:04.2341632Z * [new tag] ciflow/inductor/162795 -> ciflow/inductor/162795 2025-12-04T08:57:04.2342931Z * [new tag] ciflow/inductor/163245 -> ciflow/inductor/163245 2025-12-04T08:57:04.2343928Z * [new tag] ciflow/inductor/163335 -> ciflow/inductor/163335 2025-12-04T08:57:04.2345076Z * [new tag] ciflow/inductor/163503 -> ciflow/inductor/163503 2025-12-04T08:57:04.2346216Z * [new tag] ciflow/inductor/163942 -> ciflow/inductor/163942 2025-12-04T08:57:04.2347467Z * [new tag] ciflow/inductor/165270 -> ciflow/inductor/165270 2025-12-04T08:57:04.2348553Z * [new tag] ciflow/inductor/165274 -> ciflow/inductor/165274 2025-12-04T08:57:04.2349645Z * [new tag] ciflow/inductor/165322 -> ciflow/inductor/165322 2025-12-04T08:57:04.2350751Z * [new tag] ciflow/inductor/165597 -> ciflow/inductor/165597 2025-12-04T08:57:04.2351880Z * [new tag] ciflow/inductor/166063 -> ciflow/inductor/166063 2025-12-04T08:57:04.2352965Z * [new tag] ciflow/inductor/166075 -> ciflow/inductor/166075 2025-12-04T08:57:04.2354073Z * [new tag] ciflow/inductor/166165 -> ciflow/inductor/166165 2025-12-04T08:57:04.2355535Z * [new tag] ciflow/inductor/166254 -> ciflow/inductor/166254 2025-12-04T08:57:04.2356555Z * [new tag] ciflow/inductor/166483 -> ciflow/inductor/166483 2025-12-04T08:57:04.2357659Z * [new tag] ciflow/inductor/166494 -> ciflow/inductor/166494 2025-12-04T08:57:04.2358756Z * [new tag] ciflow/inductor/166545 -> ciflow/inductor/166545 2025-12-04T08:57:04.2359867Z * [new tag] ciflow/inductor/166788 -> ciflow/inductor/166788 2025-12-04T08:57:04.2361110Z * [new tag] ciflow/inductor/166846 -> ciflow/inductor/166846 2025-12-04T08:57:04.2362202Z * [new tag] ciflow/inductor/167300 -> ciflow/inductor/167300 2025-12-04T08:57:04.2363316Z * [new tag] ciflow/inductor/167407 -> ciflow/inductor/167407 2025-12-04T08:57:04.2364521Z * [new tag] ciflow/inductor/167536 -> ciflow/inductor/167536 2025-12-04T08:57:04.2365629Z * [new tag] ciflow/inductor/167552 -> ciflow/inductor/167552 2025-12-04T08:57:04.2366737Z * [new tag] ciflow/inductor/167555 -> ciflow/inductor/167555 2025-12-04T08:57:04.2367977Z * [new tag] ciflow/inductor/167583 -> ciflow/inductor/167583 2025-12-04T08:57:04.2369063Z * [new tag] ciflow/inductor/167599 -> ciflow/inductor/167599 2025-12-04T08:57:04.2370165Z * [new tag] ciflow/inductor/167647 -> ciflow/inductor/167647 2025-12-04T08:57:04.2371264Z * [new tag] ciflow/inductor/167677 -> ciflow/inductor/167677 2025-12-04T08:57:04.2372375Z * [new tag] ciflow/inductor/167680 -> ciflow/inductor/167680 2025-12-04T08:57:04.2373552Z * [new tag] ciflow/inductor/167695 -> ciflow/inductor/167695 2025-12-04T08:57:04.2374722Z * [new tag] ciflow/inductor/167742 -> ciflow/inductor/167742 2025-12-04T08:57:04.2375834Z * [new tag] ciflow/inductor/167768 -> ciflow/inductor/167768 2025-12-04T08:57:04.2377490Z * [new tag] ciflow/inductor/167773 -> ciflow/inductor/167773 2025-12-04T08:57:04.2378614Z * [new tag] ciflow/inductor/167781 -> ciflow/inductor/167781 2025-12-04T08:57:04.2379779Z * [new tag] ciflow/inductor/167880 -> ciflow/inductor/167880 2025-12-04T08:57:04.2380882Z * [new tag] ciflow/inductor/167887 -> ciflow/inductor/167887 2025-12-04T08:57:04.2381999Z * [new tag] ciflow/inductor/167972 -> ciflow/inductor/167972 2025-12-04T08:57:04.2383090Z * [new tag] ciflow/inductor/167989 -> ciflow/inductor/167989 2025-12-04T08:57:04.2384257Z * [new tag] ciflow/inductor/168002 -> ciflow/inductor/168002 2025-12-04T08:57:04.2385388Z * [new tag] ciflow/inductor/168050 -> ciflow/inductor/168050 2025-12-04T08:57:04.2386467Z * [new tag] ciflow/inductor/168051 -> ciflow/inductor/168051 2025-12-04T08:57:04.2387626Z * [new tag] ciflow/inductor/168052 -> ciflow/inductor/168052 2025-12-04T08:57:04.2388817Z * [new tag] ciflow/inductor/168073 -> ciflow/inductor/168073 2025-12-04T08:57:04.2389902Z * [new tag] ciflow/inductor/168096 -> ciflow/inductor/168096 2025-12-04T08:57:04.2391004Z * [new tag] ciflow/inductor/168114 -> ciflow/inductor/168114 2025-12-04T08:57:04.2392113Z * [new tag] ciflow/inductor/168115 -> ciflow/inductor/168115 2025-12-04T08:57:04.2393276Z * [new tag] ciflow/inductor/168127 -> ciflow/inductor/168127 2025-12-04T08:57:04.2394595Z * [new tag] ciflow/inductor/168129 -> ciflow/inductor/168129 2025-12-04T08:57:04.2395801Z * [new tag] ciflow/inductor/168157 -> ciflow/inductor/168157 2025-12-04T08:57:04.2396904Z * [new tag] ciflow/inductor/168175 -> ciflow/inductor/168175 2025-12-04T08:57:04.2398135Z * [new tag] ciflow/inductor/168185 -> ciflow/inductor/168185 2025-12-04T08:57:04.2399016Z * [new tag] ciflow/inductor/168195 -> ciflow/inductor/168195 2025-12-04T08:57:04.2400216Z * [new tag] ciflow/inductor/168209 -> ciflow/inductor/168209 2025-12-04T08:57:04.2401327Z * [new tag] ciflow/inductor/168266 -> ciflow/inductor/168266 2025-12-04T08:57:04.2402453Z * [new tag] ciflow/inductor/168316 -> ciflow/inductor/168316 2025-12-04T08:57:04.2403847Z * [new tag] ciflow/inductor/168326 -> ciflow/inductor/168326 2025-12-04T08:57:04.2405257Z * [new tag] ciflow/inductor/168368 -> ciflow/inductor/168368 2025-12-04T08:57:04.2406559Z * [new tag] ciflow/inductor/168894 -> ciflow/inductor/168894 2025-12-04T08:57:04.2407667Z * [new tag] ciflow/inductor/168934 -> ciflow/inductor/168934 2025-12-04T08:57:04.2408786Z * [new tag] ciflow/inductor/168939 -> ciflow/inductor/168939 2025-12-04T08:57:04.2409894Z * [new tag] ciflow/inductor/168946 -> ciflow/inductor/168946 2025-12-04T08:57:04.2411041Z * [new tag] ciflow/inductor/168950 -> ciflow/inductor/168950 2025-12-04T08:57:04.2412516Z * [new tag] ciflow/inductor/168951 -> ciflow/inductor/168951 2025-12-04T08:57:04.2413719Z * [new tag] ciflow/inductor/168952 -> ciflow/inductor/168952 2025-12-04T08:57:04.2414893Z * [new tag] ciflow/inductor/168955 -> ciflow/inductor/168955 2025-12-04T08:57:04.2416022Z * [new tag] ciflow/inductor/168971 -> ciflow/inductor/168971 2025-12-04T08:57:04.2417097Z * [new tag] ciflow/inductor/168979 -> ciflow/inductor/168979 2025-12-04T08:57:04.2418211Z * [new tag] ciflow/inductor/168980 -> ciflow/inductor/168980 2025-12-04T08:57:04.2419479Z * [new tag] ciflow/inductor/168983 -> ciflow/inductor/168983 2025-12-04T08:57:04.2420620Z * [new tag] ciflow/inductor/169006 -> ciflow/inductor/169006 2025-12-04T08:57:04.2421701Z * [new tag] ciflow/inductor/169023 -> ciflow/inductor/169023 2025-12-04T08:57:04.2422802Z * [new tag] ciflow/inductor/169024 -> ciflow/inductor/169024 2025-12-04T08:57:04.2424024Z * [new tag] ciflow/inductor/169025 -> ciflow/inductor/169025 2025-12-04T08:57:04.2425108Z * [new tag] ciflow/inductor/169066 -> ciflow/inductor/169066 2025-12-04T08:57:04.2426192Z * [new tag] ciflow/inductor/169091 -> ciflow/inductor/169091 2025-12-04T08:57:04.2427281Z * [new tag] ciflow/inductor/169102 -> ciflow/inductor/169102 2025-12-04T08:57:04.2428395Z * [new tag] ciflow/inductor/169103 -> ciflow/inductor/169103 2025-12-04T08:57:04.2429536Z * [new tag] ciflow/inductor/169121 -> ciflow/inductor/169121 2025-12-04T08:57:04.2430601Z * [new tag] ciflow/inductor/169134 -> ciflow/inductor/169134 2025-12-04T08:57:04.2431721Z * [new tag] ciflow/inductor/169135 -> ciflow/inductor/169135 2025-12-04T08:57:04.2432821Z * [new tag] ciflow/inductor/169141 -> ciflow/inductor/169141 2025-12-04T08:57:04.2434179Z * [new tag] ciflow/inductor/169151 -> ciflow/inductor/169151 2025-12-04T08:57:04.2435719Z * [new tag] ciflow/inductor/169161 -> ciflow/inductor/169161 2025-12-04T08:57:04.2436836Z * [new tag] ciflow/inductor/169167 -> ciflow/inductor/169167 2025-12-04T08:57:04.2438121Z * [new tag] ciflow/inductor/169177 -> ciflow/inductor/169177 2025-12-04T08:57:04.2439353Z * [new tag] ciflow/inductor/169185 -> ciflow/inductor/169185 2025-12-04T08:57:04.2440450Z * [new tag] ciflow/inductor/169196 -> ciflow/inductor/169196 2025-12-04T08:57:04.2441712Z * [new tag] ciflow/inductor/169200 -> ciflow/inductor/169200 2025-12-04T08:57:04.2442633Z * [new tag] ciflow/inductor/169204 -> ciflow/inductor/169204 2025-12-04T08:57:04.2443863Z * [new tag] ciflow/inductor/169216 -> ciflow/inductor/169216 2025-12-04T08:57:04.2444933Z * [new tag] ciflow/inductor/169219 -> ciflow/inductor/169219 2025-12-04T08:57:04.2446041Z * [new tag] ciflow/inductor/169220 -> ciflow/inductor/169220 2025-12-04T08:57:04.2447261Z * [new tag] ciflow/inductor/169230 -> ciflow/inductor/169230 2025-12-04T08:57:04.2448376Z * [new tag] ciflow/inductor/169242 -> ciflow/inductor/169242 2025-12-04T08:57:04.2449488Z * [new tag] ciflow/inductor/169245 -> ciflow/inductor/169245 2025-12-04T08:57:04.2450732Z * [new tag] ciflow/inductor/169260 -> ciflow/inductor/169260 2025-12-04T08:57:04.2451929Z * [new tag] ciflow/inductor/169282 -> ciflow/inductor/169282 2025-12-04T08:57:04.2453066Z * [new tag] ciflow/inductor/169286 -> ciflow/inductor/169286 2025-12-04T08:57:04.2454338Z * [new tag] ciflow/inductor/169299 -> ciflow/inductor/169299 2025-12-04T08:57:04.2455551Z * [new tag] ciflow/inductor/169304 -> ciflow/inductor/169304 2025-12-04T08:57:04.2457011Z * [new tag] ciflow/inductor/169305 -> ciflow/inductor/169305 2025-12-04T08:57:04.2458134Z * [new tag] ciflow/inductor/169308 -> ciflow/inductor/169308 2025-12-04T08:57:04.2459326Z * [new tag] ciflow/inductor/169319 -> ciflow/inductor/169319 2025-12-04T08:57:04.2460439Z * [new tag] ciflow/inductor/169326 -> ciflow/inductor/169326 2025-12-04T08:57:04.2461554Z * [new tag] ciflow/inductor/169332 -> ciflow/inductor/169332 2025-12-04T08:57:04.2462677Z * [new tag] ciflow/inductor/169333 -> ciflow/inductor/169333 2025-12-04T08:57:04.2463914Z * [new tag] ciflow/inductor/169336 -> ciflow/inductor/169336 2025-12-04T08:57:04.2465031Z * [new tag] ciflow/inductor/169340 -> ciflow/inductor/169340 2025-12-04T08:57:04.2466124Z * [new tag] ciflow/inductor/169341 -> ciflow/inductor/169341 2025-12-04T08:57:04.2467313Z * [new tag] ciflow/inductor/169343 -> ciflow/inductor/169343 2025-12-04T08:57:04.2468406Z * [new tag] ciflow/inductor/169346 -> ciflow/inductor/169346 2025-12-04T08:57:04.2469621Z * [new tag] ciflow/inductor/169348 -> ciflow/inductor/169348 2025-12-04T08:57:04.2470884Z * [new tag] ciflow/inductor/169350 -> ciflow/inductor/169350 2025-12-04T08:57:04.2472007Z * [new tag] ciflow/inductor/169355 -> ciflow/inductor/169355 2025-12-04T08:57:04.2473164Z * [new tag] ciflow/inductor/169370 -> ciflow/inductor/169370 2025-12-04T08:57:04.2474511Z * [new tag] ciflow/inductor/169375 -> ciflow/inductor/169375 2025-12-04T08:57:04.2475619Z * [new tag] ciflow/inductor/169389 -> ciflow/inductor/169389 2025-12-04T08:57:04.2476861Z * [new tag] ciflow/inductor/169391 -> ciflow/inductor/169391 2025-12-04T08:57:04.2477866Z * [new tag] ciflow/inductor/169393 -> ciflow/inductor/169393 2025-12-04T08:57:04.2479008Z * [new tag] ciflow/inductor/169399 -> ciflow/inductor/169399 2025-12-04T08:57:04.2480273Z * [new tag] ciflow/inductor/169400 -> ciflow/inductor/169400 2025-12-04T08:57:04.2481378Z * [new tag] ciflow/inductor/169415 -> ciflow/inductor/169415 2025-12-04T08:57:04.2482497Z * [new tag] ciflow/inductor/169417 -> ciflow/inductor/169417 2025-12-04T08:57:04.2483854Z * [new tag] ciflow/inductor/169418 -> ciflow/inductor/169418 2025-12-04T08:57:04.2485082Z * [new tag] ciflow/inductor/169430 -> ciflow/inductor/169430 2025-12-04T08:57:04.2486078Z * [new tag] ciflow/inductor/169432 -> ciflow/inductor/169432 2025-12-04T08:57:04.2487712Z * [new tag] ciflow/inductor/169436 -> ciflow/inductor/169436 2025-12-04T08:57:04.2488937Z * [new tag] ciflow/inductor/169437 -> ciflow/inductor/169437 2025-12-04T08:57:04.2490162Z * [new tag] ciflow/inductor/169438 -> ciflow/inductor/169438 2025-12-04T08:57:04.2491255Z * [new tag] ciflow/inductor/169441 -> ciflow/inductor/169441 2025-12-04T08:57:04.2492330Z * [new tag] ciflow/inductor/169446 -> ciflow/inductor/169446 2025-12-04T08:57:04.2493648Z * [new tag] ciflow/inductor/169447 -> ciflow/inductor/169447 2025-12-04T08:57:04.2494789Z * [new tag] ciflow/inductor/169452 -> ciflow/inductor/169452 2025-12-04T08:57:04.2495987Z * [new tag] ciflow/inductor/169455 -> ciflow/inductor/169455 2025-12-04T08:57:04.2497094Z * [new tag] ciflow/inductor/169459 -> ciflow/inductor/169459 2025-12-04T08:57:04.2498311Z * [new tag] ciflow/inductor/169463 -> ciflow/inductor/169463 2025-12-04T08:57:04.2499597Z * [new tag] ciflow/inductor/169476 -> ciflow/inductor/169476 2025-12-04T08:57:04.2500717Z * [new tag] ciflow/inductor/169485 -> ciflow/inductor/169485 2025-12-04T08:57:04.2501820Z * [new tag] ciflow/inductor/169493 -> ciflow/inductor/169493 2025-12-04T08:57:04.2502928Z * [new tag] ciflow/inductor/169496 -> ciflow/inductor/169496 2025-12-04T08:57:04.2504101Z * [new tag] ciflow/inductor/169497 -> ciflow/inductor/169497 2025-12-04T08:57:04.2505490Z * [new tag] ciflow/inductor/169503 -> ciflow/inductor/169503 2025-12-04T08:57:04.2506605Z * [new tag] ciflow/inductor/169504 -> ciflow/inductor/169504 2025-12-04T08:57:04.2508058Z * [new tag] ciflow/inductor/169505 -> ciflow/inductor/169505 2025-12-04T08:57:04.2509534Z * [new tag] ciflow/inductor/169508 -> ciflow/inductor/169508 2025-12-04T08:57:04.2510654Z * [new tag] ciflow/inductor/169509 -> ciflow/inductor/169509 2025-12-04T08:57:04.2511767Z * [new tag] ciflow/inductor/169513 -> ciflow/inductor/169513 2025-12-04T08:57:04.2512962Z * [new tag] ciflow/inductor/169514 -> ciflow/inductor/169514 2025-12-04T08:57:04.2514088Z * [new tag] ciflow/inductor/169515 -> ciflow/inductor/169515 2025-12-04T08:57:04.2515192Z * [new tag] ciflow/inductor/169517 -> ciflow/inductor/169517 2025-12-04T08:57:04.2516331Z * [new tag] ciflow/inductor/169519 -> ciflow/inductor/169519 2025-12-04T08:57:04.2517440Z * [new tag] ciflow/inductor/169520 -> ciflow/inductor/169520 2025-12-04T08:57:04.2518591Z * [new tag] ciflow/inductor/169521 -> ciflow/inductor/169521 2025-12-04T08:57:04.2519722Z * [new tag] ciflow/inductor/169524 -> ciflow/inductor/169524 2025-12-04T08:57:04.2520844Z * [new tag] ciflow/inductor/169527 -> ciflow/inductor/169527 2025-12-04T08:57:04.2521962Z * [new tag] ciflow/inductor/169528 -> ciflow/inductor/169528 2025-12-04T08:57:04.2523196Z * [new tag] ciflow/inductor/169532 -> ciflow/inductor/169532 2025-12-04T08:57:04.2524329Z * [new tag] ciflow/inductor/169535 -> ciflow/inductor/169535 2025-12-04T08:57:04.2525444Z * [new tag] ciflow/inductor/169536 -> ciflow/inductor/169536 2025-12-04T08:57:04.2526574Z * [new tag] ciflow/inductor/169547 -> ciflow/inductor/169547 2025-12-04T08:57:04.2528018Z * [new tag] ciflow/inductor/169548 -> ciflow/inductor/169548 2025-12-04T08:57:04.2528863Z * [new tag] ciflow/inductor/169549 -> ciflow/inductor/169549 2025-12-04T08:57:04.2530115Z * [new tag] ciflow/inductor/169551 -> ciflow/inductor/169551 2025-12-04T08:57:04.2531230Z * [new tag] ciflow/inductor/169552 -> ciflow/inductor/169552 2025-12-04T08:57:04.2532325Z * [new tag] ciflow/inductor/169553 -> ciflow/inductor/169553 2025-12-04T08:57:04.2533758Z * [new tag] ciflow/inductor/3b9a386 -> ciflow/inductor/3b9a386 2025-12-04T08:57:04.2535076Z * [new tag] ciflow/inductor/3d4b92b -> ciflow/inductor/3d4b92b 2025-12-04T08:57:04.2536398Z * [new tag] ciflow/inductor/d224ac7 -> ciflow/inductor/d224ac7 2025-12-04T08:57:04.2537690Z * [new tag] ciflow/linux-aarch64/157994 -> ciflow/linux-aarch64/157994 2025-12-04T08:57:04.2538594Z * [new tag] ciflow/linux-aarch64/166075 -> ciflow/linux-aarch64/166075 2025-12-04T08:57:04.2539734Z * [new tag] ciflow/linux-aarch64/166876 -> ciflow/linux-aarch64/166876 2025-12-04T08:57:04.2540837Z * [new tag] ciflow/linux-aarch64/167981 -> ciflow/linux-aarch64/167981 2025-12-04T08:57:04.2542049Z * [new tag] ciflow/mps/166254 -> ciflow/mps/166254 2025-12-04T08:57:04.2543064Z * [new tag] ciflow/mps/169017 -> ciflow/mps/169017 2025-12-04T08:57:04.2544519Z * [new tag] ciflow/mps/169372 -> ciflow/mps/169372 2025-12-04T08:57:04.2545590Z * [new tag] ciflow/mps/169478 -> ciflow/mps/169478 2025-12-04T08:57:04.2546862Z * [new tag] ciflow/op-benchmark/157994 -> ciflow/op-benchmark/157994 2025-12-04T08:57:04.2547852Z * [new tag] ciflow/op-benchmark/166075 -> ciflow/op-benchmark/166075 2025-12-04T08:57:04.2548939Z * [new tag] ciflow/op-benchmark/169544 -> ciflow/op-benchmark/169544 2025-12-04T08:57:04.2550314Z * [new tag] ciflow/periodic-rocm-mi200/165997 -> ciflow/periodic-rocm-mi200/165997 2025-12-04T08:57:04.2551355Z * [new tag] ciflow/periodic-rocm-mi200/166517 -> ciflow/periodic-rocm-mi200/166517 2025-12-04T08:57:04.2552284Z * [new tag] ciflow/periodic-rocm-mi200/169063 -> ciflow/periodic-rocm-mi200/169063 2025-12-04T08:57:04.2553455Z * [new tag] ciflow/periodic-rocm-mi200/169425 -> ciflow/periodic-rocm-mi200/169425 2025-12-04T08:57:04.2554692Z * [new tag] ciflow/periodic-rocm-mi300/166517 -> ciflow/periodic-rocm-mi300/166517 2025-12-04T08:57:04.2555818Z * [new tag] ciflow/periodic-rocm-mi300/169063 -> ciflow/periodic-rocm-mi300/169063 2025-12-04T08:57:04.2556711Z * [new tag] ciflow/periodic-rocm-mi300/169425 -> ciflow/periodic-rocm-mi300/169425 2025-12-04T08:57:04.2558330Z * [new tag] ciflow/periodic/054a2fd -> ciflow/periodic/054a2fd 2025-12-04T08:57:04.2559342Z * [new tag] ciflow/periodic/167207 -> ciflow/periodic/167207 2025-12-04T08:57:04.2560537Z * [new tag] ciflow/periodic/167978 -> ciflow/periodic/167978 2025-12-04T08:57:04.2561520Z * [new tag] ciflow/periodic/168096 -> ciflow/periodic/168096 2025-12-04T08:57:04.2562611Z * [new tag] ciflow/periodic/169286 -> ciflow/periodic/169286 2025-12-04T08:57:04.2563821Z * [new tag] ciflow/periodic/2a6d37d -> ciflow/periodic/2a6d37d 2025-12-04T08:57:04.2565045Z * [new tag] ciflow/periodic/317eeb8 -> ciflow/periodic/317eeb8 2025-12-04T08:57:04.2566220Z * [new tag] ciflow/periodic/3c32 -> ciflow/periodic/3c32 2025-12-04T08:57:04.2567490Z * [new tag] ciflow/periodic/3e98831 -> ciflow/periodic/3e98831 2025-12-04T08:57:04.2569340Z * [new tag] ciflow/periodic/7c648509a7470ace9fb2bae960dd4790f7e943e9 -> ciflow/periodic/7c648509a7470ace9fb2bae960dd4790f7e943e9 2025-12-04T08:57:04.2570621Z * [new tag] ciflow/periodic/94512-point -> ciflow/periodic/94512-point 2025-12-04T08:57:04.2572129Z * [new tag] ciflow/periodic/csl/test87519 -> ciflow/periodic/csl/test87519 2025-12-04T08:57:04.2573440Z * [new tag] ciflow/periodic/csltest88275 -> ciflow/periodic/csltest88275 2025-12-04T08:57:04.2574829Z * [new tag] ciflow/periodic/csltest88761 -> ciflow/periodic/csltest88761 2025-12-04T08:57:04.2576103Z * [new tag] ciflow/periodic/release_1.12 -> ciflow/periodic/release_1.12 2025-12-04T08:57:04.2577503Z * [new tag] ciflow/periodic/release_1.12.0 -> ciflow/periodic/release_1.12.0 2025-12-04T08:57:04.2578824Z * [new tag] ciflow/periodic/sha-ec5b83 -> ciflow/periodic/sha-ec5b83 2025-12-04T08:57:04.2580099Z * [new tag] ciflow/pull/167207 -> ciflow/pull/167207 2025-12-04T08:57:04.2581561Z * [new tag] ciflow/quantization-periodic/169207 -> ciflow/quantization-periodic/169207 2025-12-04T08:57:04.2582750Z * [new tag] ciflow/rocm-mi200/165545 -> ciflow/rocm-mi200/165545 2025-12-04T08:57:04.2583825Z * [new tag] ciflow/rocm-mi200/165997 -> ciflow/rocm-mi200/165997 2025-12-04T08:57:04.2584923Z * [new tag] ciflow/rocm-mi200/168096 -> ciflow/rocm-mi200/168096 2025-12-04T08:57:04.2586063Z * [new tag] ciflow/rocm-mi200/168275 -> ciflow/rocm-mi200/168275 2025-12-04T08:57:04.2586940Z * [new tag] ciflow/rocm-mi200/169063 -> ciflow/rocm-mi200/169063 2025-12-04T08:57:04.2588251Z * [new tag] ciflow/rocm-mi200/169356 -> ciflow/rocm-mi200/169356 2025-12-04T08:57:04.2589319Z * [new tag] ciflow/rocm-mi200/169425 -> ciflow/rocm-mi200/169425 2025-12-04T08:57:04.2590557Z * [new tag] ciflow/rocm-mi300/165545 -> ciflow/rocm-mi300/165545 2025-12-04T08:57:04.2591818Z * [new tag] ciflow/rocm-mi300/167157 -> ciflow/rocm-mi300/167157 2025-12-04T08:57:04.2592824Z * [new tag] ciflow/rocm-mi300/168096 -> ciflow/rocm-mi300/168096 2025-12-04T08:57:04.2593696Z * [new tag] ciflow/rocm-mi300/169063 -> ciflow/rocm-mi300/169063 2025-12-04T08:57:04.2595295Z * [new tag] ciflow/rocm-mi300/169425 -> ciflow/rocm-mi300/169425 2025-12-04T08:57:04.2596639Z * [new tag] ciflow/rocm-mi355/167157 -> ciflow/rocm-mi355/167157 2025-12-04T08:57:04.2597634Z * [new tag] ciflow/rocm-mi355/168275 -> ciflow/rocm-mi355/168275 2025-12-04T08:57:04.2598701Z * [new tag] ciflow/rocm-mi355/169425 -> ciflow/rocm-mi355/169425 2025-12-04T08:57:04.2600060Z * [new tag] ciflow/rocm-navi31/168275 -> ciflow/rocm-navi31/168275 2025-12-04T08:57:04.2600951Z * [new tag] ciflow/rocm-navi31/169425 -> ciflow/rocm-navi31/169425 2025-12-04T08:57:04.2602372Z * [new tag] ciflow/rocm/115316 -> ciflow/rocm/115316 2025-12-04T08:57:04.2603451Z * [new tag] ciflow/rocm/148492 -> ciflow/rocm/148492 2025-12-04T08:57:04.2604305Z * [new tag] ciflow/rocm/160685 -> ciflow/rocm/160685 2025-12-04T08:57:04.2607393Z * [new tag] ciflow/rocm/161607 -> ciflow/rocm/161607 2025-12-04T08:57:04.2608483Z * [new tag] ciflow/rocm/162052 -> ciflow/rocm/162052 2025-12-04T08:57:04.2609479Z * [new tag] ciflow/rocm/165997 -> ciflow/rocm/165997 2025-12-04T08:57:04.2610537Z * [new tag] ciflow/rocm/166165 -> ciflow/rocm/166165 2025-12-04T08:57:04.2611538Z * [new tag] ciflow/rocm/166517 -> ciflow/rocm/166517 2025-12-04T08:57:04.2613105Z * [new tag] ciflow/rocm/167207 -> ciflow/rocm/167207 2025-12-04T08:57:04.2614430Z * [new tag] ciflow/rocm/167536 -> ciflow/rocm/167536 2025-12-04T08:57:04.2615184Z * [new tag] ciflow/rocm/167781 -> ciflow/rocm/167781 2025-12-04T08:57:04.2616649Z * [new tag] ciflow/rocm/167989 -> ciflow/rocm/167989 2025-12-04T08:57:04.2618088Z * [new tag] ciflow/rocm/168073 -> ciflow/rocm/168073 2025-12-04T08:57:04.2619360Z * [new tag] ciflow/rocm/168195 -> ciflow/rocm/168195 2025-12-04T08:57:04.2620496Z * [new tag] ciflow/rocm/168939 -> ciflow/rocm/168939 2025-12-04T08:57:04.2621580Z * [new tag] ciflow/rocm/168971 -> ciflow/rocm/168971 2025-12-04T08:57:04.2622708Z * [new tag] ciflow/rocm/169024 -> ciflow/rocm/169024 2025-12-04T08:57:04.2623967Z * [new tag] ciflow/rocm/169200 -> ciflow/rocm/169200 2025-12-04T08:57:04.2625194Z * [new tag] ciflow/rocm/169216 -> ciflow/rocm/169216 2025-12-04T08:57:04.2626364Z * [new tag] ciflow/rocm/169312 -> ciflow/rocm/169312 2025-12-04T08:57:04.2627586Z * [new tag] ciflow/rocm/169380 -> ciflow/rocm/169380 2025-12-04T08:57:04.2628671Z * [new tag] ciflow/rocm/169427 -> ciflow/rocm/169427 2025-12-04T08:57:04.2629768Z * [new tag] ciflow/rocm/169455 -> ciflow/rocm/169455 2025-12-04T08:57:04.2630872Z * [new tag] ciflow/rocm/169470 -> ciflow/rocm/169470 2025-12-04T08:57:04.2632011Z * [new tag] ciflow/rocm/169471 -> ciflow/rocm/169471 2025-12-04T08:57:04.2633083Z * [new tag] ciflow/rocm/169472 -> ciflow/rocm/169472 2025-12-04T08:57:04.2634262Z * [new tag] ciflow/rocm/169514 -> ciflow/rocm/169514 2025-12-04T08:57:04.2635650Z * [new tag] ciflow/slow/01c7106 -> ciflow/slow/01c7106 2025-12-04T08:57:04.2636839Z * [new tag] ciflow/slow/0577043 -> ciflow/slow/0577043 2025-12-04T08:57:04.2638364Z * [new tag] ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym -> ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym 2025-12-04T08:57:04.2639198Z * [new tag] ciflow/slow/0e81104 -> ciflow/slow/0e81104 2025-12-04T08:57:04.2640377Z * [new tag] ciflow/slow/167207 -> ciflow/slow/167207 2025-12-04T08:57:04.2641441Z * [new tag] ciflow/slow/168050 -> ciflow/slow/168050 2025-12-04T08:57:04.2642605Z * [new tag] ciflow/slow/1732077 -> ciflow/slow/1732077 2025-12-04T08:57:04.2643897Z * [new tag] ciflow/slow/187eb7c -> ciflow/slow/187eb7c 2025-12-04T08:57:04.2645363Z * [new tag] ciflow/slow/1faef89 -> ciflow/slow/1faef89 2025-12-04T08:57:04.2647003Z * [new tag] ciflow/slow/3920ec1 -> ciflow/slow/3920ec1 2025-12-04T08:57:04.2648424Z * [new tag] ciflow/slow/3b7c6b2 -> ciflow/slow/3b7c6b2 2025-12-04T08:57:04.2649728Z * [new tag] ciflow/slow/59a3759 -> ciflow/slow/59a3759 2025-12-04T08:57:04.2650975Z * [new tag] ciflow/slow/70ef0bb -> ciflow/slow/70ef0bb 2025-12-04T08:57:04.2652231Z * [new tag] ciflow/slow/788ff06 -> ciflow/slow/788ff06 2025-12-04T08:57:04.2653876Z * [new tag] ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym -> ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym 2025-12-04T08:57:04.2654969Z * [new tag] ciflow/slow/9d85864 -> ciflow/slow/9d85864 2025-12-04T08:57:04.2656205Z * [new tag] ciflow/slow/9ffad5b -> ciflow/slow/9ffad5b 2025-12-04T08:57:04.2657448Z * [new tag] ciflow/slow/a206e8b -> ciflow/slow/a206e8b 2025-12-04T08:57:04.2658787Z * [new tag] ciflow/slow/a837609 -> ciflow/slow/a837609 2025-12-04T08:57:04.2660012Z * [new tag] ciflow/slow/af841f3 -> ciflow/slow/af841f3 2025-12-04T08:57:04.2662062Z * [new tag] ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym -> ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym 2025-12-04T08:57:04.2663164Z * [new tag] ciflow/torchbench/168175 -> ciflow/torchbench/168175 2025-12-04T08:57:04.2664458Z * [new tag] ciflow/trunk/148492 -> ciflow/trunk/148492 2025-12-04T08:57:04.2665455Z * [new tag] ciflow/trunk/157149 -> ciflow/trunk/157149 2025-12-04T08:57:04.2666543Z * [new tag] ciflow/trunk/157994 -> ciflow/trunk/157994 2025-12-04T08:57:04.2667540Z * [new tag] ciflow/trunk/159718 -> ciflow/trunk/159718 2025-12-04T08:57:04.2668659Z * [new tag] ciflow/trunk/160685 -> ciflow/trunk/160685 2025-12-04T08:57:04.2669656Z * [new tag] ciflow/trunk/160729 -> ciflow/trunk/160729 2025-12-04T08:57:04.2670748Z * [new tag] ciflow/trunk/162275 -> ciflow/trunk/162275 2025-12-04T08:57:04.2671742Z * [new tag] ciflow/trunk/162795 -> ciflow/trunk/162795 2025-12-04T08:57:04.2672830Z * [new tag] ciflow/trunk/163245 -> ciflow/trunk/163245 2025-12-04T08:57:04.2673871Z * [new tag] ciflow/trunk/163942 -> ciflow/trunk/163942 2025-12-04T08:57:04.2674758Z * [new tag] ciflow/trunk/165274 -> ciflow/trunk/165274 2025-12-04T08:57:04.2676353Z * [new tag] ciflow/trunk/165483 -> ciflow/trunk/165483 2025-12-04T08:57:04.2677775Z * [new tag] ciflow/trunk/165728 -> ciflow/trunk/165728 2025-12-04T08:57:04.2679090Z * [new tag] ciflow/trunk/165922 -> ciflow/trunk/165922 2025-12-04T08:57:04.2680309Z * [new tag] ciflow/trunk/166075 -> ciflow/trunk/166075 2025-12-04T08:57:04.2681410Z * [new tag] ciflow/trunk/166165 -> ciflow/trunk/166165 2025-12-04T08:57:04.2682502Z * [new tag] ciflow/trunk/166829 -> ciflow/trunk/166829 2025-12-04T08:57:04.2683777Z * [new tag] ciflow/trunk/166843 -> ciflow/trunk/166843 2025-12-04T08:57:04.2684901Z * [new tag] ciflow/trunk/166876 -> ciflow/trunk/166876 2025-12-04T08:57:04.2686028Z * [new tag] ciflow/trunk/167207 -> ciflow/trunk/167207 2025-12-04T08:57:04.2687105Z * [new tag] ciflow/trunk/167536 -> ciflow/trunk/167536 2025-12-04T08:57:04.2688225Z * [new tag] ciflow/trunk/167552 -> ciflow/trunk/167552 2025-12-04T08:57:04.2689336Z * [new tag] ciflow/trunk/167555 -> ciflow/trunk/167555 2025-12-04T08:57:04.2690462Z * [new tag] ciflow/trunk/167599 -> ciflow/trunk/167599 2025-12-04T08:57:04.2691573Z * [new tag] ciflow/trunk/167659 -> ciflow/trunk/167659 2025-12-04T08:57:04.2692849Z * [new tag] ciflow/trunk/167672 -> ciflow/trunk/167672 2025-12-04T08:57:04.2694115Z * [new tag] ciflow/trunk/167742 -> ciflow/trunk/167742 2025-12-04T08:57:04.2695224Z * [new tag] ciflow/trunk/167781 -> ciflow/trunk/167781 2025-12-04T08:57:04.2696526Z * [new tag] ciflow/trunk/167837 -> ciflow/trunk/167837 2025-12-04T08:57:04.2697623Z * [new tag] ciflow/trunk/167887 -> ciflow/trunk/167887 2025-12-04T08:57:04.2698723Z * [new tag] ciflow/trunk/167978 -> ciflow/trunk/167978 2025-12-04T08:57:04.2699818Z * [new tag] ciflow/trunk/168050 -> ciflow/trunk/168050 2025-12-04T08:57:04.2700967Z * [new tag] ciflow/trunk/168051 -> ciflow/trunk/168051 2025-12-04T08:57:04.2702180Z * [new tag] ciflow/trunk/168096 -> ciflow/trunk/168096 2025-12-04T08:57:04.2703131Z * [new tag] ciflow/trunk/168127 -> ciflow/trunk/168127 2025-12-04T08:57:04.2704306Z * [new tag] ciflow/trunk/168157 -> ciflow/trunk/168157 2025-12-04T08:57:04.2705693Z * [new tag] ciflow/trunk/168175 -> ciflow/trunk/168175 2025-12-04T08:57:04.2706802Z * [new tag] ciflow/trunk/168209 -> ciflow/trunk/168209 2025-12-04T08:57:04.2708049Z * [new tag] ciflow/trunk/168213 -> ciflow/trunk/168213 2025-12-04T08:57:04.2709274Z * [new tag] ciflow/trunk/168226 -> ciflow/trunk/168226 2025-12-04T08:57:04.2710369Z * [new tag] ciflow/trunk/168262 -> ciflow/trunk/168262 2025-12-04T08:57:04.2711491Z * [new tag] ciflow/trunk/168275 -> ciflow/trunk/168275 2025-12-04T08:57:04.2712707Z * [new tag] ciflow/trunk/168328 -> ciflow/trunk/168328 2025-12-04T08:57:04.2713907Z * [new tag] ciflow/trunk/168368 -> ciflow/trunk/168368 2025-12-04T08:57:04.2715259Z * [new tag] ciflow/trunk/168917 -> ciflow/trunk/168917 2025-12-04T08:57:04.2716818Z * [new tag] ciflow/trunk/168933 -> ciflow/trunk/168933 2025-12-04T08:57:04.2718101Z * [new tag] ciflow/trunk/168941 -> ciflow/trunk/168941 2025-12-04T08:57:04.2719170Z * [new tag] ciflow/trunk/168955 -> ciflow/trunk/168955 2025-12-04T08:57:04.2720299Z * [new tag] ciflow/trunk/168980 -> ciflow/trunk/168980 2025-12-04T08:57:04.2721565Z * [new tag] ciflow/trunk/169004 -> ciflow/trunk/169004 2025-12-04T08:57:04.2722720Z * [new tag] ciflow/trunk/169006 -> ciflow/trunk/169006 2025-12-04T08:57:04.2723988Z * [new tag] ciflow/trunk/169023 -> ciflow/trunk/169023 2025-12-04T08:57:04.2725175Z * [new tag] ciflow/trunk/169025 -> ciflow/trunk/169025 2025-12-04T08:57:04.2726295Z * [new tag] ciflow/trunk/169048 -> ciflow/trunk/169048 2025-12-04T08:57:04.2727410Z * [new tag] ciflow/trunk/169066 -> ciflow/trunk/169066 2025-12-04T08:57:04.2728505Z * [new tag] ciflow/trunk/169091 -> ciflow/trunk/169091 2025-12-04T08:57:04.2729607Z * [new tag] ciflow/trunk/169102 -> ciflow/trunk/169102 2025-12-04T08:57:04.2730725Z * [new tag] ciflow/trunk/169103 -> ciflow/trunk/169103 2025-12-04T08:57:04.2732002Z * [new tag] ciflow/trunk/169125 -> ciflow/trunk/169125 2025-12-04T08:57:04.2733276Z * [new tag] ciflow/trunk/169139 -> ciflow/trunk/169139 2025-12-04T08:57:04.2734581Z * [new tag] ciflow/trunk/169148 -> ciflow/trunk/169148 2025-12-04T08:57:04.2735673Z * [new tag] ciflow/trunk/169151 -> ciflow/trunk/169151 2025-12-04T08:57:04.2736828Z * [new tag] ciflow/trunk/169156 -> ciflow/trunk/169156 2025-12-04T08:57:04.2738110Z * [new tag] ciflow/trunk/169176 -> ciflow/trunk/169176 2025-12-04T08:57:04.2739378Z * [new tag] ciflow/trunk/169204 -> ciflow/trunk/169204 2025-12-04T08:57:04.2740502Z * [new tag] ciflow/trunk/169207 -> ciflow/trunk/169207 2025-12-04T08:57:04.2741629Z * [new tag] ciflow/trunk/169211 -> ciflow/trunk/169211 2025-12-04T08:57:04.2742846Z * [new tag] ciflow/trunk/169229 -> ciflow/trunk/169229 2025-12-04T08:57:04.2744193Z * [new tag] ciflow/trunk/169231 -> ciflow/trunk/169231 2025-12-04T08:57:04.2745302Z * [new tag] ciflow/trunk/169260 -> ciflow/trunk/169260 2025-12-04T08:57:04.2746677Z * [new tag] ciflow/trunk/169271 -> ciflow/trunk/169271 2025-12-04T08:57:04.2747925Z * [new tag] ciflow/trunk/169280 -> ciflow/trunk/169280 2025-12-04T08:57:04.2748807Z * [new tag] ciflow/trunk/169281 -> ciflow/trunk/169281 2025-12-04T08:57:04.2750050Z * [new tag] ciflow/trunk/169286 -> ciflow/trunk/169286 2025-12-04T08:57:04.2751315Z * [new tag] ciflow/trunk/169293 -> ciflow/trunk/169293 2025-12-04T08:57:04.2752437Z * [new tag] ciflow/trunk/169296 -> ciflow/trunk/169296 2025-12-04T08:57:04.2753558Z * [new tag] ciflow/trunk/169304 -> ciflow/trunk/169304 2025-12-04T08:57:04.2754715Z * [new tag] ciflow/trunk/169305 -> ciflow/trunk/169305 2025-12-04T08:57:04.2755914Z * [new tag] ciflow/trunk/169312 -> ciflow/trunk/169312 2025-12-04T08:57:04.2757264Z * [new tag] ciflow/trunk/169328 -> ciflow/trunk/169328 2025-12-04T08:57:04.2758383Z * [new tag] ciflow/trunk/169343 -> ciflow/trunk/169343 2025-12-04T08:57:04.2759479Z * [new tag] ciflow/trunk/169355 -> ciflow/trunk/169355 2025-12-04T08:57:04.2760596Z * [new tag] ciflow/trunk/169370 -> ciflow/trunk/169370 2025-12-04T08:57:04.2761833Z * [new tag] ciflow/trunk/169379 -> ciflow/trunk/169379 2025-12-04T08:57:04.2762943Z * [new tag] ciflow/trunk/169380 -> ciflow/trunk/169380 2025-12-04T08:57:04.2764048Z * [new tag] ciflow/trunk/169385 -> ciflow/trunk/169385 2025-12-04T08:57:04.2765127Z * [new tag] ciflow/trunk/169387 -> ciflow/trunk/169387 2025-12-04T08:57:04.2766391Z * [new tag] ciflow/trunk/169410 -> ciflow/trunk/169410 2025-12-04T08:57:04.2767524Z * [new tag] ciflow/trunk/169412 -> ciflow/trunk/169412 2025-12-04T08:57:04.2768650Z * [new tag] ciflow/trunk/169418 -> ciflow/trunk/169418 2025-12-04T08:57:04.2769775Z * [new tag] ciflow/trunk/169423 -> ciflow/trunk/169423 2025-12-04T08:57:04.2770869Z * [new tag] ciflow/trunk/169427 -> ciflow/trunk/169427 2025-12-04T08:57:04.2771983Z * [new tag] ciflow/trunk/169430 -> ciflow/trunk/169430 2025-12-04T08:57:04.2773099Z * [new tag] ciflow/trunk/169437 -> ciflow/trunk/169437 2025-12-04T08:57:04.2774398Z * [new tag] ciflow/trunk/169442 -> ciflow/trunk/169442 2025-12-04T08:57:04.2775502Z * [new tag] ciflow/trunk/169452 -> ciflow/trunk/169452 2025-12-04T08:57:04.2776676Z * [new tag] ciflow/trunk/169454 -> ciflow/trunk/169454 2025-12-04T08:57:04.2777792Z * [new tag] ciflow/trunk/169459 -> ciflow/trunk/169459 2025-12-04T08:57:04.2779042Z * [new tag] ciflow/trunk/169474 -> ciflow/trunk/169474 2025-12-04T08:57:04.2780169Z * [new tag] ciflow/trunk/169475 -> ciflow/trunk/169475 2025-12-04T08:57:04.2781293Z * [new tag] ciflow/trunk/169476 -> ciflow/trunk/169476 2025-12-04T08:57:04.2782527Z * [new tag] ciflow/trunk/169487 -> ciflow/trunk/169487 2025-12-04T08:57:04.2783663Z * [new tag] ciflow/trunk/169497 -> ciflow/trunk/169497 2025-12-04T08:57:04.2784818Z * [new tag] ciflow/trunk/169503 -> ciflow/trunk/169503 2025-12-04T08:57:04.2785947Z * [new tag] ciflow/trunk/169505 -> ciflow/trunk/169505 2025-12-04T08:57:04.2787094Z * [new tag] ciflow/trunk/169507 -> ciflow/trunk/169507 2025-12-04T08:57:04.2788183Z * [new tag] ciflow/trunk/169514 -> ciflow/trunk/169514 2025-12-04T08:57:04.2789291Z * [new tag] ciflow/trunk/169517 -> ciflow/trunk/169517 2025-12-04T08:57:04.2790523Z * [new tag] ciflow/trunk/169519 -> ciflow/trunk/169519 2025-12-04T08:57:04.2791962Z * [new tag] ciflow/trunk/169528 -> ciflow/trunk/169528 2025-12-04T08:57:04.2793056Z * [new tag] ciflow/trunk/169541 -> ciflow/trunk/169541 2025-12-04T08:57:04.2794271Z * [new tag] ciflow/trunk/169555 -> ciflow/trunk/169555 2025-12-04T08:57:04.2795820Z * [new tag] ciflow/unstable/123 -> ciflow/unstable/123 2025-12-04T08:57:04.2797128Z * [new tag] ciflow/vllm/165270 -> ciflow/vllm/165270 2025-12-04T08:57:04.2798203Z * [new tag] ciflow/vllm/165274 -> ciflow/vllm/165274 2025-12-04T08:57:04.2799199Z * [new tag] ciflow/vllm/166494 -> ciflow/vllm/166494 2025-12-04T08:57:04.2800215Z * [new tag] ciflow/vllm/169219 -> ciflow/vllm/169219 2025-12-04T08:57:04.2801296Z * [new tag] ciflow/vllm/169220 -> ciflow/vllm/169220 2025-12-04T08:57:04.2802585Z * [new tag] ciflow/xpu/157994 -> ciflow/xpu/157994 2025-12-04T08:57:04.2803650Z * [new tag] ciflow/xpu/159718 -> ciflow/xpu/159718 2025-12-04T08:57:04.2804477Z * [new tag] ciflow/xpu/161940 -> ciflow/xpu/161940 2025-12-04T08:57:04.2806061Z * [new tag] ciflow/xpu/163251 -> ciflow/xpu/163251 2025-12-04T08:57:04.2807184Z * [new tag] ciflow/xpu/166829 -> ciflow/xpu/166829 2025-12-04T08:57:04.2808171Z * [new tag] ciflow/xpu/166843 -> ciflow/xpu/166843 2025-12-04T08:57:04.2809260Z * [new tag] ciflow/xpu/167972 -> ciflow/xpu/167972 2025-12-04T08:57:04.2810388Z * [new tag] ciflow/xpu/167981 -> ciflow/xpu/167981 2025-12-04T08:57:04.2811442Z * [new tag] ciflow/xpu/168213 -> ciflow/xpu/168213 2025-12-04T08:57:04.2812290Z * [new tag] ciflow/xpu/168262 -> ciflow/xpu/168262 2025-12-04T08:57:04.2813549Z * [new tag] ciflow/xpu/168328 -> ciflow/xpu/168328 2025-12-04T08:57:04.2815096Z * [new tag] ciflow/xpu/168950 -> ciflow/xpu/168950 2025-12-04T08:57:04.2816616Z * [new tag] ciflow/xpu/169039 -> ciflow/xpu/169039 2025-12-04T08:57:04.2817976Z * [new tag] ciflow/xpu/169200 -> ciflow/xpu/169200 2025-12-04T08:57:04.2819120Z * [new tag] ciflow/xpu/169203 -> ciflow/xpu/169203 2025-12-04T08:57:04.2820290Z * [new tag] ciflow/xpu/169229 -> ciflow/xpu/169229 2025-12-04T08:57:04.2821371Z * [new tag] ciflow/xpu/169230 -> ciflow/xpu/169230 2025-12-04T08:57:04.2822469Z * [new tag] ciflow/xpu/169231 -> ciflow/xpu/169231 2025-12-04T08:57:04.2823722Z * [new tag] ciflow/xpu/169241 -> ciflow/xpu/169241 2025-12-04T08:57:04.2824935Z * [new tag] ciflow/xpu/169280 -> ciflow/xpu/169280 2025-12-04T08:57:04.2825960Z * [new tag] ciflow/xpu/169296 -> ciflow/xpu/169296 2025-12-04T08:57:04.2827236Z * [new tag] ciflow/xpu/169353 -> ciflow/xpu/169353 2025-12-04T08:57:04.2828342Z * [new tag] ciflow/xpu/169410 -> ciflow/xpu/169410 2025-12-04T08:57:04.2829511Z * [new tag] ciflow/xpu/169442 -> ciflow/xpu/169442 2025-12-04T08:57:04.2830671Z * [new tag] ciflow/xpu/169555 -> ciflow/xpu/169555 2025-12-04T08:57:04.2831783Z * [new tag] cslpull75 -> cslpull75 2025-12-04T08:57:04.2832851Z * [new tag] cslpull76 -> cslpull76 2025-12-04T08:57:04.2833979Z * [new tag] cslpull77 -> cslpull77 2025-12-04T08:57:04.2835039Z * [new tag] cslpull78 -> cslpull78 2025-12-04T08:57:04.2836536Z * [new tag] cslpull79 -> cslpull79 2025-12-04T08:57:04.2837879Z * [new tag] cslpull80 -> cslpull80 2025-12-04T08:57:04.2839042Z * [new tag] cslpull81 -> cslpull81 2025-12-04T08:57:04.2840230Z * [new tag] cslpull82 -> cslpull82 2025-12-04T08:57:04.2841389Z * [new tag] cslpull83 -> cslpull83 2025-12-04T08:57:04.2842547Z * [new tag] cslpull84 -> cslpull84 2025-12-04T08:57:04.2843800Z * [new tag] cslpull85 -> cslpull85 2025-12-04T08:57:04.2845047Z * [new tag] cslpull86 -> cslpull86 2025-12-04T08:57:04.2846249Z * [new tag] cslpull87 -> cslpull87 2025-12-04T08:57:04.2847419Z * [new tag] cslpull88 -> cslpull88 2025-12-04T08:57:04.2848769Z * [new tag] cslpull89 -> cslpull89 2025-12-04T08:57:04.2849609Z * [new tag] cslpull90 -> cslpull90 2025-12-04T08:57:04.2851224Z * [new tag] cslpull91 -> cslpull91 2025-12-04T08:57:04.2852361Z * [new tag] cslpull92 -> cslpull92 2025-12-04T08:57:04.2853707Z * [new tag] flight_5 -> flight_5 2025-12-04T08:57:04.2855083Z * [new tag] flight_5.1 -> flight_5.1 2025-12-04T08:57:04.2856274Z * [new tag] flight_5.2 -> flight_5.2 2025-12-04T08:57:04.2857527Z * [new tag] flight_5.3 -> flight_5.3 2025-12-04T08:57:04.2858779Z * [new tag] forpull1 -> forpull1 2025-12-04T08:57:04.2860158Z * [new tag] malfet/tag-2ef5611 -> malfet/tag-2ef5611 2025-12-04T08:57:04.2861339Z * [new tag] malfet/tag-317b1a0 -> malfet/tag-317b1a0 2025-12-04T08:57:04.2862497Z * [new tag] malfet/tag-ec6f767 -> malfet/tag-ec6f767 2025-12-04T08:57:04.2863897Z * [new tag] nightly-binary -> nightly-binary 2025-12-04T08:57:04.2865021Z * [new tag] sqzhang_flight4_plus -> sqzhang_flight4_plus 2025-12-04T08:57:04.2866326Z * [new tag] sqzhang_flight_3 -> sqzhang_flight_3 2025-12-04T08:57:04.2867863Z * [new tag] trunk/02d8bd6974cf84b721680d773dbdb1b6f40ce272 -> trunk/02d8bd6974cf84b721680d773dbdb1b6f40ce272 2025-12-04T08:57:04.2869120Z * [new tag] trunk/066997fb38ade71e00d78e9d572e380b5f02bd3e -> trunk/066997fb38ade71e00d78e9d572e380b5f02bd3e 2025-12-04T08:57:04.2870636Z * [new tag] trunk/076e7b19fa1d481ad778d06d2b49ba57d3ce8c88 -> trunk/076e7b19fa1d481ad778d06d2b49ba57d3ce8c88 2025-12-04T08:57:04.2872051Z * [new tag] trunk/07dcc0b83db3211653a38565a24e15acdba75654 -> trunk/07dcc0b83db3211653a38565a24e15acdba75654 2025-12-04T08:57:04.2873326Z * [new tag] trunk/082e96b68dfcd16cab7cfafc4d3d055767dab3eb -> trunk/082e96b68dfcd16cab7cfafc4d3d055767dab3eb 2025-12-04T08:57:04.2874550Z * [new tag] trunk/088048f2fea28ff7d450f65c72419ca45780d30b -> trunk/088048f2fea28ff7d450f65c72419ca45780d30b 2025-12-04T08:57:04.2875689Z * [new tag] trunk/09076941a95c76f4d9ad189d064dfd8baa39e672 -> trunk/09076941a95c76f4d9ad189d064dfd8baa39e672 2025-12-04T08:57:04.2876963Z * [new tag] trunk/0b80a4c62b94402844bf221791c096b0035c6d75 -> trunk/0b80a4c62b94402844bf221791c096b0035c6d75 2025-12-04T08:57:04.2878395Z * [new tag] trunk/0bbbdf1750567a980634ad907a325357ba8ba8f2 -> trunk/0bbbdf1750567a980634ad907a325357ba8ba8f2 2025-12-04T08:57:04.2879657Z * [new tag] trunk/0c281dd78773b2bc17c58ead0e4cd4ac46e775c5 -> trunk/0c281dd78773b2bc17c58ead0e4cd4ac46e775c5 2025-12-04T08:57:04.2881062Z * [new tag] trunk/135f3753c418a6879b1954904184937b67e61688 -> trunk/135f3753c418a6879b1954904184937b67e61688 2025-12-04T08:57:04.2882095Z * [new tag] trunk/15da21026cb13cd20257dc9e96830db108743c10 -> trunk/15da21026cb13cd20257dc9e96830db108743c10 2025-12-04T08:57:04.2883471Z * [new tag] trunk/166efdad2ac827f30fb02504c6017520257f88ec -> trunk/166efdad2ac827f30fb02504c6017520257f88ec 2025-12-04T08:57:04.2884812Z * [new tag] trunk/174272c15fae553d8488140af931f7d8050a313f -> trunk/174272c15fae553d8488140af931f7d8050a313f 2025-12-04T08:57:04.2887192Z * [new tag] trunk/18f3ca08f13b8de61307f5e8cd7d4cccb67e9d11 -> trunk/18f3ca08f13b8de61307f5e8cd7d4cccb67e9d11 2025-12-04T08:57:04.2887574Z * [new tag] trunk/1902eddfe655a15ebcf2c72bd81ade110fdeef63 -> trunk/1902eddfe655a15ebcf2c72bd81ade110fdeef63 2025-12-04T08:57:04.2888696Z * [new tag] trunk/195f92e98d3d66738577f11f22c4b5c8a1c76dd5 -> trunk/195f92e98d3d66738577f11f22c4b5c8a1c76dd5 2025-12-04T08:57:04.2889817Z * [new tag] trunk/1aa13e17de39e3c768ea7aebaad166ce72a06676 -> trunk/1aa13e17de39e3c768ea7aebaad166ce72a06676 2025-12-04T08:57:04.2891040Z * [new tag] trunk/1afe2832f58e24e54a5bfda5a5afa9b96fdea40e -> trunk/1afe2832f58e24e54a5bfda5a5afa9b96fdea40e 2025-12-04T08:57:04.2892258Z * [new tag] trunk/1c87554d74140eaee964ca8b1832cede67f5f520 -> trunk/1c87554d74140eaee964ca8b1832cede67f5f520 2025-12-04T08:57:04.2893583Z * [new tag] trunk/1ccb743b7b5be955f49736c162c4f5004b8a0dd8 -> trunk/1ccb743b7b5be955f49736c162c4f5004b8a0dd8 2025-12-04T08:57:04.2894941Z * [new tag] trunk/1cee47d6ce0a02227185b566593f002dd639ca0c -> trunk/1cee47d6ce0a02227185b566593f002dd639ca0c 2025-12-04T08:57:04.2895875Z * [new tag] trunk/1d21b4df2babe322e5d085ceb6de884eb260a62d -> trunk/1d21b4df2babe322e5d085ceb6de884eb260a62d 2025-12-04T08:57:04.2897254Z * [new tag] trunk/1e34fb2550e4aa650314f7a6d9f6daf4da7478a8 -> trunk/1e34fb2550e4aa650314f7a6d9f6daf4da7478a8 2025-12-04T08:57:04.2898549Z * [new tag] trunk/1e526fb5b1d93bfc70691c5c3955fdffc1b7b7de -> trunk/1e526fb5b1d93bfc70691c5c3955fdffc1b7b7de 2025-12-04T08:57:04.2899785Z * [new tag] trunk/1ee32a8b1f554a312d79bad01ded24f38cd95543 -> trunk/1ee32a8b1f554a312d79bad01ded24f38cd95543 2025-12-04T08:57:04.2901046Z * [new tag] trunk/201e2c4117eb9744594dad6a5c18213d7b4705d7 -> trunk/201e2c4117eb9744594dad6a5c18213d7b4705d7 2025-12-04T08:57:04.2902279Z * [new tag] trunk/2353a0f60eb4b4cb6675907a7fa9fbedc1c02e7f -> trunk/2353a0f60eb4b4cb6675907a7fa9fbedc1c02e7f 2025-12-04T08:57:04.2903663Z * [new tag] trunk/285779b1621cf9f073a062b0889a642d200308d9 -> trunk/285779b1621cf9f073a062b0889a642d200308d9 2025-12-04T08:57:04.2904865Z * [new tag] trunk/2887faaec6295d081580d09fce161201826c6d87 -> trunk/2887faaec6295d081580d09fce161201826c6d87 2025-12-04T08:57:04.2906357Z * [new tag] trunk/296e67c92635443c67b11c0ae1bd045f03ebb7bc -> trunk/296e67c92635443c67b11c0ae1bd045f03ebb7bc 2025-12-04T08:57:04.2907669Z * [new tag] trunk/29856679769b3dede478767e2fe6cfb51197cb25 -> trunk/29856679769b3dede478767e2fe6cfb51197cb25 2025-12-04T08:57:04.2909008Z * [new tag] trunk/29e5455a4740c326ab187c7aa7b5ef98034ea563 -> trunk/29e5455a4740c326ab187c7aa7b5ef98034ea563 2025-12-04T08:57:04.2910272Z * [new tag] trunk/2ac3ef882afb23136adc188975f0a8802fc68adf -> trunk/2ac3ef882afb23136adc188975f0a8802fc68adf 2025-12-04T08:57:04.2911218Z * [new tag] trunk/2bec68e73b64715354af076ad309335f943e36cd -> trunk/2bec68e73b64715354af076ad309335f943e36cd 2025-12-04T08:57:04.2912590Z * [new tag] trunk/2c87367e6f88662cd5cedbd1537748b7948c38e1 -> trunk/2c87367e6f88662cd5cedbd1537748b7948c38e1 2025-12-04T08:57:04.2913998Z * [new tag] trunk/2d1f78fe3ec13820f136a2e0336da12a25f41708 -> trunk/2d1f78fe3ec13820f136a2e0336da12a25f41708 2025-12-04T08:57:04.2914998Z * [new tag] trunk/2df6058f116a65722a0e03073402feb242572d35 -> trunk/2df6058f116a65722a0e03073402feb242572d35 2025-12-04T08:57:04.2916403Z * [new tag] trunk/2e0c2e170fe658c440775c8e5c44228aafcc47ec -> trunk/2e0c2e170fe658c440775c8e5c44228aafcc47ec 2025-12-04T08:57:04.2917759Z * [new tag] trunk/2f9b7dad7b5419b063bd0f2e204de192720ebb94 -> trunk/2f9b7dad7b5419b063bd0f2e204de192720ebb94 2025-12-04T08:57:04.2918991Z * [new tag] trunk/305168768a95d69c444df5cd334bb774edfe06f1 -> trunk/305168768a95d69c444df5cd334bb774edfe06f1 2025-12-04T08:57:04.2920204Z * [new tag] trunk/31fc12773026e8e00f054dd79ad9b2491e693b48 -> trunk/31fc12773026e8e00f054dd79ad9b2491e693b48 2025-12-04T08:57:04.2921460Z * [new tag] trunk/320de0c6b0a3e7c6d2693ea5c28d5d0156ba7991 -> trunk/320de0c6b0a3e7c6d2693ea5c28d5d0156ba7991 2025-12-04T08:57:04.2922752Z * [new tag] trunk/3418bd29475dff06695045fcdf93e7d0dac67da8 -> trunk/3418bd29475dff06695045fcdf93e7d0dac67da8 2025-12-04T08:57:04.2924163Z * [new tag] trunk/34a98608afa0cb5b48f0d6d30432fdd0a2614ddf -> trunk/34a98608afa0cb5b48f0d6d30432fdd0a2614ddf 2025-12-04T08:57:04.2925435Z * [new tag] trunk/35b7a9a26c5923d98aebaa41a031dae21788a9ee -> trunk/35b7a9a26c5923d98aebaa41a031dae21788a9ee 2025-12-04T08:57:04.2926685Z * [new tag] trunk/39d07dbf03a911bdd45d1af78d8638dc92074938 -> trunk/39d07dbf03a911bdd45d1af78d8638dc92074938 2025-12-04T08:57:04.2927854Z * [new tag] trunk/3cd98b4205ada151042cc7ff097a82d4a4b18725 -> trunk/3cd98b4205ada151042cc7ff097a82d4a4b18725 2025-12-04T08:57:04.2929134Z * [new tag] trunk/3d35fd20a78ff4d016fa80f4e5fad37191d7bcae -> trunk/3d35fd20a78ff4d016fa80f4e5fad37191d7bcae 2025-12-04T08:57:04.2930428Z * [new tag] trunk/409a5fee945c46a3edaf5df162812f201bfd7b2f -> trunk/409a5fee945c46a3edaf5df162812f201bfd7b2f 2025-12-04T08:57:04.2931725Z * [new tag] trunk/42e9005cda22da3f1c559c3649218cebd671027c -> trunk/42e9005cda22da3f1c559c3649218cebd671027c 2025-12-04T08:57:04.2932995Z * [new tag] trunk/43b94713bbf340d3c124fde02d0f73add4021247 -> trunk/43b94713bbf340d3c124fde02d0f73add4021247 2025-12-04T08:57:04.2934741Z * [new tag] trunk/44ac69388a4a5eb463dbd2a13f00d1e3b924566c -> trunk/44ac69388a4a5eb463dbd2a13f00d1e3b924566c 2025-12-04T08:57:04.2935967Z * [new tag] trunk/45d14e2497292be06ad36eaa1aaaf7c630a2586a -> trunk/45d14e2497292be06ad36eaa1aaaf7c630a2586a 2025-12-04T08:57:04.2936979Z * [new tag] trunk/45d310ad84854dff730c0b12e577d7998d978686 -> trunk/45d310ad84854dff730c0b12e577d7998d978686 2025-12-04T08:57:04.2938642Z * [new tag] trunk/47b28ddf7bd74b50fa93b307a7d3b183a6d77f54 -> trunk/47b28ddf7bd74b50fa93b307a7d3b183a6d77f54 2025-12-04T08:57:04.2939585Z * [new tag] trunk/481e5ab336275bd3acd5fa8a611b05b4469012af -> trunk/481e5ab336275bd3acd5fa8a611b05b4469012af 2025-12-04T08:57:04.2940955Z * [new tag] trunk/491731647f6b8a9345dcfb3bc9416aea254a7d96 -> trunk/491731647f6b8a9345dcfb3bc9416aea254a7d96 2025-12-04T08:57:04.2942201Z * [new tag] trunk/49a04d26088acc17d948ddd66920f3e16371e873 -> trunk/49a04d26088acc17d948ddd66920f3e16371e873 2025-12-04T08:57:04.2943427Z * [new tag] trunk/4bebc827c47d2f1f0fa1a417a5201a97aef3d985 -> trunk/4bebc827c47d2f1f0fa1a417a5201a97aef3d985 2025-12-04T08:57:04.2944426Z * [new tag] trunk/4c246677784c6a14bc2dbb9ff8773ef0a3a3222f -> trunk/4c246677784c6a14bc2dbb9ff8773ef0a3a3222f 2025-12-04T08:57:04.2945951Z * [new tag] trunk/4cfb47ff548b6d996641058cf04a70e311a4c3aa -> trunk/4cfb47ff548b6d996641058cf04a70e311a4c3aa 2025-12-04T08:57:04.2947179Z * [new tag] trunk/4e0061c1aa52f606dda8cfab0bd7591e588faf2c -> trunk/4e0061c1aa52f606dda8cfab0bd7591e588faf2c 2025-12-04T08:57:04.2948913Z * [new tag] trunk/4fefb8e7e942386ffac764a41b232241f82bea3a -> trunk/4fefb8e7e942386ffac764a41b232241f82bea3a 2025-12-04T08:57:04.2949924Z * [new tag] trunk/503b2640023521f5a35cd9a52fc8033d73a95d0d -> trunk/503b2640023521f5a35cd9a52fc8033d73a95d0d 2025-12-04T08:57:04.2951307Z * [new tag] trunk/518c2b1b3dab9a2ef2849e04b3bc2f20c1c41db9 -> trunk/518c2b1b3dab9a2ef2849e04b3bc2f20c1c41db9 2025-12-04T08:57:04.2952564Z * [new tag] trunk/5191b2fa68ba19960912bfd7fd721c79d76bb1f3 -> trunk/5191b2fa68ba19960912bfd7fd721c79d76bb1f3 2025-12-04T08:57:04.2953882Z * [new tag] trunk/52ac0f0dc4acacd219f1317fbc28ec631c01e07a -> trunk/52ac0f0dc4acacd219f1317fbc28ec631c01e07a 2025-12-04T08:57:04.2955170Z * [new tag] trunk/539ba711b029de9f191070f4f0d12f18f5b7f292 -> trunk/539ba711b029de9f191070f4f0d12f18f5b7f292 2025-12-04T08:57:04.2956436Z * [new tag] trunk/556375b55deebebbc56cb7aef81f4d52f031ba28 -> trunk/556375b55deebebbc56cb7aef81f4d52f031ba28 2025-12-04T08:57:04.2957796Z * [new tag] trunk/55c4ab554845481d0a69a3811937575fe8bb1a66 -> trunk/55c4ab554845481d0a69a3811937575fe8bb1a66 2025-12-04T08:57:04.2959061Z * [new tag] trunk/5634469fda9e5d98869c82c7d03bb08914245f96 -> trunk/5634469fda9e5d98869c82c7d03bb08914245f96 2025-12-04T08:57:04.2960014Z * [new tag] trunk/5778f6ff894686a975a9a23645178ae4c87ad5dc -> trunk/5778f6ff894686a975a9a23645178ae4c87ad5dc 2025-12-04T08:57:04.2961425Z * [new tag] trunk/587d63a3e07de5dc91065f9ef70bcacda9989068 -> trunk/587d63a3e07de5dc91065f9ef70bcacda9989068 2025-12-04T08:57:04.2962715Z * [new tag] trunk/597930f6b568852356ca9795dac76f9e4653adbd -> trunk/597930f6b568852356ca9795dac76f9e4653adbd 2025-12-04T08:57:04.2963663Z * [new tag] trunk/597df3a4e2a67b9fdbe1a89b2f4d74f822274db6 -> trunk/597df3a4e2a67b9fdbe1a89b2f4d74f822274db6 2025-12-04T08:57:04.2965156Z * [new tag] trunk/59abd50e931f4efb21b053f7a2911f5d8a49d883 -> trunk/59abd50e931f4efb21b053f7a2911f5d8a49d883 2025-12-04T08:57:04.2966409Z * [new tag] trunk/5a607febc04c3a2b5824c75f3f60307867439a2c -> trunk/5a607febc04c3a2b5824c75f3f60307867439a2c 2025-12-04T08:57:04.2967696Z * [new tag] trunk/5bf1cdf4755c54ef462b44cb8041b0a57311556b -> trunk/5bf1cdf4755c54ef462b44cb8041b0a57311556b 2025-12-04T08:57:04.2968827Z * [new tag] trunk/5f0030ba63d334d7e8c93a09e41403b89e4c573c -> trunk/5f0030ba63d334d7e8c93a09e41403b89e4c573c 2025-12-04T08:57:04.2970087Z * [new tag] trunk/5f21d27e71268464d362a96c9ac09ea475f7f202 -> trunk/5f21d27e71268464d362a96c9ac09ea475f7f202 2025-12-04T08:57:04.2971386Z * [new tag] trunk/5fafc13038c9988d9ac21fa793fbd5890604b447 -> trunk/5fafc13038c9988d9ac21fa793fbd5890604b447 2025-12-04T08:57:04.2972712Z * [new tag] trunk/61be54a31dc09b59d99b62176fb935aee0b924ef -> trunk/61be54a31dc09b59d99b62176fb935aee0b924ef 2025-12-04T08:57:04.2974020Z * [new tag] trunk/62d3ccd71484ed6a760d909b41487101bbc65719 -> trunk/62d3ccd71484ed6a760d909b41487101bbc65719 2025-12-04T08:57:04.2975275Z * [new tag] trunk/641cdb68ae27668eb441d0e49c87a0602c120c2b -> trunk/641cdb68ae27668eb441d0e49c87a0602c120c2b 2025-12-04T08:57:04.2976500Z * [new tag] trunk/65c4620d6bb0c6029f69762c22b91dda2294da9a -> trunk/65c4620d6bb0c6029f69762c22b91dda2294da9a 2025-12-04T08:57:04.2977841Z * [new tag] trunk/66004b993744b4106bf8afaba71f3c228a804206 -> trunk/66004b993744b4106bf8afaba71f3c228a804206 2025-12-04T08:57:04.2979122Z * [new tag] trunk/6658a04c7ca67acb64512341342e7b3ee13ee386 -> trunk/6658a04c7ca67acb64512341342e7b3ee13ee386 2025-12-04T08:57:04.2980333Z * [new tag] trunk/6864e309092a71f8ab0ca6a4dc7f8a4073fd31c4 -> trunk/6864e309092a71f8ab0ca6a4dc7f8a4073fd31c4 2025-12-04T08:57:04.2981718Z * [new tag] trunk/6c261c6cb07892c90ca19ed51c9705b1659a3f7d -> trunk/6c261c6cb07892c90ca19ed51c9705b1659a3f7d 2025-12-04T08:57:04.2982757Z * [new tag] trunk/6c8b6a043f1628188b6396b3a2a6e000ca68362b -> trunk/6c8b6a043f1628188b6396b3a2a6e000ca68362b 2025-12-04T08:57:04.2984090Z * [new tag] trunk/6ceb4a32f92ae67ce5d7d97931d17401ebf5ffa5 -> trunk/6ceb4a32f92ae67ce5d7d97931d17401ebf5ffa5 2025-12-04T08:57:04.2985351Z * [new tag] trunk/6e404e9b7d6f5fb0de86aa73888c3038248c17f8 -> trunk/6e404e9b7d6f5fb0de86aa73888c3038248c17f8 2025-12-04T08:57:04.2986670Z * [new tag] trunk/6ec30b490aee1db6bcdc7340abddef25784f08ec -> trunk/6ec30b490aee1db6bcdc7340abddef25784f08ec 2025-12-04T08:57:04.2987819Z * [new tag] trunk/6f2783a6c08e1db34275ff25176ffe9aebc30a71 -> trunk/6f2783a6c08e1db34275ff25176ffe9aebc30a71 2025-12-04T08:57:04.2989137Z * [new tag] trunk/6f53fefeb90ad3281119b5cfc4aa9ffd8a066e3d -> trunk/6f53fefeb90ad3281119b5cfc4aa9ffd8a066e3d 2025-12-04T08:57:04.2990433Z * [new tag] trunk/6f7dcf51e46d0c880db1a2f5c70de57adb576f4a -> trunk/6f7dcf51e46d0c880db1a2f5c70de57adb576f4a 2025-12-04T08:57:04.2991713Z * [new tag] trunk/6ff831180d2fa436c7f1c1af3adac641fce9d60e -> trunk/6ff831180d2fa436c7f1c1af3adac641fce9d60e 2025-12-04T08:57:04.2992844Z * [new tag] trunk/70076464a63ab218a7ceefb0e76ccd7131deb8f8 -> trunk/70076464a63ab218a7ceefb0e76ccd7131deb8f8 2025-12-04T08:57:04.2994127Z * [new tag] trunk/70d797a5fc109b20a517646fcaa819477cd0d485 -> trunk/70d797a5fc109b20a517646fcaa819477cd0d485 2025-12-04T08:57:04.2995260Z * [new tag] trunk/7348cb355ff0a6f79cd4871215aea72185748734 -> trunk/7348cb355ff0a6f79cd4871215aea72185748734 2025-12-04T08:57:04.2996689Z * [new tag] trunk/74fe26a1ebe32931783569f2e762e3c2c974901f -> trunk/74fe26a1ebe32931783569f2e762e3c2c974901f 2025-12-04T08:57:04.2997999Z * [new tag] trunk/76aeb8c7e0f795b3fddca134cbea9a69da3ee696 -> trunk/76aeb8c7e0f795b3fddca134cbea9a69da3ee696 2025-12-04T08:57:04.2999271Z * [new tag] trunk/7741edd4ed665f3988052e260863efb508d61a03 -> trunk/7741edd4ed665f3988052e260863efb508d61a03 2025-12-04T08:57:04.3000543Z * [new tag] trunk/78adb3b3df41b45d2368b67226d2f864b78939a6 -> trunk/78adb3b3df41b45d2368b67226d2f864b78939a6 2025-12-04T08:57:04.3001861Z * [new tag] trunk/79d7b178225e5ed24d4e1db74e5abbff848f5fb7 -> trunk/79d7b178225e5ed24d4e1db74e5abbff848f5fb7 2025-12-04T08:57:04.3002801Z * [new tag] trunk/7a1e316115fc6996b3f2336822ba5d5f6179f0c3 -> trunk/7a1e316115fc6996b3f2336822ba5d5f6179f0c3 2025-12-04T08:57:04.3004172Z * [new tag] trunk/7a41b66367c38d0af3e8a90f7be48d6b281e7bca -> trunk/7a41b66367c38d0af3e8a90f7be48d6b281e7bca 2025-12-04T08:57:04.3005719Z * [new tag] trunk/7b7af390ea8541c611d1ce2018a6934188fc197b -> trunk/7b7af390ea8541c611d1ce2018a6934188fc197b 2025-12-04T08:57:04.3007002Z * [new tag] trunk/7ba4680f3755a560af81aa0f688791e367aa3609 -> trunk/7ba4680f3755a560af81aa0f688791e367aa3609 2025-12-04T08:57:04.3008292Z * [new tag] trunk/7bc2a66ded06a0b2549aa51d807edc5dc3e73d1b -> trunk/7bc2a66ded06a0b2549aa51d807edc5dc3e73d1b 2025-12-04T08:57:04.3009560Z * [new tag] trunk/7c648509a7470ace9fb2bae960dd4790f7e943e9 -> trunk/7c648509a7470ace9fb2bae960dd4790f7e943e9 2025-12-04T08:57:04.3010543Z * [new tag] trunk/7cbc2d034cecd21ab5c9707d0a9c525c17143fb8 -> trunk/7cbc2d034cecd21ab5c9707d0a9c525c17143fb8 2025-12-04T08:57:04.3011932Z * [new tag] trunk/7d1bbaf4ba301ea3fba6f3c7bc02d58f6417aaed -> trunk/7d1bbaf4ba301ea3fba6f3c7bc02d58f6417aaed 2025-12-04T08:57:04.3013252Z * [new tag] trunk/7d2a33e4ebf60b217a3cd77feae19231eb996fc8 -> trunk/7d2a33e4ebf60b217a3cd77feae19231eb996fc8 2025-12-04T08:57:04.3014698Z * [new tag] trunk/7eb625920054b1126a7d2d99818aaa188c6ba95e -> trunk/7eb625920054b1126a7d2d99818aaa188c6ba95e 2025-12-04T08:57:04.3016135Z * [new tag] trunk/7f55ba19c456a3d6cc443dd9edb6bb7cca677ead -> trunk/7f55ba19c456a3d6cc443dd9edb6bb7cca677ead 2025-12-04T08:57:04.3017357Z * [new tag] trunk/81af382128efa094d8702e18f2c133760904c718 -> trunk/81af382128efa094d8702e18f2c133760904c718 2025-12-04T08:57:04.3018805Z * [new tag] trunk/84149583d483e9c973c9a0feda70e4f3964947b0 -> trunk/84149583d483e9c973c9a0feda70e4f3964947b0 2025-12-04T08:57:04.3020255Z * [new tag] trunk/85a315917efe82c24306be805c584ec044951c75 -> trunk/85a315917efe82c24306be805c584ec044951c75 2025-12-04T08:57:04.3021486Z * [new tag] trunk/87329491c82a5f8c1cc4ec11d8f55a5de2551ece -> trunk/87329491c82a5f8c1cc4ec11d8f55a5de2551ece 2025-12-04T08:57:04.3022489Z * [new tag] trunk/892640e25aeefa8007c5af837214b4502b6b62a6 -> trunk/892640e25aeefa8007c5af837214b4502b6b62a6 2025-12-04T08:57:04.3024123Z * [new tag] trunk/89e3bbcb5b5321dc8b9520b4d5a8ee60cea1d0b4 -> trunk/89e3bbcb5b5321dc8b9520b4d5a8ee60cea1d0b4 2025-12-04T08:57:04.3025345Z * [new tag] trunk/8c73bbbb02159223c0c97d268a0a74cb78158a1c -> trunk/8c73bbbb02159223c0c97d268a0a74cb78158a1c 2025-12-04T08:57:04.3026592Z * [new tag] trunk/8d56e98c8db988a22cb2dfaeefb30bc7d2a3cc43 -> trunk/8d56e98c8db988a22cb2dfaeefb30bc7d2a3cc43 2025-12-04T08:57:04.3027854Z * [new tag] trunk/8d9dd9603e5ee26c01007f0cd4f018e584840922 -> trunk/8d9dd9603e5ee26c01007f0cd4f018e584840922 2025-12-04T08:57:04.3029232Z * [new tag] trunk/8ef0c0b02b062d75e7c9be2594914a3e784d23ca -> trunk/8ef0c0b02b062d75e7c9be2594914a3e784d23ca 2025-12-04T08:57:04.3030535Z * [new tag] trunk/90b27e7e8352cde97d32ddad24740ef819633f38 -> trunk/90b27e7e8352cde97d32ddad24740ef819633f38 2025-12-04T08:57:04.3031513Z * [new tag] trunk/90f0139e64b2951815d524b6a373bed20c4fbf90 -> trunk/90f0139e64b2951815d524b6a373bed20c4fbf90 2025-12-04T08:57:04.3032774Z * [new tag] trunk/93d0d6838c56af59b0dba794e6aa08f0c1c7799c -> trunk/93d0d6838c56af59b0dba794e6aa08f0c1c7799c 2025-12-04T08:57:04.3034082Z * [new tag] trunk/94ca8d5f1e81fea3ae488650a0fb6795049a9f87 -> trunk/94ca8d5f1e81fea3ae488650a0fb6795049a9f87 2025-12-04T08:57:04.3035246Z * [new tag] trunk/9844fbeadd5cebdf1281d6fbf79164139c352693 -> trunk/9844fbeadd5cebdf1281d6fbf79164139c352693 2025-12-04T08:57:04.3036626Z * [new tag] trunk/99024dec888ec1e50b546822a32b6fb2f35e5eaa -> trunk/99024dec888ec1e50b546822a32b6fb2f35e5eaa 2025-12-04T08:57:04.3037884Z * [new tag] trunk/9a296e640fc88aa44d275b48cd9cc30c573b169d -> trunk/9a296e640fc88aa44d275b48cd9cc30c573b169d 2025-12-04T08:57:04.3039136Z * [new tag] trunk/9b3e34d8589b29f7b4e7fab6f78711b7ca6e4639 -> trunk/9b3e34d8589b29f7b4e7fab6f78711b7ca6e4639 2025-12-04T08:57:04.3040409Z * [new tag] trunk/9cd055e547e9b67a5f9827f8999c38d7eda1bcb8 -> trunk/9cd055e547e9b67a5f9827f8999c38d7eda1bcb8 2025-12-04T08:57:04.3041648Z * [new tag] trunk/9f0df5686cb4ada94f94620acba2e3c3f363b11d -> trunk/9f0df5686cb4ada94f94620acba2e3c3f363b11d 2025-12-04T08:57:04.3042936Z * [new tag] trunk/9f7fceb887d0cfa0326a59b887821c63ff11340a -> trunk/9f7fceb887d0cfa0326a59b887821c63ff11340a 2025-12-04T08:57:04.3044455Z * [new tag] trunk/9f8ef8855d3078d70f7b782540ff2aaf158d6742 -> trunk/9f8ef8855d3078d70f7b782540ff2aaf158d6742 2025-12-04T08:57:04.3045912Z * [new tag] trunk/9fb52efc797b47a1f425a03aa5e47b866d8b1098 -> trunk/9fb52efc797b47a1f425a03aa5e47b866d8b1098 2025-12-04T08:57:04.3047135Z * [new tag] trunk/9ff4a2ebc5762d46c73e46b1b523d7ff349fedfa -> trunk/9ff4a2ebc5762d46c73e46b1b523d7ff349fedfa 2025-12-04T08:57:04.3048368Z * [new tag] trunk/a0f3937b94422354538ebbd47202d5b0e8a3fd0d -> trunk/a0f3937b94422354538ebbd47202d5b0e8a3fd0d 2025-12-04T08:57:04.3049903Z * [new tag] trunk/a15066c28b3145e6edbfc88359d0411d14cfc70c -> trunk/a15066c28b3145e6edbfc88359d0411d14cfc70c 2025-12-04T08:57:04.3051233Z * [new tag] trunk/a20f775e82564d2a9979221ed7f3b8d7cf54ce90 -> trunk/a20f775e82564d2a9979221ed7f3b8d7cf54ce90 2025-12-04T08:57:04.3052380Z * [new tag] trunk/a2973fb00ec002dd4b6bbf07385f066efb259b8c -> trunk/a2973fb00ec002dd4b6bbf07385f066efb259b8c 2025-12-04T08:57:04.3053788Z * [new tag] trunk/a7dc6dab9ad911259d4801c502907e531594db45 -> trunk/a7dc6dab9ad911259d4801c502907e531594db45 2025-12-04T08:57:04.3055403Z * [new tag] trunk/a951a9cee65c01660bbc6e6fded90ecb10fa6109 -> trunk/a951a9cee65c01660bbc6e6fded90ecb10fa6109 2025-12-04T08:57:04.3056657Z * [new tag] trunk/abfa1a6d65c7c159e35c72c25979b9da4971689e -> trunk/abfa1a6d65c7c159e35c72c25979b9da4971689e 2025-12-04T08:57:04.3057917Z * [new tag] trunk/ae3a2395bf66151078e2d201716f7d63ce1c6f3e -> trunk/ae3a2395bf66151078e2d201716f7d63ce1c6f3e 2025-12-04T08:57:04.3058926Z * [new tag] trunk/afdff7f0325080dedac44d080cb5a3b0e65e6c5e -> trunk/afdff7f0325080dedac44d080cb5a3b0e65e6c5e 2025-12-04T08:57:04.3060217Z * [new tag] trunk/b1aed4e7a72c03a38f44543aaea0dae2e9b76d48 -> trunk/b1aed4e7a72c03a38f44543aaea0dae2e9b76d48 2025-12-04T08:57:04.3061579Z * [new tag] trunk/b1decff555cd50e2123c8c6e25cc0d447c411f62 -> trunk/b1decff555cd50e2123c8c6e25cc0d447c411f62 2025-12-04T08:57:04.3062895Z * [new tag] trunk/b2b6b034c9fd08672c40e63ef243556ad4c49bd2 -> trunk/b2b6b034c9fd08672c40e63ef243556ad4c49bd2 2025-12-04T08:57:04.3064175Z * [new tag] trunk/b39813b4a04931682b0491adba2138d01d716d99 -> trunk/b39813b4a04931682b0491adba2138d01d716d99 2025-12-04T08:57:04.3065520Z * [new tag] trunk/b3a7edb2311367974cc7cd764cfb11a5d6758b24 -> trunk/b3a7edb2311367974cc7cd764cfb11a5d6758b24 2025-12-04T08:57:04.3066812Z * [new tag] trunk/b4cc1329c86acaef6d42c1fac7169b8d870ab0d7 -> trunk/b4cc1329c86acaef6d42c1fac7169b8d870ab0d7 2025-12-04T08:57:04.3068167Z * [new tag] trunk/b555c39217f765759954a4f9f9bd1e9b87bed11a -> trunk/b555c39217f765759954a4f9f9bd1e9b87bed11a 2025-12-04T08:57:04.3069438Z * [new tag] trunk/b6b6c80379388b7f9932c3e6a0f9907bf430e417 -> trunk/b6b6c80379388b7f9932c3e6a0f9907bf430e417 2025-12-04T08:57:04.3070766Z * [new tag] trunk/b6b6d912df0b6f4082f8e50b18bd1de1dd7325f4 -> trunk/b6b6d912df0b6f4082f8e50b18bd1de1dd7325f4 2025-12-04T08:57:04.3072094Z * [new tag] trunk/b7d60685f8cbc939b68a20871e90db67e729329b -> trunk/b7d60685f8cbc939b68a20871e90db67e729329b 2025-12-04T08:57:04.3073393Z * [new tag] trunk/b7f6b9a4fc6259f7af068f31868b3119bb1bac3e -> trunk/b7f6b9a4fc6259f7af068f31868b3119bb1bac3e 2025-12-04T08:57:04.3074688Z * [new tag] trunk/b8c4ba3593761e7b2a3ebd86f040fb07b47c02cf -> trunk/b8c4ba3593761e7b2a3ebd86f040fb07b47c02cf 2025-12-04T08:57:04.3075940Z * [new tag] trunk/b9c8f3a4884befb965ff42620ce44a71b04887f5 -> trunk/b9c8f3a4884befb965ff42620ce44a71b04887f5 2025-12-04T08:57:04.3077202Z * [new tag] trunk/ba1412546f3082c0958c077acc2025e4dbc33f1f -> trunk/ba1412546f3082c0958c077acc2025e4dbc33f1f 2025-12-04T08:57:04.3078620Z * [new tag] trunk/bac403c0b38c63bdbcc0c31f1c2b0bc0260f610f -> trunk/bac403c0b38c63bdbcc0c31f1c2b0bc0260f610f 2025-12-04T08:57:04.3079910Z * [new tag] trunk/bb3034198b459401fabeab254e1b99f0115046e2 -> trunk/bb3034198b459401fabeab254e1b99f0115046e2 2025-12-04T08:57:04.3081167Z * [new tag] trunk/bc39b2b3bc7a6e19a42e62bd576974035086fe55 -> trunk/bc39b2b3bc7a6e19a42e62bd576974035086fe55 2025-12-04T08:57:04.3082578Z * [new tag] trunk/bc43d5b297f207a11d83d77ddf0152bdaabe15a8 -> trunk/bc43d5b297f207a11d83d77ddf0152bdaabe15a8 2025-12-04T08:57:04.3083999Z * [new tag] trunk/bc6a4863c7246a6493d16d4ea6eee71ec07c6a09 -> trunk/bc6a4863c7246a6493d16d4ea6eee71ec07c6a09 2025-12-04T08:57:04.3085311Z * [new tag] trunk/bea4912944defdbcb8b061800caab6cbbbd01df5 -> trunk/bea4912944defdbcb8b061800caab6cbbbd01df5 2025-12-04T08:57:04.3086758Z * [new tag] trunk/c04e2c656f48d82d1521b867bbbf03967b9b7564 -> trunk/c04e2c656f48d82d1521b867bbbf03967b9b7564 2025-12-04T08:57:04.3088020Z * [new tag] trunk/c0660bcee27e7d7731634e274576a7081882bede -> trunk/c0660bcee27e7d7731634e274576a7081882bede 2025-12-04T08:57:04.3089314Z * [new tag] trunk/c178ed43d3d99cbefe84fbfb21d6f282b20d62ac -> trunk/c178ed43d3d99cbefe84fbfb21d6f282b20d62ac 2025-12-04T08:57:04.3090616Z * [new tag] trunk/c55b1e8f61d041ee436d697449eb028931d574fb -> trunk/c55b1e8f61d041ee436d697449eb028931d574fb 2025-12-04T08:57:04.3091746Z * [new tag] trunk/c6ae7579fe12fe75f1a8f7043a494c90567273f1 -> trunk/c6ae7579fe12fe75f1a8f7043a494c90567273f1 2025-12-04T08:57:04.3093499Z * [new tag] trunk/c8210e7d94bad5ae21ac389fa4ba8a463c76c4d0 -> trunk/c8210e7d94bad5ae21ac389fa4ba8a463c76c4d0 2025-12-04T08:57:04.3095051Z * [new tag] trunk/cc0853af42122f8185321f542616f4474e717f09 -> trunk/cc0853af42122f8185321f542616f4474e717f09 2025-12-04T08:57:04.3096217Z * [new tag] trunk/cddec6562eabfa390d014fa3741a5659cf9c94c9 -> trunk/cddec6562eabfa390d014fa3741a5659cf9c94c9 2025-12-04T08:57:04.3097558Z * [new tag] trunk/ce5e7e3bf1f4b69a4f4f93d288ba75b906df492a -> trunk/ce5e7e3bf1f4b69a4f4f93d288ba75b906df492a 2025-12-04T08:57:04.3098948Z * [new tag] trunk/d038b0130ec7c20ebcac219301292fd8e98a1ace -> trunk/d038b0130ec7c20ebcac219301292fd8e98a1ace 2025-12-04T08:57:04.3100644Z * [new tag] trunk/d16447dacaf2420ea175f0c275c75da951f57d39 -> trunk/d16447dacaf2420ea175f0c275c75da951f57d39 2025-12-04T08:57:04.3101912Z * [new tag] trunk/d19f1e8cab6810bb2e99141f9976665954c67a50 -> trunk/d19f1e8cab6810bb2e99141f9976665954c67a50 2025-12-04T08:57:04.3103233Z * [new tag] trunk/d1c9f03b2a5af4104721712f8cdffe9b4f340c01 -> trunk/d1c9f03b2a5af4104721712f8cdffe9b4f340c01 2025-12-04T08:57:04.3104978Z * [new tag] trunk/d40f4950f2b7f7aa380a22fe0f6166e71680fbcf -> trunk/d40f4950f2b7f7aa380a22fe0f6166e71680fbcf 2025-12-04T08:57:04.3107916Z * [new tag] trunk/d5038950bacfe36bbf24a47a455fe76901deb8e8 -> trunk/d5038950bacfe36bbf24a47a455fe76901deb8e8 2025-12-04T08:57:04.3109050Z * [new tag] trunk/d54ff42903c2ae0533931ff11d23b35f875bdb3d -> trunk/d54ff42903c2ae0533931ff11d23b35f875bdb3d 2025-12-04T08:57:04.3110450Z * [new tag] trunk/d76697633a2d2b9cced1ae21161849b33bfe7e47 -> trunk/d76697633a2d2b9cced1ae21161849b33bfe7e47 2025-12-04T08:57:04.3111743Z * [new tag] trunk/d78f52b199c547106d4cd9d2856dd0805c118bf1 -> trunk/d78f52b199c547106d4cd9d2856dd0805c118bf1 2025-12-04T08:57:04.3113036Z * [new tag] trunk/d8fd5c6eed28e5004150691d048a3f6785e19a8e -> trunk/d8fd5c6eed28e5004150691d048a3f6785e19a8e 2025-12-04T08:57:04.3114340Z * [new tag] trunk/d900f5e86745dec76713f4b0ef07005ef36b2f5a -> trunk/d900f5e86745dec76713f4b0ef07005ef36b2f5a 2025-12-04T08:57:04.3115614Z * [new tag] trunk/d973dc6b87d763859fe1c5bd1287e3b6b1c49d1b -> trunk/d973dc6b87d763859fe1c5bd1287e3b6b1c49d1b 2025-12-04T08:57:04.3116914Z * [new tag] trunk/d998c03304cb6ede76e1ed535b4ddeb6c2bf40ec -> trunk/d998c03304cb6ede76e1ed535b4ddeb6c2bf40ec 2025-12-04T08:57:04.3118241Z * [new tag] trunk/d9cb8a70833101dbbe16b99520cfbdd70d0a87bf -> trunk/d9cb8a70833101dbbe16b99520cfbdd70d0a87bf 2025-12-04T08:57:04.3119529Z * [new tag] trunk/d9d5e91b43f70eb8637af55db6856d49be391ffd -> trunk/d9d5e91b43f70eb8637af55db6856d49be391ffd 2025-12-04T08:57:04.3120832Z * [new tag] trunk/dd18a75336a4fbd7497955cc5665904724fce889 -> trunk/dd18a75336a4fbd7497955cc5665904724fce889 2025-12-04T08:57:04.3122259Z * [new tag] trunk/ded9bcd61a059bf723e6e84689552962b480ea77 -> trunk/ded9bcd61a059bf723e6e84689552962b480ea77 2025-12-04T08:57:04.3123690Z * [new tag] trunk/dfbd3714d15c37a7b83b322a6b60f997fc00f50c -> trunk/dfbd3714d15c37a7b83b322a6b60f997fc00f50c 2025-12-04T08:57:04.3125003Z * [new tag] trunk/e115f9f4e4b039f8e9a642aaa2bd8254a920541b -> trunk/e115f9f4e4b039f8e9a642aaa2bd8254a920541b 2025-12-04T08:57:04.3126133Z * [new tag] trunk/e3f24fd73ad74c6e7176687986436956c7c18235 -> trunk/e3f24fd73ad74c6e7176687986436956c7c18235 2025-12-04T08:57:04.3127544Z * [new tag] trunk/e7d24d3ff93d1503ba63860b7057438ad93f918e -> trunk/e7d24d3ff93d1503ba63860b7057438ad93f918e 2025-12-04T08:57:04.3128856Z * [new tag] trunk/ea7035f462a0d2830865ee86c832bd101e1427fc -> trunk/ea7035f462a0d2830865ee86c832bd101e1427fc 2025-12-04T08:57:04.3130120Z * [new tag] trunk/eabb7ad2128580ef674446027b95bcf4e21e8df3 -> trunk/eabb7ad2128580ef674446027b95bcf4e21e8df3 2025-12-04T08:57:04.3131411Z * [new tag] trunk/eb5c63652a33da42e7018c23df5f20a3eb4c6ccf -> trunk/eb5c63652a33da42e7018c23df5f20a3eb4c6ccf 2025-12-04T08:57:04.3132704Z * [new tag] trunk/ec2c71f5c85021b8938cdafadce24c15a36fd93e -> trunk/ec2c71f5c85021b8938cdafadce24c15a36fd93e 2025-12-04T08:57:04.3134184Z * [new tag] trunk/ecbcc3f6bf327856b435b259ac63cc2f328c4b4e -> trunk/ecbcc3f6bf327856b435b259ac63cc2f328c4b4e 2025-12-04T08:57:04.3135738Z * [new tag] trunk/ee87bbe876c42575e961b32a0827d76bc9782ca2 -> trunk/ee87bbe876c42575e961b32a0827d76bc9782ca2 2025-12-04T08:57:04.3137034Z * [new tag] trunk/ef019d1d431c4c5a95b594cb90d40a50cd00f5e4 -> trunk/ef019d1d431c4c5a95b594cb90d40a50cd00f5e4 2025-12-04T08:57:04.3138420Z * [new tag] trunk/ef8ecc13830a86c4b231f1aad9aba7851db61b53 -> trunk/ef8ecc13830a86c4b231f1aad9aba7851db61b53 2025-12-04T08:57:04.3139653Z * [new tag] trunk/f1076f5510920044912247b1abb8760cb820f598 -> trunk/f1076f5510920044912247b1abb8760cb820f598 2025-12-04T08:57:04.3141000Z * [new tag] trunk/f2d6a75a00a1d648ca9a0abc6a33e14c3dea6c40 -> trunk/f2d6a75a00a1d648ca9a0abc6a33e14c3dea6c40 2025-12-04T08:57:04.3142268Z * [new tag] trunk/f47dd0ddef1359e5b43e4b962412f67b30ecde56 -> trunk/f47dd0ddef1359e5b43e4b962412f67b30ecde56 2025-12-04T08:57:04.3143570Z * [new tag] trunk/f49d32dfa4730dcfb1b60eeeb369b5889da983c8 -> trunk/f49d32dfa4730dcfb1b60eeeb369b5889da983c8 2025-12-04T08:57:04.3144641Z * [new tag] trunk/f4dedf78fc30fd4b93975787ca6074ee89db9467 -> trunk/f4dedf78fc30fd4b93975787ca6074ee89db9467 2025-12-04T08:57:04.3146028Z * [new tag] trunk/f7c0d03819ebed05c4038f095d66d1b8c54aca17 -> trunk/f7c0d03819ebed05c4038f095d66d1b8c54aca17 2025-12-04T08:57:04.3147310Z * [new tag] trunk/f7e1bd80a063e17453c361837ba6ea2570920a73 -> trunk/f7e1bd80a063e17453c361837ba6ea2570920a73 2025-12-04T08:57:04.3148509Z * [new tag] trunk/f9bd6c53624c7c0ea3772de78498326e84c2f0e7 -> trunk/f9bd6c53624c7c0ea3772de78498326e84c2f0e7 2025-12-04T08:57:04.3149820Z * [new tag] trunk/fb5be221a46b51bfc9509013b0d85bc5a9d4f15b -> trunk/fb5be221a46b51bfc9509013b0d85bc5a9d4f15b 2025-12-04T08:57:04.3151097Z * [new tag] trunk/fdf863d5e1de3b2688c9511e96876e34581dbfd7 -> trunk/fdf863d5e1de3b2688c9511e96876e34581dbfd7 2025-12-04T08:57:04.3152796Z * [new tag] trunk/fe0e65adfc0e7ca6e5f57e6ea8b16bd5cc967307 -> trunk/fe0e65adfc0e7ca6e5f57e6ea8b16bd5cc967307 2025-12-04T08:57:04.3154244Z * [new tag] trunk/fec710bf89173f5355468a7ce1afe9157c3d9009 -> trunk/fec710bf89173f5355468a7ce1afe9157c3d9009 2025-12-04T08:57:04.3155633Z * [new tag] trunk/ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 -> trunk/ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T08:57:04.3156880Z * [new tag] v0.1.1 -> v0.1.1 2025-12-04T08:57:04.3158000Z * [new tag] v0.1.10 -> v0.1.10 2025-12-04T08:57:04.3158937Z * [new tag] v0.1.11 -> v0.1.11 2025-12-04T08:57:04.3160231Z * [new tag] v0.1.12 -> v0.1.12 2025-12-04T08:57:04.3161410Z * [new tag] v0.1.2 -> v0.1.2 2025-12-04T08:57:04.3162590Z * [new tag] v0.1.3 -> v0.1.3 2025-12-04T08:57:04.3163856Z * [new tag] v0.1.4 -> v0.1.4 2025-12-04T08:57:04.3165210Z * [new tag] v0.1.5 -> v0.1.5 2025-12-04T08:57:04.3166439Z * [new tag] v0.1.6 -> v0.1.6 2025-12-04T08:57:04.3167597Z * [new tag] v0.1.7 -> v0.1.7 2025-12-04T08:57:04.3168732Z * [new tag] v0.1.8 -> v0.1.8 2025-12-04T08:57:04.3169861Z * [new tag] v0.1.9 -> v0.1.9 2025-12-04T08:57:04.3171098Z * [new tag] v0.2.0 -> v0.2.0 2025-12-04T08:57:04.3172407Z * [new tag] v0.3.0 -> v0.3.0 2025-12-04T08:57:04.3173715Z * [new tag] v0.3.1 -> v0.3.1 2025-12-04T08:57:04.3174935Z * [new tag] v0.4.0 -> v0.4.0 2025-12-04T08:57:04.3176100Z * [new tag] v0.4.1 -> v0.4.1 2025-12-04T08:57:04.3177815Z * [new tag] v1.0.0 -> v1.0.0 2025-12-04T08:57:04.3179089Z * [new tag] v1.0.0a0 -> v1.0.0a0 2025-12-04T08:57:04.3180277Z * [new tag] v1.0.1 -> v1.0.1 2025-12-04T08:57:04.3181478Z * [new tag] v1.0rc0 -> v1.0rc0 2025-12-04T08:57:04.3182521Z * [new tag] v1.0rc1 -> v1.0rc1 2025-12-04T08:57:04.3183675Z * [new tag] v1.1.0 -> v1.1.0 2025-12-04T08:57:04.3184897Z * [new tag] v1.1.0a0 -> v1.1.0a0 2025-12-04T08:57:04.3186254Z * [new tag] v1.10.0 -> v1.10.0 2025-12-04T08:57:04.3187661Z * [new tag] v1.10.0-rc1 -> v1.10.0-rc1 2025-12-04T08:57:04.3188891Z * [new tag] v1.10.0-rc2 -> v1.10.0-rc2 2025-12-04T08:57:04.3189869Z * [new tag] v1.10.0-rc3 -> v1.10.0-rc3 2025-12-04T08:57:04.3191149Z * [new tag] v1.10.1 -> v1.10.1 2025-12-04T08:57:04.3192205Z * [new tag] v1.10.1-rc1 -> v1.10.1-rc1 2025-12-04T08:57:04.3193266Z * [new tag] v1.10.2 -> v1.10.2 2025-12-04T08:57:04.3194351Z * [new tag] v1.10.2-rc1 -> v1.10.2-rc1 2025-12-04T08:57:04.3195596Z * [new tag] v1.11.0 -> v1.11.0 2025-12-04T08:57:04.3196854Z * [new tag] v1.11.0-rc1 -> v1.11.0-rc1 2025-12-04T08:57:04.3198112Z * [new tag] v1.11.0-rc2 -> v1.11.0-rc2 2025-12-04T08:57:04.3199366Z * [new tag] v1.11.0-rc3 -> v1.11.0-rc3 2025-12-04T08:57:04.3200719Z * [new tag] v1.11.0-rc4 -> v1.11.0-rc4 2025-12-04T08:57:04.3201978Z * [new tag] v1.11.0-rc5 -> v1.11.0-rc5 2025-12-04T08:57:04.3202951Z * [new tag] v1.11.0-rc6 -> v1.11.0-rc6 2025-12-04T08:57:04.3204005Z * [new tag] v1.11.0-rc7 -> v1.11.0-rc7 2025-12-04T08:57:04.3205479Z * [new tag] v1.12.0 -> v1.12.0 2025-12-04T08:57:04.3206719Z * [new tag] v1.12.0-rc1 -> v1.12.0-rc1 2025-12-04T08:57:04.3208133Z * [new tag] v1.12.0-rc2 -> v1.12.0-rc2 2025-12-04T08:57:04.3209141Z * [new tag] v1.12.0-rc3 -> v1.12.0-rc3 2025-12-04T08:57:04.3210498Z * [new tag] v1.12.0-rc4 -> v1.12.0-rc4 2025-12-04T08:57:04.3211700Z * [new tag] v1.12.0-rc5 -> v1.12.0-rc5 2025-12-04T08:57:04.3212964Z * [new tag] v1.12.0-rc6 -> v1.12.0-rc6 2025-12-04T08:57:04.3214191Z * [new tag] v1.12.0-rc7 -> v1.12.0-rc7 2025-12-04T08:57:04.3215169Z * [new tag] v1.12.0-rc8 -> v1.12.0-rc8 2025-12-04T08:57:04.3216246Z * [new tag] v1.12.1 -> v1.12.1 2025-12-04T08:57:04.3217612Z * [new tag] v1.12.1-rc1 -> v1.12.1-rc1 2025-12-04T08:57:04.3218857Z * [new tag] v1.12.1-rc2 -> v1.12.1-rc2 2025-12-04T08:57:04.3220134Z * [new tag] v1.12.1-rc3 -> v1.12.1-rc3 2025-12-04T08:57:04.3221343Z * [new tag] v1.12.1-rc4 -> v1.12.1-rc4 2025-12-04T08:57:04.3222327Z * [new tag] v1.12.1-rc5 -> v1.12.1-rc5 2025-12-04T08:57:04.3223628Z * [new tag] v1.13.0 -> v1.13.0 2025-12-04T08:57:04.3224871Z * [new tag] v1.13.0-rc1 -> v1.13.0-rc1 2025-12-04T08:57:04.3226035Z * [new tag] v1.13.0-rc2 -> v1.13.0-rc2 2025-12-04T08:57:04.3227235Z * [new tag] v1.13.0-rc3 -> v1.13.0-rc3 2025-12-04T08:57:04.3228600Z * [new tag] v1.13.0-rc4 -> v1.13.0-rc4 2025-12-04T08:57:04.3229565Z * [new tag] v1.13.0-rc5 -> v1.13.0-rc5 2025-12-04T08:57:04.3230641Z * [new tag] v1.13.0-rc6 -> v1.13.0-rc6 2025-12-04T08:57:04.3231884Z * [new tag] v1.13.1 -> v1.13.1 2025-12-04T08:57:04.3232931Z * [new tag] v1.13.1-rc1 -> v1.13.1-rc1 2025-12-04T08:57:04.3234063Z * [new tag] v1.2.0 -> v1.2.0 2025-12-04T08:57:04.3235265Z * [new tag] v1.2.0a0 -> v1.2.0a0 2025-12-04T08:57:04.3236473Z * [new tag] v1.3.0 -> v1.3.0 2025-12-04T08:57:04.3237702Z * [new tag] v1.3.0a0 -> v1.3.0a0 2025-12-04T08:57:04.3238680Z * [new tag] v1.3.1 -> v1.3.1 2025-12-04T08:57:04.3239914Z * [new tag] v1.4.0 -> v1.4.0 2025-12-04T08:57:04.3241139Z * [new tag] v1.4.0a0 -> v1.4.0a0 2025-12-04T08:57:04.3242105Z * [new tag] v1.4.1 -> v1.4.1 2025-12-04T08:57:04.3243405Z * [new tag] v1.5.0 -> v1.5.0 2025-12-04T08:57:04.3244676Z * [new tag] v1.5.0-rc1 -> v1.5.0-rc1 2025-12-04T08:57:04.3245887Z * [new tag] v1.5.0-rc2 -> v1.5.0-rc2 2025-12-04T08:57:04.3247179Z * [new tag] v1.5.0-rc3 -> v1.5.0-rc3 2025-12-04T08:57:04.3248306Z * [new tag] v1.5.0-rc4 -> v1.5.0-rc4 2025-12-04T08:57:04.3249278Z * [new tag] v1.5.0-rc5 -> v1.5.0-rc5 2025-12-04T08:57:04.3250602Z * [new tag] v1.5.1 -> v1.5.1 2025-12-04T08:57:04.3251650Z * [new tag] v1.5.1-rc1 -> v1.5.1-rc1 2025-12-04T08:57:04.3252606Z * [new tag] v1.6.0 -> v1.6.0 2025-12-04T08:57:04.3254528Z * [new tag] v1.6.0-rc1 -> v1.6.0-rc1 2025-12-04T08:57:04.3255824Z * [new tag] v1.6.0-rc2 -> v1.6.0-rc2 2025-12-04T08:57:04.3257107Z * [new tag] v1.6.0-rc3 -> v1.6.0-rc3 2025-12-04T08:57:04.3258371Z * [new tag] v1.6.0-rc4 -> v1.6.0-rc4 2025-12-04T08:57:04.3259567Z * [new tag] v1.6.0-rc5 -> v1.6.0-rc5 2025-12-04T08:57:04.3260691Z * [new tag] v1.6.0-rc6 -> v1.6.0-rc6 2025-12-04T08:57:04.3261692Z * [new tag] v1.6.0-rc7 -> v1.6.0-rc7 2025-12-04T08:57:04.3262952Z * [new tag] v1.7.0 -> v1.7.0 2025-12-04T08:57:04.3264202Z * [new tag] v1.7.0-rc1 -> v1.7.0-rc1 2025-12-04T08:57:04.3265526Z * [new tag] v1.7.0-rc2 -> v1.7.0-rc2 2025-12-04T08:57:04.3266691Z * [new tag] v1.7.0-rc3 -> v1.7.0-rc3 2025-12-04T08:57:04.3267730Z * [new tag] v1.7.0-rc4 -> v1.7.0-rc4 2025-12-04T08:57:04.3268976Z * [new tag] v1.7.1 -> v1.7.1 2025-12-04T08:57:04.3270290Z * [new tag] v1.7.1-rc1 -> v1.7.1-rc1 2025-12-04T08:57:04.3271546Z * [new tag] v1.7.1-rc2 -> v1.7.1-rc2 2025-12-04T08:57:04.3272627Z * [new tag] v1.7.1-rc3 -> v1.7.1-rc3 2025-12-04T08:57:04.3273892Z * [new tag] v1.8.0 -> v1.8.0 2025-12-04T08:57:04.3274960Z * [new tag] v1.8.0-rc1 -> v1.8.0-rc1 2025-12-04T08:57:04.3276179Z * [new tag] v1.8.0-rc2 -> v1.8.0-rc2 2025-12-04T08:57:04.3277407Z * [new tag] v1.8.0-rc3 -> v1.8.0-rc3 2025-12-04T08:57:04.3278574Z * [new tag] v1.8.0-rc4 -> v1.8.0-rc4 2025-12-04T08:57:04.3279549Z * [new tag] v1.8.0-rc5 -> v1.8.0-rc5 2025-12-04T08:57:04.3280601Z * [new tag] v1.8.1 -> v1.8.1 2025-12-04T08:57:04.3281860Z * [new tag] v1.8.1-rc1 -> v1.8.1-rc1 2025-12-04T08:57:04.3282845Z * [new tag] v1.8.1-rc2 -> v1.8.1-rc2 2025-12-04T08:57:04.3283896Z * [new tag] v1.8.1-rc3 -> v1.8.1-rc3 2025-12-04T08:57:04.3285472Z * [new tag] v1.8.2 -> v1.8.2 2025-12-04T08:57:04.3286436Z * [new tag] v1.8.2-rc1 -> v1.8.2-rc1 2025-12-04T08:57:04.3287706Z * [new tag] v1.9.0 -> v1.9.0 2025-12-04T08:57:04.3288949Z * [new tag] v1.9.0-rc1 -> v1.9.0-rc1 2025-12-04T08:57:04.3290236Z * [new tag] v1.9.0-rc2 -> v1.9.0-rc2 2025-12-04T08:57:04.3291478Z * [new tag] v1.9.0-rc3 -> v1.9.0-rc3 2025-12-04T08:57:04.3292540Z * [new tag] v1.9.0-rc4 -> v1.9.0-rc4 2025-12-04T08:57:04.3293961Z * [new tag] v1.9.1 -> v1.9.1 2025-12-04T08:57:04.3295390Z * [new tag] v1.9.1-rc1 -> v1.9.1-rc1 2025-12-04T08:57:04.3296374Z * [new tag] v1.9.1-rc2 -> v1.9.1-rc2 2025-12-04T08:57:04.3297654Z * [new tag] v2.0.0 -> v2.0.0 2025-12-04T08:57:04.3298846Z * [new tag] v2.0.0-rc1 -> v2.0.0-rc1 2025-12-04T08:57:04.3300116Z * [new tag] v2.0.0-rc2 -> v2.0.0-rc2 2025-12-04T08:57:04.3301361Z * [new tag] v2.0.0-rc3 -> v2.0.0-rc3 2025-12-04T08:57:04.3302583Z * [new tag] v2.0.0-rc4 -> v2.0.0-rc4 2025-12-04T08:57:04.3303842Z * [new tag] v2.0.0-rc5 -> v2.0.0-rc5 2025-12-04T08:57:04.3304761Z * [new tag] v2.0.0-rc6 -> v2.0.0-rc6 2025-12-04T08:57:04.3306426Z * [new tag] v2.0.1 -> v2.0.1 2025-12-04T08:57:04.3307859Z * [new tag] v2.0.1-rc1 -> v2.0.1-rc1 2025-12-04T08:57:04.3308621Z * [new tag] v2.0.1-rc2 -> v2.0.1-rc2 2025-12-04T08:57:04.3309954Z * [new tag] v2.0.1-rc3 -> v2.0.1-rc3 2025-12-04T08:57:04.3310988Z * [new tag] v2.0.1-rc4 -> v2.0.1-rc4 2025-12-04T08:57:04.3312602Z * [new tag] v2.1.0 -> v2.1.0 2025-12-04T08:57:04.3313776Z * [new tag] v2.1.0-rc1 -> v2.1.0-rc1 2025-12-04T08:57:04.3315073Z * [new tag] v2.1.0-rc2 -> v2.1.0-rc2 2025-12-04T08:57:04.3316423Z * [new tag] v2.1.0-rc3 -> v2.1.0-rc3 2025-12-04T08:57:04.3317778Z * [new tag] v2.1.0-rc4 -> v2.1.0-rc4 2025-12-04T08:57:04.3319008Z * [new tag] v2.1.0-rc5 -> v2.1.0-rc5 2025-12-04T08:57:04.3319991Z * [new tag] v2.1.0-rc6 -> v2.1.0-rc6 2025-12-04T08:57:04.3321278Z * [new tag] v2.1.1 -> v2.1.1 2025-12-04T08:57:04.3322626Z * [new tag] v2.1.1-rc1 -> v2.1.1-rc1 2025-12-04T08:57:04.3323951Z * [new tag] v2.1.1-rc2 -> v2.1.1-rc2 2025-12-04T08:57:04.3325278Z * [new tag] v2.1.1-rc3 -> v2.1.1-rc3 2025-12-04T08:57:04.3326536Z * [new tag] v2.1.1-rc4 -> v2.1.1-rc4 2025-12-04T08:57:04.3327744Z * [new tag] v2.1.1-rc5 -> v2.1.1-rc5 2025-12-04T08:57:04.3328712Z * [new tag] v2.1.1-rc6 -> v2.1.1-rc6 2025-12-04T08:57:04.3329947Z * [new tag] v2.1.2 -> v2.1.2 2025-12-04T08:57:04.3331235Z * [new tag] v2.1.2-rc1 -> v2.1.2-rc1 2025-12-04T08:57:04.3332909Z * [new tag] v2.1.2-rc2 -> v2.1.2-rc2 2025-12-04T08:57:04.3334175Z * [new tag] v2.1.2-rc3 -> v2.1.2-rc3 2025-12-04T08:57:04.3335437Z * [new tag] v2.2.0 -> v2.2.0 2025-12-04T08:57:04.3336685Z * [new tag] v2.2.0-rc1 -> v2.2.0-rc1 2025-12-04T08:57:04.3337882Z * [new tag] v2.2.0-rc2 -> v2.2.0-rc2 2025-12-04T08:57:04.3339020Z * [new tag] v2.2.0-rc3 -> v2.2.0-rc3 2025-12-04T08:57:04.3340226Z * [new tag] v2.2.0-rc4 -> v2.2.0-rc4 2025-12-04T08:57:04.3341431Z * [new tag] v2.2.0-rc5 -> v2.2.0-rc5 2025-12-04T08:57:04.3342733Z * [new tag] v2.2.0-rc6 -> v2.2.0-rc6 2025-12-04T08:57:04.3343708Z * [new tag] v2.2.0-rc7 -> v2.2.0-rc7 2025-12-04T08:57:04.3344787Z * [new tag] v2.2.0-rc8 -> v2.2.0-rc8 2025-12-04T08:57:04.3346061Z * [new tag] v2.2.1 -> v2.2.1 2025-12-04T08:57:04.3347314Z * [new tag] v2.2.1-rc1 -> v2.2.1-rc1 2025-12-04T08:57:04.3348289Z * [new tag] v2.2.1-rc2 -> v2.2.1-rc2 2025-12-04T08:57:04.3349363Z * [new tag] v2.2.1-rc3 -> v2.2.1-rc3 2025-12-04T08:57:04.3350324Z * [new tag] v2.2.2 -> v2.2.2 2025-12-04T08:57:04.3351741Z * [new tag] v2.2.2-rc1 -> v2.2.2-rc1 2025-12-04T08:57:04.3352715Z * [new tag] v2.2.2-rc2 -> v2.2.2-rc2 2025-12-04T08:57:04.3353796Z * [new tag] v2.2.2-rc3 -> v2.2.2-rc3 2025-12-04T08:57:04.3355037Z * [new tag] v2.3.0 -> v2.3.0 2025-12-04T08:57:04.3356212Z * [new tag] v2.3.0-rc1 -> v2.3.0-rc1 2025-12-04T08:57:04.3357616Z * [new tag] v2.3.0-rc10 -> v2.3.0-rc10 2025-12-04T08:57:04.3358820Z * [new tag] v2.3.0-rc11 -> v2.3.0-rc11 2025-12-04T08:57:04.3359956Z * [new tag] v2.3.0-rc12 -> v2.3.0-rc12 2025-12-04T08:57:04.3361135Z * [new tag] v2.3.0-rc2 -> v2.3.0-rc2 2025-12-04T08:57:04.3362378Z * [new tag] v2.3.0-rc3 -> v2.3.0-rc3 2025-12-04T08:57:04.3363592Z * [new tag] v2.3.0-rc4 -> v2.3.0-rc4 2025-12-04T08:57:04.3364876Z * [new tag] v2.3.0-rc5 -> v2.3.0-rc5 2025-12-04T08:57:04.3365853Z * [new tag] v2.3.0-rc6 -> v2.3.0-rc6 2025-12-04T08:57:04.3367129Z * [new tag] v2.3.0-rc7 -> v2.3.0-rc7 2025-12-04T08:57:04.3368350Z * [new tag] v2.3.0-rc8 -> v2.3.0-rc8 2025-12-04T08:57:04.3369329Z * [new tag] v2.3.0-rc9 -> v2.3.0-rc9 2025-12-04T08:57:04.3370432Z * [new tag] v2.3.1 -> v2.3.1 2025-12-04T08:57:04.3371774Z * [new tag] v2.3.1-rc1 -> v2.3.1-rc1 2025-12-04T08:57:04.3373106Z * [new tag] v2.3.1-rc2 -> v2.3.1-rc2 2025-12-04T08:57:04.3374495Z * [new tag] v2.3.1-rc3 -> v2.3.1-rc3 2025-12-04T08:57:04.3375727Z * [new tag] v2.4.0 -> v2.4.0 2025-12-04T08:57:04.3376962Z * [new tag] v2.4.0-rc1 -> v2.4.0-rc1 2025-12-04T08:57:04.3378148Z * [new tag] v2.4.0-rc2 -> v2.4.0-rc2 2025-12-04T08:57:04.3379349Z * [new tag] v2.4.0-rc3 -> v2.4.0-rc3 2025-12-04T08:57:04.3380614Z * [new tag] v2.4.0-rc4 -> v2.4.0-rc4 2025-12-04T08:57:04.3381968Z * [new tag] v2.4.0-rc5 -> v2.4.0-rc5 2025-12-04T08:57:04.3383187Z * [new tag] v2.4.0-rc6 -> v2.4.0-rc6 2025-12-04T08:57:04.3384395Z * [new tag] v2.4.0-rc7 -> v2.4.0-rc7 2025-12-04T08:57:04.3385629Z * [new tag] v2.4.0-rc8 -> v2.4.0-rc8 2025-12-04T08:57:04.3386891Z * [new tag] v2.4.0-rc9 -> v2.4.0-rc9 2025-12-04T08:57:04.3387895Z * [new tag] v2.4.1 -> v2.4.1 2025-12-04T08:57:04.3389195Z * [new tag] v2.4.1-rc1 -> v2.4.1-rc1 2025-12-04T08:57:04.3390429Z * [new tag] v2.4.1-rc2 -> v2.4.1-rc2 2025-12-04T08:57:04.3391698Z * [new tag] v2.4.1-rc3 -> v2.4.1-rc3 2025-12-04T08:57:04.3392917Z * [new tag] v2.5.0 -> v2.5.0 2025-12-04T08:57:04.3394140Z * [new tag] v2.5.0-rc1 -> v2.5.0-rc1 2025-12-04T08:57:04.3395255Z * [new tag] v2.5.0-rc10 -> v2.5.0-rc10 2025-12-04T08:57:04.3396415Z * [new tag] v2.5.0-rc2 -> v2.5.0-rc2 2025-12-04T08:57:04.3397596Z * [new tag] v2.5.0-rc3 -> v2.5.0-rc3 2025-12-04T08:57:04.3398884Z * [new tag] v2.5.0-rc4 -> v2.5.0-rc4 2025-12-04T08:57:04.3400095Z * [new tag] v2.5.0-rc5 -> v2.5.0-rc5 2025-12-04T08:57:04.3401386Z * [new tag] v2.5.0-rc6 -> v2.5.0-rc6 2025-12-04T08:57:04.3402601Z * [new tag] v2.5.0-rc7 -> v2.5.0-rc7 2025-12-04T08:57:04.3403904Z * [new tag] v2.5.0-rc8 -> v2.5.0-rc8 2025-12-04T08:57:04.3405869Z * [new tag] v2.5.0-rc9 -> v2.5.0-rc9 2025-12-04T08:57:04.3408304Z * [new tag] v2.5.1 -> v2.5.1 2025-12-04T08:57:04.3409514Z * [new tag] v2.5.1-rc1 -> v2.5.1-rc1 2025-12-04T08:57:04.3410289Z * [new tag] v2.6.0 -> v2.6.0 2025-12-04T08:57:04.3412151Z * [new tag] v2.6.0-rc1 -> v2.6.0-rc1 2025-12-04T08:57:04.3413565Z * [new tag] v2.6.0-rc2 -> v2.6.0-rc2 2025-12-04T08:57:04.3414857Z * [new tag] v2.6.0-rc3 -> v2.6.0-rc3 2025-12-04T08:57:04.3416016Z * [new tag] v2.6.0-rc4 -> v2.6.0-rc4 2025-12-04T08:57:04.3417425Z * [new tag] v2.6.0-rc5 -> v2.6.0-rc5 2025-12-04T08:57:04.3418712Z * [new tag] v2.6.0-rc6 -> v2.6.0-rc6 2025-12-04T08:57:04.3419975Z * [new tag] v2.6.0-rc7 -> v2.6.0-rc7 2025-12-04T08:57:04.3421309Z * [new tag] v2.6.0-rc8 -> v2.6.0-rc8 2025-12-04T08:57:04.3422527Z * [new tag] v2.6.0-rc9 -> v2.6.0-rc9 2025-12-04T08:57:04.3423983Z * [new tag] v2.7.0 -> v2.7.0 2025-12-04T08:57:04.3425216Z * [new tag] v2.7.0-rc1 -> v2.7.0-rc1 2025-12-04T08:57:04.3426275Z * [new tag] v2.7.0-rc10 -> v2.7.0-rc10 2025-12-04T08:57:04.3427630Z * [new tag] v2.7.0-rc2 -> v2.7.0-rc2 2025-12-04T08:57:04.3428844Z * [new tag] v2.7.0-rc3 -> v2.7.0-rc3 2025-12-04T08:57:04.3430180Z * [new tag] v2.7.0-rc4 -> v2.7.0-rc4 2025-12-04T08:57:04.3431404Z * [new tag] v2.7.0-rc5 -> v2.7.0-rc5 2025-12-04T08:57:04.3432577Z * [new tag] v2.7.0-rc6 -> v2.7.0-rc6 2025-12-04T08:57:04.3434040Z * [new tag] v2.7.0-rc7 -> v2.7.0-rc7 2025-12-04T08:57:04.3435474Z * [new tag] v2.7.0-rc8 -> v2.7.0-rc8 2025-12-04T08:57:04.3436857Z * [new tag] v2.7.0-rc9 -> v2.7.0-rc9 2025-12-04T08:57:04.3437918Z * [new tag] v2.7.1 -> v2.7.1 2025-12-04T08:57:04.3439270Z * [new tag] v2.7.1-rc1 -> v2.7.1-rc1 2025-12-04T08:57:04.3440563Z * [new tag] v2.7.1-rc2 -> v2.7.1-rc2 2025-12-04T08:57:04.3441856Z * [new tag] v2.7.1-rc3 -> v2.7.1-rc3 2025-12-04T08:57:04.3443169Z * [new tag] v2.7.1-rc4 -> v2.7.1-rc4 2025-12-04T08:57:04.3444361Z * [new tag] v2.7.1-rc5 -> v2.7.1-rc5 2025-12-04T08:57:04.3445464Z * [new tag] v2.8.0 -> v2.8.0 2025-12-04T08:57:04.3446720Z * [new tag] v2.8.0-rc1 -> v2.8.0-rc1 2025-12-04T08:57:04.3447954Z * [new tag] v2.8.0-rc2 -> v2.8.0-rc2 2025-12-04T08:57:04.3449262Z * [new tag] v2.8.0-rc3 -> v2.8.0-rc3 2025-12-04T08:57:04.3450621Z * [new tag] v2.8.0-rc4 -> v2.8.0-rc4 2025-12-04T08:57:04.3451891Z * [new tag] v2.8.0-rc5 -> v2.8.0-rc5 2025-12-04T08:57:04.3453269Z * [new tag] v2.8.0-rc6 -> v2.8.0-rc6 2025-12-04T08:57:04.3454675Z * [new tag] v2.8.0-rc7 -> v2.8.0-rc7 2025-12-04T08:57:04.3455884Z * [new tag] v2.8.0-rc8 -> v2.8.0-rc8 2025-12-04T08:57:04.3457183Z * [new tag] v2.9.0 -> v2.9.0 2025-12-04T08:57:04.3458430Z * [new tag] v2.9.0-rc1 -> v2.9.0-rc1 2025-12-04T08:57:04.3459711Z * [new tag] v2.9.0-rc10 -> v2.9.0-rc10 2025-12-04T08:57:04.3460999Z * [new tag] v2.9.0-rc11 -> v2.9.0-rc11 2025-12-04T08:57:04.3462571Z * [new tag] v2.9.0-rc2 -> v2.9.0-rc2 2025-12-04T08:57:04.3463781Z * [new tag] v2.9.0-rc3 -> v2.9.0-rc3 2025-12-04T08:57:04.3465055Z * [new tag] v2.9.0-rc4 -> v2.9.0-rc4 2025-12-04T08:57:04.3466371Z * [new tag] v2.9.0-rc5 -> v2.9.0-rc5 2025-12-04T08:57:04.3467833Z * [new tag] v2.9.0-rc6 -> v2.9.0-rc6 2025-12-04T08:57:04.3469096Z * [new tag] v2.9.0-rc7 -> v2.9.0-rc7 2025-12-04T08:57:04.3470506Z * [new tag] v2.9.0-rc8 -> v2.9.0-rc8 2025-12-04T08:57:04.3471586Z * [new tag] v2.9.0-rc9 -> v2.9.0-rc9 2025-12-04T08:57:04.3472678Z * [new tag] v2.9.1 -> v2.9.1 2025-12-04T08:57:04.3473956Z * [new tag] v2.9.1-rc1 -> v2.9.1-rc1 2025-12-04T08:57:04.3475220Z * [new tag] v2.9.1-rc2 -> v2.9.1-rc2 2025-12-04T08:57:04.3476875Z * [new tag] viable/strict/1759343184 -> viable/strict/1759343184 2025-12-04T08:57:04.3478099Z * [new tag] viable/strict/1759346540 -> viable/strict/1759346540 2025-12-04T08:57:04.3479272Z * [new tag] viable/strict/1759348181 -> viable/strict/1759348181 2025-12-04T08:57:04.3480451Z * [new tag] viable/strict/1759350324 -> viable/strict/1759350324 2025-12-04T08:57:04.3481660Z * [new tag] viable/strict/1759351793 -> viable/strict/1759351793 2025-12-04T08:57:04.3482890Z * [new tag] viable/strict/1759353844 -> viable/strict/1759353844 2025-12-04T08:57:04.3484109Z * [new tag] viable/strict/1759355374 -> viable/strict/1759355374 2025-12-04T08:57:04.3485647Z * [new tag] viable/strict/1759357472 -> viable/strict/1759357472 2025-12-04T08:57:04.3486865Z * [new tag] viable/strict/1759361002 -> viable/strict/1759361002 2025-12-04T08:57:04.3488334Z * [new tag] viable/strict/1759362585 -> viable/strict/1759362585 2025-12-04T08:57:04.3489818Z * [new tag] viable/strict/1759365359 -> viable/strict/1759365359 2025-12-04T08:57:04.3491079Z * [new tag] viable/strict/1759370089 -> viable/strict/1759370089 2025-12-04T08:57:04.3492353Z * [new tag] viable/strict/1759377554 -> viable/strict/1759377554 2025-12-04T08:57:04.3493752Z * [new tag] viable/strict/1759379133 -> viable/strict/1759379133 2025-12-04T08:57:04.3495071Z * [new tag] viable/strict/1759389871 -> viable/strict/1759389871 2025-12-04T08:57:04.3496332Z * [new tag] viable/strict/1759393562 -> viable/strict/1759393562 2025-12-04T08:57:04.3497611Z * [new tag] viable/strict/1759395076 -> viable/strict/1759395076 2025-12-04T08:57:04.3498880Z * [new tag] viable/strict/1759398579 -> viable/strict/1759398579 2025-12-04T08:57:04.3500143Z * [new tag] viable/strict/1759404142 -> viable/strict/1759404142 2025-12-04T08:57:04.3501376Z * [new tag] viable/strict/1759405773 -> viable/strict/1759405773 2025-12-04T08:57:04.3502596Z * [new tag] viable/strict/1759408041 -> viable/strict/1759408041 2025-12-04T08:57:04.3503938Z * [new tag] viable/strict/1759411593 -> viable/strict/1759411593 2025-12-04T08:57:04.3505068Z * [new tag] viable/strict/1759427395 -> viable/strict/1759427395 2025-12-04T08:57:04.3506659Z * [new tag] viable/strict/1759434582 -> viable/strict/1759434582 2025-12-04T08:57:04.3507927Z * [new tag] viable/strict/1759436720 -> viable/strict/1759436720 2025-12-04T08:57:04.3509190Z * [new tag] viable/strict/1759440219 -> viable/strict/1759440219 2025-12-04T08:57:04.3510436Z * [new tag] viable/strict/1759441948 -> viable/strict/1759441948 2025-12-04T08:57:04.3511881Z * [new tag] viable/strict/1759443860 -> viable/strict/1759443860 2025-12-04T08:57:04.3512979Z * [new tag] viable/strict/1759445377 -> viable/strict/1759445377 2025-12-04T08:57:04.3514320Z * [new tag] viable/strict/1759447415 -> viable/strict/1759447415 2025-12-04T08:57:04.3515522Z * [new tag] viable/strict/1759451750 -> viable/strict/1759451750 2025-12-04T08:57:04.3516773Z * [new tag] viable/strict/1759453910 -> viable/strict/1759453910 2025-12-04T08:57:04.3518068Z * [new tag] viable/strict/1759456483 -> viable/strict/1759456483 2025-12-04T08:57:04.3519369Z * [new tag] viable/strict/1759459279 -> viable/strict/1759459279 2025-12-04T08:57:04.3520605Z * [new tag] viable/strict/1759460742 -> viable/strict/1759460742 2025-12-04T08:57:04.3521862Z * [new tag] viable/strict/1759462025 -> viable/strict/1759462025 2025-12-04T08:57:04.3523170Z * [new tag] viable/strict/1759469086 -> viable/strict/1759469086 2025-12-04T08:57:04.3524396Z * [new tag] viable/strict/1759470581 -> viable/strict/1759470581 2025-12-04T08:57:04.3525727Z * [new tag] viable/strict/1759472786 -> viable/strict/1759472786 2025-12-04T08:57:04.3526965Z * [new tag] viable/strict/1759476294 -> viable/strict/1759476294 2025-12-04T08:57:04.3528212Z * [new tag] viable/strict/1759479963 -> viable/strict/1759479963 2025-12-04T08:57:04.3529436Z * [new tag] viable/strict/1759492177 -> viable/strict/1759492177 2025-12-04T08:57:04.3530712Z * [new tag] viable/strict/1759519278 -> viable/strict/1759519278 2025-12-04T08:57:04.3531953Z * [new tag] viable/strict/1759524580 -> viable/strict/1759524580 2025-12-04T08:57:04.3533330Z * [new tag] viable/strict/1759528193 -> viable/strict/1759528193 2025-12-04T08:57:04.3534850Z * [new tag] viable/strict/1759533797 -> viable/strict/1759533797 2025-12-04T08:57:04.3536087Z * [new tag] viable/strict/1759542780 -> viable/strict/1759542780 2025-12-04T08:57:04.3537358Z * [new tag] viable/strict/1759549779 -> viable/strict/1759549779 2025-12-04T08:57:04.3538629Z * [new tag] viable/strict/1759555455 -> viable/strict/1759555455 2025-12-04T08:57:04.3539895Z * [new tag] viable/strict/1759559176 -> viable/strict/1759559176 2025-12-04T08:57:04.3541190Z * [new tag] viable/strict/1759560629 -> viable/strict/1759560629 2025-12-04T08:57:04.3542392Z * [new tag] viable/strict/1759569848 -> viable/strict/1759569848 2025-12-04T08:57:04.3543819Z * [new tag] viable/strict/1759571382 -> viable/strict/1759571382 2025-12-04T08:57:04.3545093Z * [new tag] viable/strict/1759573474 -> viable/strict/1759573474 2025-12-04T08:57:04.3546344Z * [new tag] viable/strict/1759618187 -> viable/strict/1759618187 2025-12-04T08:57:04.3547591Z * [new tag] viable/strict/1759626742 -> viable/strict/1759626742 2025-12-04T08:57:04.3548888Z * [new tag] viable/strict/1759632427 -> viable/strict/1759632427 2025-12-04T08:57:04.3550174Z * [new tag] viable/strict/1759634971 -> viable/strict/1759634971 2025-12-04T08:57:04.3551529Z * [new tag] viable/strict/1759661382 -> viable/strict/1759661382 2025-12-04T08:57:04.3552849Z * [new tag] viable/strict/1759663294 -> viable/strict/1759663294 2025-12-04T08:57:04.3553969Z * [new tag] viable/strict/1759708178 -> viable/strict/1759708178 2025-12-04T08:57:04.3555244Z * [new tag] viable/strict/1759715695 -> viable/strict/1759715695 2025-12-04T08:57:04.3556579Z * [new tag] viable/strict/1759728293 -> viable/strict/1759728293 2025-12-04T08:57:04.3557755Z * [new tag] viable/strict/1759735513 -> viable/strict/1759735513 2025-12-04T08:57:04.3559075Z * [new tag] viable/strict/1759739177 -> viable/strict/1759739177 2025-12-04T08:57:04.3560305Z * [new tag] viable/strict/1759758635 -> viable/strict/1759758635 2025-12-04T08:57:04.3561552Z * [new tag] viable/strict/1759765784 -> viable/strict/1759765784 2025-12-04T08:57:04.3562767Z * [new tag] viable/strict/1759767948 -> viable/strict/1759767948 2025-12-04T08:57:04.3564052Z * [new tag] viable/strict/1759771461 -> viable/strict/1759771461 2025-12-04T08:57:04.3565123Z * [new tag] viable/strict/1759776706 -> viable/strict/1759776706 2025-12-04T08:57:04.3566408Z * [new tag] viable/strict/1759782317 -> viable/strict/1759782317 2025-12-04T08:57:04.3568198Z * [new tag] viable/strict/1759783777 -> viable/strict/1759783777 2025-12-04T08:57:04.3569525Z * [new tag] viable/strict/1759785815 -> viable/strict/1759785815 2025-12-04T08:57:04.3570752Z * [new tag] viable/strict/1759789459 -> viable/strict/1759789459 2025-12-04T08:57:04.3571976Z * [new tag] viable/strict/1759790974 -> viable/strict/1759790974 2025-12-04T08:57:04.3573169Z * [new tag] viable/strict/1759794583 -> viable/strict/1759794583 2025-12-04T08:57:04.3574662Z * [new tag] viable/strict/1759797408 -> viable/strict/1759797408 2025-12-04T08:57:04.3575950Z * [new tag] viable/strict/1759799518 -> viable/strict/1759799518 2025-12-04T08:57:04.3577204Z * [new tag] viable/strict/1759804909 -> viable/strict/1759804909 2025-12-04T08:57:04.3578452Z * [new tag] viable/strict/1759807643 -> viable/strict/1759807643 2025-12-04T08:57:04.3579758Z * [new tag] viable/strict/1759809089 -> viable/strict/1759809089 2025-12-04T08:57:04.3580998Z * [new tag] viable/strict/1759811145 -> viable/strict/1759811145 2025-12-04T08:57:04.3582243Z * [new tag] viable/strict/1759812581 -> viable/strict/1759812581 2025-12-04T08:57:04.3583560Z * [new tag] viable/strict/1759814683 -> viable/strict/1759814683 2025-12-04T08:57:04.3584922Z * [new tag] viable/strict/1759821889 -> viable/strict/1759821889 2025-12-04T08:57:04.3586209Z * [new tag] viable/strict/1759823376 -> viable/strict/1759823376 2025-12-04T08:57:04.3587464Z * [new tag] viable/strict/1759827107 -> viable/strict/1759827107 2025-12-04T08:57:04.3588699Z * [new tag] viable/strict/1759830577 -> viable/strict/1759830577 2025-12-04T08:57:04.3590057Z * [new tag] viable/strict/1759832720 -> viable/strict/1759832720 2025-12-04T08:57:04.3591269Z * [new tag] viable/strict/1759842063 -> viable/strict/1759842063 2025-12-04T08:57:04.3592502Z * [new tag] viable/strict/1759847121 -> viable/strict/1759847121 2025-12-04T08:57:04.3593998Z * [new tag] viable/strict/1759850721 -> viable/strict/1759850721 2025-12-04T08:57:04.3595285Z * [new tag] viable/strict/1759857870 -> viable/strict/1759857870 2025-12-04T08:57:04.3596548Z * [new tag] viable/strict/1759863143 -> viable/strict/1759863143 2025-12-04T08:57:04.3597795Z * [new tag] viable/strict/1759875874 -> viable/strict/1759875874 2025-12-04T08:57:04.3598942Z * [new tag] viable/strict/1759877385 -> viable/strict/1759877385 2025-12-04T08:57:04.3600231Z * [new tag] viable/strict/1759883801 -> viable/strict/1759883801 2025-12-04T08:57:04.3601480Z * [new tag] viable/strict/1759885922 -> viable/strict/1759885922 2025-12-04T08:57:04.3602842Z * [new tag] viable/strict/1759888488 -> viable/strict/1759888488 2025-12-04T08:57:04.3604011Z * [new tag] viable/strict/1759895471 -> viable/strict/1759895471 2025-12-04T08:57:04.3605470Z * [new tag] viable/strict/1759904803 -> viable/strict/1759904803 2025-12-04T08:57:04.3606926Z * [new tag] viable/strict/1759908300 -> viable/strict/1759908300 2025-12-04T08:57:04.3608227Z * [new tag] viable/strict/1759915520 -> viable/strict/1759915520 2025-12-04T08:57:04.3609511Z * [new tag] viable/strict/1759916978 -> viable/strict/1759916978 2025-12-04T08:57:04.3610654Z * [new tag] viable/strict/1759930024 -> viable/strict/1759930024 2025-12-04T08:57:04.3611925Z * [new tag] viable/strict/1759948122 -> viable/strict/1759948122 2025-12-04T08:57:04.3613233Z * [new tag] viable/strict/1759952983 -> viable/strict/1759952983 2025-12-04T08:57:04.3614664Z * [new tag] viable/strict/1759955121 -> viable/strict/1759955121 2025-12-04T08:57:04.3615968Z * [new tag] viable/strict/1759962298 -> viable/strict/1759962298 2025-12-04T08:57:04.3617306Z * [new tag] viable/strict/1759965837 -> viable/strict/1759965837 2025-12-04T08:57:04.3618573Z * [new tag] viable/strict/1759970213 -> viable/strict/1759970213 2025-12-04T08:57:04.3619943Z * [new tag] viable/strict/1759974894 -> viable/strict/1759974894 2025-12-04T08:57:04.3621177Z * [new tag] viable/strict/1759977763 -> viable/strict/1759977763 2025-12-04T08:57:04.3622466Z * [new tag] viable/strict/1759979241 -> viable/strict/1759979241 2025-12-04T08:57:04.3623741Z * [new tag] viable/strict/1759985417 -> viable/strict/1759985417 2025-12-04T08:57:04.3625039Z * [new tag] viable/strict/1759987490 -> viable/strict/1759987490 2025-12-04T08:57:04.3626327Z * [new tag] viable/strict/1759996180 -> viable/strict/1759996180 2025-12-04T08:57:04.3627554Z * [new tag] viable/strict/1760065682 -> viable/strict/1760065682 2025-12-04T08:57:04.3628870Z * [new tag] viable/strict/1760066894 -> viable/strict/1760066894 2025-12-04T08:57:04.3630234Z * [new tag] viable/strict/1760070345 -> viable/strict/1760070345 2025-12-04T08:57:04.3631497Z * [new tag] viable/strict/1760089782 -> viable/strict/1760089782 2025-12-04T08:57:04.3632777Z * [new tag] viable/strict/1760091921 -> viable/strict/1760091921 2025-12-04T08:57:04.3634009Z * [new tag] viable/strict/1760127924 -> viable/strict/1760127924 2025-12-04T08:57:04.3635345Z * [new tag] viable/strict/1760129489 -> viable/strict/1760129489 2025-12-04T08:57:04.3636662Z * [new tag] viable/strict/1760132980 -> viable/strict/1760132980 2025-12-04T08:57:04.3637963Z * [new tag] viable/strict/1760135060 -> viable/strict/1760135060 2025-12-04T08:57:04.3650480Z * [new tag] viable/strict/1760215782 -> viable/strict/1760215782 2025-12-04T08:57:04.3650762Z * [new tag] viable/strict/1760273849 -> viable/strict/1760273849 2025-12-04T08:57:04.3650912Z * [new tag] viable/strict/1760275517 -> viable/strict/1760275517 2025-12-04T08:57:04.3651057Z * [new tag] viable/strict/1760276979 -> viable/strict/1760276979 2025-12-04T08:57:04.3651181Z * [new tag] viable/strict/1760279007 -> viable/strict/1760279007 2025-12-04T08:57:04.3651307Z * [new tag] viable/strict/1760286328 -> viable/strict/1760286328 2025-12-04T08:57:04.3651438Z * [new tag] viable/strict/1760493304 -> viable/strict/1760493304 2025-12-04T08:57:04.3651563Z * [new tag] viable/strict/1760496298 -> viable/strict/1760496298 2025-12-04T08:57:04.3651860Z * [new tag] viable/strict/1760518396 -> viable/strict/1760518396 2025-12-04T08:57:04.3652095Z * [new tag] viable/strict/1760534864 -> viable/strict/1760534864 2025-12-04T08:57:04.3652233Z * [new tag] viable/strict/1760549062 -> viable/strict/1760549062 2025-12-04T08:57:04.3653669Z * [new tag] viable/strict/1760552799 -> viable/strict/1760552799 2025-12-04T08:57:04.3654920Z * [new tag] viable/strict/1760554355 -> viable/strict/1760554355 2025-12-04T08:57:04.3656319Z * [new tag] viable/strict/1760556275 -> viable/strict/1760556275 2025-12-04T08:57:04.3657374Z * [new tag] viable/strict/1760564979 -> viable/strict/1760564979 2025-12-04T08:57:04.3658921Z * [new tag] viable/strict/1760567049 -> viable/strict/1760567049 2025-12-04T08:57:04.3660483Z * [new tag] viable/strict/1760568585 -> viable/strict/1760568585 2025-12-04T08:57:04.3661723Z * [new tag] viable/strict/1760570630 -> viable/strict/1760570630 2025-12-04T08:57:04.3662955Z * [new tag] viable/strict/1760572180 -> viable/strict/1760572180 2025-12-04T08:57:04.3664277Z * [new tag] viable/strict/1760575094 -> viable/strict/1760575094 2025-12-04T08:57:04.3665611Z * [new tag] viable/strict/1760579709 -> viable/strict/1760579709 2025-12-04T08:57:04.3667234Z * [new tag] viable/strict/1760582614 -> viable/strict/1760582614 2025-12-04T08:57:04.3668591Z * [new tag] viable/strict/1760586815 -> viable/strict/1760586815 2025-12-04T08:57:04.3669726Z * [new tag] viable/strict/1760588829 -> viable/strict/1760588829 2025-12-04T08:57:04.3670986Z * [new tag] viable/strict/1760590200 -> viable/strict/1760590200 2025-12-04T08:57:04.3672332Z * [new tag] viable/strict/1760592311 -> viable/strict/1760592311 2025-12-04T08:57:04.3673539Z * [new tag] viable/strict/1760619733 -> viable/strict/1760619733 2025-12-04T08:57:04.3674700Z * [new tag] viable/strict/1760628335 -> viable/strict/1760628335 2025-12-04T08:57:04.3675939Z * [new tag] viable/strict/1760635490 -> viable/strict/1760635490 2025-12-04T08:57:04.3677220Z * [new tag] viable/strict/1760640743 -> viable/strict/1760640743 2025-12-04T08:57:04.3678508Z * [new tag] viable/strict/1760642528 -> viable/strict/1760642528 2025-12-04T08:57:04.3679852Z * [new tag] viable/strict/1760646330 -> viable/strict/1760646330 2025-12-04T08:57:04.3681096Z * [new tag] viable/strict/1760666101 -> viable/strict/1760666101 2025-12-04T08:57:04.3682351Z * [new tag] viable/strict/1760668990 -> viable/strict/1760668990 2025-12-04T08:57:04.3683616Z * [new tag] viable/strict/1760670600 -> viable/strict/1760670600 2025-12-04T08:57:04.3684856Z * [new tag] viable/strict/1760671704 -> viable/strict/1760671704 2025-12-04T08:57:04.3686112Z * [new tag] viable/strict/1760673121 -> viable/strict/1760673121 2025-12-04T08:57:04.3687380Z * [new tag] viable/strict/1760675352 -> viable/strict/1760675352 2025-12-04T08:57:04.3688664Z * [new tag] viable/strict/1760696731 -> viable/strict/1760696731 2025-12-04T08:57:04.3691048Z * [new tag] viable/strict/1760723515 -> viable/strict/1760723515 2025-12-04T08:57:04.3692328Z * [new tag] viable/strict/1760727234 -> viable/strict/1760727234 2025-12-04T08:57:04.3693715Z * [new tag] viable/strict/1760730578 -> viable/strict/1760730578 2025-12-04T08:57:04.3694926Z * [new tag] viable/strict/1760732726 -> viable/strict/1760732726 2025-12-04T08:57:04.3696247Z * [new tag] viable/strict/1760734180 -> viable/strict/1760734180 2025-12-04T08:57:04.3697620Z * [new tag] viable/strict/1760736251 -> viable/strict/1760736251 2025-12-04T08:57:04.3698839Z * [new tag] viable/strict/1760737772 -> viable/strict/1760737772 2025-12-04T08:57:04.3700251Z * [new tag] viable/strict/1760758005 -> viable/strict/1760758005 2025-12-04T08:57:04.3701565Z * [new tag] viable/strict/1760761532 -> viable/strict/1760761532 2025-12-04T08:57:04.3702885Z * [new tag] viable/strict/1760802581 -> viable/strict/1760802581 2025-12-04T08:57:04.3704197Z * [new tag] viable/strict/1760827772 -> viable/strict/1760827772 2025-12-04T08:57:04.3705768Z * [new tag] viable/strict/1760834524 -> viable/strict/1760834524 2025-12-04T08:57:04.3707147Z * [new tag] viable/strict/1760845009 -> viable/strict/1760845009 2025-12-04T08:57:04.3708426Z * [new tag] viable/strict/1760876836 -> viable/strict/1760876836 2025-12-04T08:57:04.3709720Z * [new tag] viable/strict/1760880329 -> viable/strict/1760880329 2025-12-04T08:57:04.3710971Z * [new tag] viable/strict/1760888987 -> viable/strict/1760888987 2025-12-04T08:57:04.3712255Z * [new tag] viable/strict/1760912664 -> viable/strict/1760912664 2025-12-04T08:57:04.3713533Z * [new tag] viable/strict/1760925321 -> viable/strict/1760925321 2025-12-04T08:57:04.3714799Z * [new tag] viable/strict/1760931488 -> viable/strict/1760931488 2025-12-04T08:57:04.3716039Z * [new tag] viable/strict/1760932693 -> viable/strict/1760932693 2025-12-04T08:57:04.3717370Z * [new tag] viable/strict/1761004184 -> viable/strict/1761004184 2025-12-04T08:57:04.3718643Z * [new tag] viable/strict/1761014748 -> viable/strict/1761014748 2025-12-04T08:57:04.3719911Z * [new tag] viable/strict/1761017491 -> viable/strict/1761017491 2025-12-04T08:57:04.3721221Z * [new tag] viable/strict/1761018806 -> viable/strict/1761018806 2025-12-04T08:57:04.3722541Z * [new tag] viable/strict/1761020754 -> viable/strict/1761020754 2025-12-04T08:57:04.3723789Z * [new tag] viable/strict/1761024303 -> viable/strict/1761024303 2025-12-04T08:57:04.3725069Z * [new tag] viable/strict/1761029582 -> viable/strict/1761029582 2025-12-04T08:57:04.3726299Z * [new tag] viable/strict/1761031535 -> viable/strict/1761031535 2025-12-04T08:57:04.3727556Z * [new tag] viable/strict/1761035196 -> viable/strict/1761035196 2025-12-04T08:57:04.3728968Z * [new tag] viable/strict/1761045825 -> viable/strict/1761045825 2025-12-04T08:57:04.3730231Z * [new tag] viable/strict/1761054796 -> viable/strict/1761054796 2025-12-04T08:57:04.3731496Z * [new tag] viable/strict/1761060314 -> viable/strict/1761060314 2025-12-04T08:57:04.3732819Z * [new tag] viable/strict/1761071198 -> viable/strict/1761071198 2025-12-04T08:57:04.3734543Z * [new tag] viable/strict/1761074628 -> viable/strict/1761074628 2025-12-04T08:57:04.3736346Z * [new tag] viable/strict/1761078351 -> viable/strict/1761078351 2025-12-04T08:57:04.3737590Z * [new tag] viable/strict/1761079822 -> viable/strict/1761079822 2025-12-04T08:57:04.3738816Z * [new tag] viable/strict/1761081873 -> viable/strict/1761081873 2025-12-04T08:57:04.3740231Z * [new tag] viable/strict/1761083392 -> viable/strict/1761083392 2025-12-04T08:57:04.3741513Z * [new tag] viable/strict/1761085465 -> viable/strict/1761085465 2025-12-04T08:57:04.3742845Z * [new tag] viable/strict/1761089099 -> viable/strict/1761089099 2025-12-04T08:57:04.3744167Z * [new tag] viable/strict/1761095535 -> viable/strict/1761095535 2025-12-04T08:57:04.3745589Z * [new tag] viable/strict/1761098119 -> viable/strict/1761098119 2025-12-04T08:57:04.3747062Z * [new tag] viable/strict/1761101330 -> viable/strict/1761101330 2025-12-04T08:57:04.3748310Z * [new tag] viable/strict/1761114425 -> viable/strict/1761114425 2025-12-04T08:57:04.3749593Z * [new tag] viable/strict/1761116036 -> viable/strict/1761116036 2025-12-04T08:57:04.3750895Z * [new tag] viable/strict/1761119379 -> viable/strict/1761119379 2025-12-04T08:57:04.3752132Z * [new tag] viable/strict/1761121601 -> viable/strict/1761121601 2025-12-04T08:57:04.3753499Z * [new tag] viable/strict/1761123234 -> viable/strict/1761123234 2025-12-04T08:57:04.3754934Z * [new tag] viable/strict/1761126621 -> viable/strict/1761126621 2025-12-04T08:57:04.3756218Z * [new tag] viable/strict/1761132259 -> viable/strict/1761132259 2025-12-04T08:57:04.3757547Z * [new tag] viable/strict/1761146746 -> viable/strict/1761146746 2025-12-04T08:57:04.3758788Z * [new tag] viable/strict/1761164752 -> viable/strict/1761164752 2025-12-04T08:57:04.3760024Z * [new tag] viable/strict/1761166198 -> viable/strict/1761166198 2025-12-04T08:57:04.3761308Z * [new tag] viable/strict/1761175424 -> viable/strict/1761175424 2025-12-04T08:57:04.3762554Z * [new tag] viable/strict/1761176983 -> viable/strict/1761176983 2025-12-04T08:57:04.3763961Z * [new tag] viable/strict/1761179891 -> viable/strict/1761179891 2025-12-04T08:57:04.3765211Z * [new tag] viable/strict/1761181930 -> viable/strict/1761181930 2025-12-04T08:57:04.3766545Z * [new tag] viable/strict/1761184516 -> viable/strict/1761184516 2025-12-04T08:57:04.3767817Z * [new tag] viable/strict/1761190179 -> viable/strict/1761190179 2025-12-04T08:57:04.3769101Z * [new tag] viable/strict/1761193558 -> viable/strict/1761193558 2025-12-04T08:57:04.3770341Z * [new tag] viable/strict/1761207990 -> viable/strict/1761207990 2025-12-04T08:57:04.3771615Z * [new tag] viable/strict/1761229539 -> viable/strict/1761229539 2025-12-04T08:57:04.3773069Z * [new tag] viable/strict/1761244031 -> viable/strict/1761244031 2025-12-04T08:57:04.3774695Z * [new tag] viable/strict/1761248986 -> viable/strict/1761248986 2025-12-04T08:57:04.3776018Z * [new tag] viable/strict/1761259791 -> viable/strict/1761259791 2025-12-04T08:57:04.3777257Z * [new tag] viable/strict/1761266139 -> viable/strict/1761266139 2025-12-04T08:57:04.3778575Z * [new tag] viable/strict/1761268316 -> viable/strict/1761268316 2025-12-04T08:57:04.3779821Z * [new tag] viable/strict/1761273805 -> viable/strict/1761273805 2025-12-04T08:57:04.3781109Z * [new tag] viable/strict/1761275261 -> viable/strict/1761275261 2025-12-04T08:57:04.3782408Z * [new tag] viable/strict/1761277913 -> viable/strict/1761277913 2025-12-04T08:57:04.3783694Z * [new tag] viable/strict/1761290701 -> viable/strict/1761290701 2025-12-04T08:57:04.3785060Z * [new tag] viable/strict/1761294396 -> viable/strict/1761294396 2025-12-04T08:57:04.3786379Z * [new tag] viable/strict/1761303047 -> viable/strict/1761303047 2025-12-04T08:57:04.3787673Z * [new tag] viable/strict/1761335388 -> viable/strict/1761335388 2025-12-04T08:57:04.3788916Z * [new tag] viable/strict/1761337551 -> viable/strict/1761337551 2025-12-04T08:57:04.3790163Z * [new tag] viable/strict/1761339007 -> viable/strict/1761339007 2025-12-04T08:57:04.3791506Z * [new tag] viable/strict/1761341050 -> viable/strict/1761341050 2025-12-04T08:57:04.3792924Z * [new tag] viable/strict/1761346188 -> viable/strict/1761346188 2025-12-04T08:57:04.3794210Z * [new tag] viable/strict/1761349792 -> viable/strict/1761349792 2025-12-04T08:57:04.3795445Z * [new tag] viable/strict/1761352620 -> viable/strict/1761352620 2025-12-04T08:57:04.3796761Z * [new tag] viable/strict/1761354730 -> viable/strict/1761354730 2025-12-04T08:57:04.3798096Z * [new tag] viable/strict/1761357298 -> viable/strict/1761357298 2025-12-04T08:57:04.3799320Z * [new tag] viable/strict/1761360201 -> viable/strict/1761360201 2025-12-04T08:57:04.3800622Z * [new tag] viable/strict/1761361753 -> viable/strict/1761361753 2025-12-04T08:57:04.3801862Z * [new tag] viable/strict/1761364351 -> viable/strict/1761364351 2025-12-04T08:57:04.3803128Z * [new tag] viable/strict/1761366338 -> viable/strict/1761366338 2025-12-04T08:57:04.3804648Z * [new tag] viable/strict/1761367802 -> viable/strict/1761367802 2025-12-04T08:57:04.3808333Z * [new tag] viable/strict/1761369889 -> viable/strict/1761369889 2025-12-04T08:57:04.3809670Z * [new tag] viable/strict/1761371385 -> viable/strict/1761371385 2025-12-04T08:57:04.3810995Z * [new tag] viable/strict/1761373581 -> viable/strict/1761373581 2025-12-04T08:57:04.3812394Z * [new tag] viable/strict/1761375054 -> viable/strict/1761375054 2025-12-04T08:57:04.3813843Z * [new tag] viable/strict/1761421785 -> viable/strict/1761421785 2025-12-04T08:57:04.3815346Z * [new tag] viable/strict/1761434614 -> viable/strict/1761434614 2025-12-04T08:57:04.3816766Z * [new tag] viable/strict/1761439254 -> viable/strict/1761439254 2025-12-04T08:57:04.3818139Z * [new tag] viable/strict/1761454187 -> viable/strict/1761454187 2025-12-04T08:57:04.3819484Z * [new tag] viable/strict/1761459991 -> viable/strict/1761459991 2025-12-04T08:57:04.3820945Z * [new tag] viable/strict/1761470668 -> viable/strict/1761470668 2025-12-04T08:57:04.3822535Z * [new tag] viable/strict/1761472188 -> viable/strict/1761472188 2025-12-04T08:57:04.3824428Z * [new tag] viable/strict/1761503178 -> viable/strict/1761503178 2025-12-04T08:57:04.3825861Z * [new tag] viable/strict/1761517492 -> viable/strict/1761517492 2025-12-04T08:57:04.3827199Z * [new tag] viable/strict/1761518981 -> viable/strict/1761518981 2025-12-04T08:57:04.3828492Z * [new tag] viable/strict/1761533609 -> viable/strict/1761533609 2025-12-04T08:57:04.3829663Z * [new tag] viable/strict/1761546438 -> viable/strict/1761546438 2025-12-04T08:57:04.3831074Z * [new tag] viable/strict/1761548133 -> viable/strict/1761548133 2025-12-04T08:57:04.3832675Z * [new tag] viable/strict/1761555186 -> viable/strict/1761555186 2025-12-04T08:57:04.3834169Z * [new tag] viable/strict/1761557178 -> viable/strict/1761557178 2025-12-04T08:57:04.3835613Z * [new tag] viable/strict/1761560772 -> viable/strict/1761560772 2025-12-04T08:57:04.3836947Z * [new tag] viable/strict/1761562266 -> viable/strict/1761562266 2025-12-04T08:57:04.3838314Z * [new tag] viable/strict/1761564260 -> viable/strict/1761564260 2025-12-04T08:57:04.3839554Z * [new tag] viable/strict/1761568072 -> viable/strict/1761568072 2025-12-04T08:57:04.3840807Z * [new tag] viable/strict/1761571683 -> viable/strict/1761571683 2025-12-04T08:57:04.3842000Z * [new tag] viable/strict/1761580199 -> viable/strict/1761580199 2025-12-04T08:57:04.3843294Z * [new tag] viable/strict/1761587383 -> viable/strict/1761587383 2025-12-04T08:57:04.3844768Z * [new tag] viable/strict/1761591165 -> viable/strict/1761591165 2025-12-04T08:57:04.3845906Z * [new tag] viable/strict/1761594575 -> viable/strict/1761594575 2025-12-04T08:57:04.3847272Z * [new tag] viable/strict/1761596710 -> viable/strict/1761596710 2025-12-04T08:57:04.3848572Z * [new tag] viable/strict/1761598189 -> viable/strict/1761598189 2025-12-04T08:57:04.3849861Z * [new tag] viable/strict/1761600254 -> viable/strict/1761600254 2025-12-04T08:57:04.3851126Z * [new tag] viable/strict/1761603879 -> viable/strict/1761603879 2025-12-04T08:57:04.3852430Z * [new tag] viable/strict/1761605429 -> viable/strict/1761605429 2025-12-04T08:57:04.3853964Z * [new tag] viable/strict/1761607468 -> viable/strict/1761607468 2025-12-04T08:57:04.3855245Z * [new tag] viable/strict/1761608983 -> viable/strict/1761608983 2025-12-04T08:57:04.3856559Z * [new tag] viable/strict/1761611846 -> viable/strict/1761611846 2025-12-04T08:57:04.3857875Z * [new tag] viable/strict/1761613922 -> viable/strict/1761613922 2025-12-04T08:57:04.3859045Z * [new tag] viable/strict/1761616504 -> viable/strict/1761616504 2025-12-04T08:57:04.3860199Z * [new tag] viable/strict/1761619599 -> viable/strict/1761619599 2025-12-04T08:57:04.3861505Z * [new tag] viable/strict/1761686693 -> viable/strict/1761686693 2025-12-04T08:57:04.3862824Z * [new tag] viable/strict/1761688179 -> viable/strict/1761688179 2025-12-04T08:57:04.3864101Z * [new tag] viable/strict/1761691973 -> viable/strict/1761691973 2025-12-04T08:57:04.3865462Z * [new tag] viable/strict/1761693884 -> viable/strict/1761693884 2025-12-04T08:57:04.3866757Z * [new tag] viable/strict/1761695389 -> viable/strict/1761695389 2025-12-04T08:57:04.3868093Z * [new tag] viable/strict/1761698408 -> viable/strict/1761698408 2025-12-04T08:57:04.3869417Z * [new tag] viable/strict/1761702931 -> viable/strict/1761702931 2025-12-04T08:57:04.3870788Z * [new tag] viable/strict/1761706307 -> viable/strict/1761706307 2025-12-04T08:57:04.3872105Z * [new tag] viable/strict/1761709065 -> viable/strict/1761709065 2025-12-04T08:57:04.3873501Z * [new tag] viable/strict/1761710285 -> viable/strict/1761710285 2025-12-04T08:57:04.3874886Z * [new tag] viable/strict/1761711983 -> viable/strict/1761711983 2025-12-04T08:57:04.3876291Z * [new tag] viable/strict/1761713514 -> viable/strict/1761713514 2025-12-04T08:57:04.3877672Z * [new tag] viable/strict/1761715523 -> viable/strict/1761715523 2025-12-04T08:57:04.3879076Z * [new tag] viable/strict/1761727973 -> viable/strict/1761727973 2025-12-04T08:57:04.3880409Z * [new tag] viable/strict/1761751558 -> viable/strict/1761751558 2025-12-04T08:57:04.3881749Z * [new tag] viable/strict/1761755187 -> viable/strict/1761755187 2025-12-04T08:57:04.3883093Z * [new tag] viable/strict/1761756826 -> viable/strict/1761756826 2025-12-04T08:57:04.3884459Z * [new tag] viable/strict/1761769551 -> viable/strict/1761769551 2025-12-04T08:57:04.3885814Z * [new tag] viable/strict/1761771032 -> viable/strict/1761771032 2025-12-04T08:57:04.3886986Z * [new tag] viable/strict/1761773101 -> viable/strict/1761773101 2025-12-04T08:57:04.3888338Z * [new tag] viable/strict/1761781792 -> viable/strict/1761781792 2025-12-04T08:57:04.3889759Z * [new tag] viable/strict/1761784788 -> viable/strict/1761784788 2025-12-04T08:57:04.3891082Z * [new tag] viable/strict/1761786740 -> viable/strict/1761786740 2025-12-04T08:57:04.3892475Z * [new tag] viable/strict/1761789332 -> viable/strict/1761789332 2025-12-04T08:57:04.3894258Z * [new tag] viable/strict/1761792569 -> viable/strict/1761792569 2025-12-04T08:57:04.3895606Z * [new tag] viable/strict/1761795289 -> viable/strict/1761795289 2025-12-04T08:57:04.3896893Z * [new tag] viable/strict/1761798345 -> viable/strict/1761798345 2025-12-04T08:57:04.3898243Z * [new tag] viable/strict/1761799827 -> viable/strict/1761799827 2025-12-04T08:57:04.3899592Z * [new tag] viable/strict/1761805604 -> viable/strict/1761805604 2025-12-04T08:57:04.3900990Z * [new tag] viable/strict/1761807202 -> viable/strict/1761807202 2025-12-04T08:57:04.3902336Z * [new tag] viable/strict/1761809094 -> viable/strict/1761809094 2025-12-04T08:57:04.3903654Z * [new tag] viable/strict/1761810576 -> viable/strict/1761810576 2025-12-04T08:57:04.3905198Z * [new tag] viable/strict/1761812771 -> viable/strict/1761812771 2025-12-04T08:57:04.3906678Z * [new tag] viable/strict/1761814363 -> viable/strict/1761814363 2025-12-04T08:57:04.3908018Z * [new tag] viable/strict/1761857410 -> viable/strict/1761857410 2025-12-04T08:57:04.3909382Z * [new tag] viable/strict/1761860985 -> viable/strict/1761860985 2025-12-04T08:57:04.3911192Z * [new tag] viable/strict/1761863094 -> viable/strict/1761863094 2025-12-04T08:57:04.3912532Z * [new tag] viable/strict/1761864590 -> viable/strict/1761864590 2025-12-04T08:57:04.3913841Z * [new tag] viable/strict/1761866675 -> viable/strict/1761866675 2025-12-04T08:57:04.3915298Z * [new tag] viable/strict/1761868178 -> viable/strict/1761868178 2025-12-04T08:57:04.3916622Z * [new tag] viable/strict/1761871111 -> viable/strict/1761871111 2025-12-04T08:57:04.3918016Z * [new tag] viable/strict/1761873126 -> viable/strict/1761873126 2025-12-04T08:57:04.3919481Z * [new tag] viable/strict/1761875714 -> viable/strict/1761875714 2025-12-04T08:57:04.3920830Z * [new tag] viable/strict/1761878924 -> viable/strict/1761878924 2025-12-04T08:57:04.3922179Z * [new tag] viable/strict/1761881727 -> viable/strict/1761881727 2025-12-04T08:57:04.3923559Z * [new tag] viable/strict/1761882959 -> viable/strict/1761882959 2025-12-04T08:57:04.3925718Z * [new tag] viable/strict/1761886268 -> viable/strict/1761886268 2025-12-04T08:57:04.3926498Z * [new tag] viable/strict/1761893641 -> viable/strict/1761893641 2025-12-04T08:57:04.3927423Z * [new tag] viable/strict/1761931517 -> viable/strict/1761931517 2025-12-04T08:57:04.3928860Z * [new tag] viable/strict/1761933080 -> viable/strict/1761933080 2025-12-04T08:57:04.3930135Z * [new tag] viable/strict/1761935217 -> viable/strict/1761935217 2025-12-04T08:57:04.3931465Z * [new tag] viable/strict/1761938533 -> viable/strict/1761938533 2025-12-04T08:57:04.3932866Z * [new tag] viable/strict/1761940184 -> viable/strict/1761940184 2025-12-04T08:57:04.3934458Z * [new tag] viable/strict/1761942338 -> viable/strict/1761942338 2025-12-04T08:57:04.3935737Z * [new tag] viable/strict/1761946100 -> viable/strict/1761946100 2025-12-04T08:57:04.3936993Z * [new tag] viable/strict/1761947374 -> viable/strict/1761947374 2025-12-04T08:57:04.3938323Z * [new tag] viable/strict/1761950978 -> viable/strict/1761950978 2025-12-04T08:57:04.3939680Z * [new tag] viable/strict/1761957727 -> viable/strict/1761957727 2025-12-04T08:57:04.3941039Z * [new tag] viable/strict/1761959532 -> viable/strict/1761959532 2025-12-04T08:57:04.3942627Z * [new tag] viable/strict/1761965366 -> viable/strict/1761965366 2025-12-04T08:57:04.3944005Z * [new tag] viable/strict/1761968066 -> viable/strict/1761968066 2025-12-04T08:57:04.3945260Z * [new tag] viable/strict/1761969322 -> viable/strict/1761969322 2025-12-04T08:57:04.3946552Z * [new tag] viable/strict/1761974723 -> viable/strict/1761974723 2025-12-04T08:57:04.3947990Z * [new tag] viable/strict/1761981837 -> viable/strict/1761981837 2025-12-04T08:57:04.3949352Z * [new tag] viable/strict/1761985546 -> viable/strict/1761985546 2025-12-04T08:57:04.3950726Z * [new tag] viable/strict/1761987030 -> viable/strict/1761987030 2025-12-04T08:57:04.3952117Z * [new tag] viable/strict/1762003554 -> viable/strict/1762003554 2025-12-04T08:57:04.3953447Z * [new tag] viable/strict/1762021560 -> viable/strict/1762021560 2025-12-04T08:57:04.3954800Z * [new tag] viable/strict/1762032190 -> viable/strict/1762032190 2025-12-04T08:57:04.3956154Z * [new tag] viable/strict/1762040981 -> viable/strict/1762040981 2025-12-04T08:57:04.3957497Z * [new tag] viable/strict/1762048525 -> viable/strict/1762048525 2025-12-04T08:57:04.3958880Z * [new tag] viable/strict/1762104223 -> viable/strict/1762104223 2025-12-04T08:57:04.3960180Z * [new tag] viable/strict/1762105778 -> viable/strict/1762105778 2025-12-04T08:57:04.3961517Z * [new tag] viable/strict/1762115109 -> viable/strict/1762115109 2025-12-04T08:57:04.3962923Z * [new tag] viable/strict/1762125840 -> viable/strict/1762125840 2025-12-04T08:57:04.3964169Z * [new tag] viable/strict/1762127377 -> viable/strict/1762127377 2025-12-04T08:57:04.3965766Z * [new tag] viable/strict/1762134925 -> viable/strict/1762134925 2025-12-04T08:57:04.3967010Z * [new tag] viable/strict/1762138338 -> viable/strict/1762138338 2025-12-04T08:57:04.3968368Z * [new tag] viable/strict/1762148993 -> viable/strict/1762148993 2025-12-04T08:57:04.3969735Z * [new tag] viable/strict/1762152871 -> viable/strict/1762152871 2025-12-04T08:57:04.3971080Z * [new tag] viable/strict/1762156183 -> viable/strict/1762156183 2025-12-04T08:57:04.3972452Z * [new tag] viable/strict/1762163457 -> viable/strict/1762163457 2025-12-04T08:57:04.3973839Z * [new tag] viable/strict/1762165569 -> viable/strict/1762165569 2025-12-04T08:57:04.3975339Z * [new tag] viable/strict/1762169035 -> viable/strict/1762169035 2025-12-04T08:57:04.3976723Z * [new tag] viable/strict/1762174936 -> viable/strict/1762174936 2025-12-04T08:57:04.3978042Z * [new tag] viable/strict/1762194412 -> viable/strict/1762194412 2025-12-04T08:57:04.3979348Z * [new tag] viable/strict/1762195876 -> viable/strict/1762195876 2025-12-04T08:57:04.3980672Z * [new tag] viable/strict/1762197788 -> viable/strict/1762197788 2025-12-04T08:57:04.3982019Z * [new tag] viable/strict/1762199389 -> viable/strict/1762199389 2025-12-04T08:57:04.3983546Z * [new tag] viable/strict/1762206585 -> viable/strict/1762206585 2025-12-04T08:57:04.3984882Z * [new tag] viable/strict/1762210184 -> viable/strict/1762210184 2025-12-04T08:57:04.3986086Z * [new tag] viable/strict/1762218736 -> viable/strict/1762218736 2025-12-04T08:57:04.3987424Z * [new tag] viable/strict/1762224529 -> viable/strict/1762224529 2025-12-04T08:57:04.3988842Z * [new tag] viable/strict/1762227253 -> viable/strict/1762227253 2025-12-04T08:57:04.3990049Z * [new tag] viable/strict/1762228515 -> viable/strict/1762228515 2025-12-04T08:57:04.3991542Z * [new tag] viable/strict/1762230349 -> viable/strict/1762230349 2025-12-04T08:57:04.3992814Z * [new tag] viable/strict/1762231859 -> viable/strict/1762231859 2025-12-04T08:57:04.3994288Z * [new tag] viable/strict/1762233925 -> viable/strict/1762233925 2025-12-04T08:57:04.3995692Z * [new tag] viable/strict/1762237630 -> viable/strict/1762237630 2025-12-04T08:57:04.3996903Z * [new tag] viable/strict/1762253522 -> viable/strict/1762253522 2025-12-04T08:57:04.3998765Z * [new tag] viable/strict/1762278588 -> viable/strict/1762278588 2025-12-04T08:57:04.4000104Z * [new tag] viable/strict/1762284203 -> viable/strict/1762284203 2025-12-04T08:57:04.4001489Z * [new tag] viable/strict/1762289446 -> viable/strict/1762289446 2025-12-04T08:57:04.4002809Z * [new tag] viable/strict/1762291515 -> viable/strict/1762291515 2025-12-04T08:57:04.4004166Z * [new tag] viable/strict/1762295100 -> viable/strict/1762295100 2025-12-04T08:57:04.4005649Z * [new tag] viable/strict/1762296590 -> viable/strict/1762296590 2025-12-04T08:57:04.4006973Z * [new tag] viable/strict/1762300179 -> viable/strict/1762300179 2025-12-04T08:57:04.4008233Z * [new tag] viable/strict/1762303207 -> viable/strict/1762303207 2025-12-04T08:57:04.4009595Z * [new tag] viable/strict/1762386584 -> viable/strict/1762386584 2025-12-04T08:57:04.4010948Z * [new tag] viable/strict/1762391537 -> viable/strict/1762391537 2025-12-04T08:57:04.4012149Z * [new tag] viable/strict/1762394119 -> viable/strict/1762394119 2025-12-04T08:57:04.4013846Z * [new tag] viable/strict/1762397437 -> viable/strict/1762397437 2025-12-04T08:57:04.4015198Z * [new tag] viable/strict/1762400256 -> viable/strict/1762400256 2025-12-04T08:57:04.4016514Z * [new tag] viable/strict/1762401469 -> viable/strict/1762401469 2025-12-04T08:57:04.4017898Z * [new tag] viable/strict/1762408195 -> viable/strict/1762408195 2025-12-04T08:57:04.4019255Z * [new tag] viable/strict/1762410411 -> viable/strict/1762410411 2025-12-04T08:57:04.4020693Z * [new tag] viable/strict/1762417613 -> viable/strict/1762417613 2025-12-04T08:57:04.4022063Z * [new tag] viable/strict/1762419198 -> viable/strict/1762419198 2025-12-04T08:57:04.4023403Z * [new tag] viable/strict/1762422656 -> viable/strict/1762422656 2025-12-04T08:57:04.4025034Z * [new tag] viable/strict/1762424746 -> viable/strict/1762424746 2025-12-04T08:57:04.4026473Z * [new tag] viable/strict/1762446386 -> viable/strict/1762446386 2025-12-04T08:57:04.4027751Z * [new tag] viable/strict/1762449912 -> viable/strict/1762449912 2025-12-04T08:57:04.4029114Z * [new tag] viable/strict/1762457031 -> viable/strict/1762457031 2025-12-04T08:57:04.4030462Z * [new tag] viable/strict/1762462441 -> viable/strict/1762462441 2025-12-04T08:57:04.4031840Z * [new tag] viable/strict/1762467909 -> viable/strict/1762467909 2025-12-04T08:57:04.4033190Z * [new tag] viable/strict/1762471493 -> viable/strict/1762471493 2025-12-04T08:57:04.4034496Z * [new tag] viable/strict/1762475990 -> viable/strict/1762475990 2025-12-04T08:57:04.4035959Z * [new tag] viable/strict/1762477933 -> viable/strict/1762477933 2025-12-04T08:57:04.4037333Z * [new tag] viable/strict/1762491053 -> viable/strict/1762491053 2025-12-04T08:57:04.4038637Z * [new tag] viable/strict/1762493118 -> viable/strict/1762493118 2025-12-04T08:57:04.4039965Z * [new tag] viable/strict/1762498442 -> viable/strict/1762498442 2025-12-04T08:57:04.4041465Z * [new tag] viable/strict/1762501778 -> viable/strict/1762501778 2025-12-04T08:57:04.4042721Z * [new tag] viable/strict/1762504001 -> viable/strict/1762504001 2025-12-04T08:57:04.4044318Z * [new tag] viable/strict/1762505583 -> viable/strict/1762505583 2025-12-04T08:57:04.4045931Z * [new tag] viable/strict/1762507523 -> viable/strict/1762507523 2025-12-04T08:57:04.4047305Z * [new tag] viable/strict/1762511140 -> viable/strict/1762511140 2025-12-04T08:57:04.4048771Z * [new tag] viable/strict/1762512632 -> viable/strict/1762512632 2025-12-04T08:57:04.4050134Z * [new tag] viable/strict/1762520467 -> viable/strict/1762520467 2025-12-04T08:57:04.4051541Z * [new tag] viable/strict/1762522016 -> viable/strict/1762522016 2025-12-04T08:57:04.4052880Z * [new tag] viable/strict/1762530591 -> viable/strict/1762530591 2025-12-04T08:57:04.4054408Z * [new tag] viable/strict/1762543405 -> viable/strict/1762543405 2025-12-04T08:57:04.4055615Z * [new tag] viable/strict/1762544998 -> viable/strict/1762544998 2025-12-04T08:57:04.4056951Z * [new tag] viable/strict/1762552182 -> viable/strict/1762552182 2025-12-04T08:57:04.4058264Z * [new tag] viable/strict/1762554297 -> viable/strict/1762554297 2025-12-04T08:57:04.4059482Z * [new tag] viable/strict/1762559381 -> viable/strict/1762559381 2025-12-04T08:57:04.4060815Z * [new tag] viable/strict/1762562222 -> viable/strict/1762562222 2025-12-04T08:57:04.4062172Z * [new tag] viable/strict/1762564319 -> viable/strict/1762564319 2025-12-04T08:57:04.4063346Z * [new tag] viable/strict/1762566904 -> viable/strict/1762566904 2025-12-04T08:57:04.4064711Z * [new tag] viable/strict/1762569781 -> viable/strict/1762569781 2025-12-04T08:57:04.4066031Z * [new tag] viable/strict/1762575940 -> viable/strict/1762575940 2025-12-04T08:57:04.4067357Z * [new tag] viable/strict/1762580974 -> viable/strict/1762580974 2025-12-04T08:57:04.4068688Z * [new tag] viable/strict/1762583185 -> viable/strict/1762583185 2025-12-04T08:57:04.4070019Z * [new tag] viable/strict/1762586647 -> viable/strict/1762586647 2025-12-04T08:57:04.4071402Z * [new tag] viable/strict/1762588183 -> viable/strict/1762588183 2025-12-04T08:57:04.4072751Z * [new tag] viable/strict/1762593886 -> viable/strict/1762593886 2025-12-04T08:57:04.4074156Z * [new tag] viable/strict/1762650743 -> viable/strict/1762650743 2025-12-04T08:57:04.4075673Z * [new tag] viable/strict/1762653328 -> viable/strict/1762653328 2025-12-04T08:57:04.4076982Z * [new tag] viable/strict/1762659342 -> viable/strict/1762659342 2025-12-04T08:57:04.4078338Z * [new tag] viable/strict/1762662360 -> viable/strict/1762662360 2025-12-04T08:57:04.4079664Z * [new tag] viable/strict/1762667377 -> viable/strict/1762667377 2025-12-04T08:57:04.4081016Z * [new tag] viable/strict/1762671090 -> viable/strict/1762671090 2025-12-04T08:57:04.4082322Z * [new tag] viable/strict/1762680284 -> viable/strict/1762680284 2025-12-04T08:57:04.4083671Z * [new tag] viable/strict/1762683900 -> viable/strict/1762683900 2025-12-04T08:57:04.4084998Z * [new tag] viable/strict/1762705541 -> viable/strict/1762705541 2025-12-04T08:57:04.4086803Z * [new tag] viable/strict/1762709004 -> viable/strict/1762709004 2025-12-04T08:57:04.4088150Z * [new tag] viable/strict/1762746004 -> viable/strict/1762746004 2025-12-04T08:57:04.4089698Z * [new tag] viable/strict/1762748799 -> viable/strict/1762748799 2025-12-04T08:57:04.4091102Z * [new tag] viable/strict/1762759504 -> viable/strict/1762759504 2025-12-04T08:57:04.4092401Z * [new tag] viable/strict/1762760973 -> viable/strict/1762760973 2025-12-04T08:57:04.4093795Z * [new tag] viable/strict/1762775374 -> viable/strict/1762775374 2025-12-04T08:57:04.4095247Z * [new tag] viable/strict/1762777661 -> viable/strict/1762777661 2025-12-04T08:57:04.4096589Z * [new tag] viable/strict/1762779774 -> viable/strict/1762779774 2025-12-04T08:57:04.4098095Z * [new tag] viable/strict/1762781259 -> viable/strict/1762781259 2025-12-04T08:57:04.4099455Z * [new tag] viable/strict/1762793628 -> viable/strict/1762793628 2025-12-04T08:57:04.4101004Z * [new tag] viable/strict/1762800711 -> viable/strict/1762800711 2025-12-04T08:57:04.4102412Z * [new tag] viable/strict/1762809894 -> viable/strict/1762809894 2025-12-04T08:57:04.4103760Z * [new tag] viable/strict/1762811384 -> viable/strict/1762811384 2025-12-04T08:57:04.4105160Z * [new tag] viable/strict/1762813841 -> viable/strict/1762813841 2025-12-04T08:57:04.4106797Z * [new tag] viable/strict/1762815047 -> viable/strict/1762815047 2025-12-04T08:57:04.4108323Z * [new tag] viable/strict/1762817094 -> viable/strict/1762817094 2025-12-04T08:57:04.4109627Z * [new tag] viable/strict/1762818582 -> viable/strict/1762818582 2025-12-04T08:57:04.4110982Z * [new tag] viable/strict/1762821623 -> viable/strict/1762821623 2025-12-04T08:57:04.4112188Z * [new tag] viable/strict/1762823531 -> viable/strict/1762823531 2025-12-04T08:57:04.4113616Z * [new tag] viable/strict/1762849583 -> viable/strict/1762849583 2025-12-04T08:57:04.4114900Z * [new tag] viable/strict/1762851200 -> viable/strict/1762851200 2025-12-04T08:57:04.4116278Z * [new tag] viable/strict/1762854603 -> viable/strict/1762854603 2025-12-04T08:57:04.4117652Z * [new tag] viable/strict/1762858276 -> viable/strict/1762858276 2025-12-04T08:57:04.4119073Z * [new tag] viable/strict/1762860891 -> viable/strict/1762860891 2025-12-04T08:57:04.4120843Z * [new tag] viable/strict/1762866174 -> viable/strict/1762866174 2025-12-04T08:57:04.4122208Z * [new tag] viable/strict/1762867653 -> viable/strict/1762867653 2025-12-04T08:57:04.4123543Z * [new tag] viable/strict/1762872669 -> viable/strict/1762872669 2025-12-04T08:57:04.4124813Z * [new tag] viable/strict/1762878380 -> viable/strict/1762878380 2025-12-04T08:57:04.4126146Z * [new tag] viable/strict/1762889003 -> viable/strict/1762889003 2025-12-04T08:57:04.4127529Z * [new tag] viable/strict/1762890589 -> viable/strict/1762890589 2025-12-04T08:57:04.4128893Z * [new tag] viable/strict/1762892743 -> viable/strict/1762892743 2025-12-04T08:57:04.4130283Z * [new tag] viable/strict/1762894271 -> viable/strict/1762894271 2025-12-04T08:57:04.4131466Z * [new tag] viable/strict/1762896287 -> viable/strict/1762896287 2025-12-04T08:57:04.4132822Z * [new tag] viable/strict/1762915871 -> viable/strict/1762915871 2025-12-04T08:57:04.4134327Z * [new tag] viable/strict/1762918569 -> viable/strict/1762918569 2025-12-04T08:57:04.4135593Z * [new tag] viable/strict/1762919776 -> viable/strict/1762919776 2025-12-04T08:57:04.4136941Z * [new tag] viable/strict/1762923072 -> viable/strict/1762923072 2025-12-04T08:57:04.4138281Z * [new tag] viable/strict/1762928826 -> viable/strict/1762928826 2025-12-04T08:57:04.4139781Z * [new tag] viable/strict/1762930451 -> viable/strict/1762930451 2025-12-04T08:57:04.4141259Z * [new tag] viable/strict/1762933780 -> viable/strict/1762933780 2025-12-04T08:57:04.4142536Z * [new tag] viable/strict/1762937638 -> viable/strict/1762937638 2025-12-04T08:57:04.4144058Z * [new tag] viable/strict/1762939545 -> viable/strict/1762939545 2025-12-04T08:57:04.4145381Z * [new tag] viable/strict/1762962692 -> viable/strict/1762962692 2025-12-04T08:57:04.4146769Z * [new tag] viable/strict/1762979143 -> viable/strict/1762979143 2025-12-04T08:57:04.4148101Z * [new tag] viable/strict/1762984188 -> viable/strict/1762984188 2025-12-04T08:57:04.4149320Z * [new tag] viable/strict/1762986306 -> viable/strict/1762986306 2025-12-04T08:57:04.4150668Z * [new tag] viable/strict/1762989903 -> viable/strict/1762989903 2025-12-04T08:57:04.4152007Z * [new tag] viable/strict/1762991377 -> viable/strict/1762991377 2025-12-04T08:57:04.4153360Z * [new tag] viable/strict/1762998921 -> viable/strict/1762998921 2025-12-04T08:57:04.4154773Z * [new tag] viable/strict/1763002287 -> viable/strict/1763002287 2025-12-04T08:57:04.4156141Z * [new tag] viable/strict/1763016840 -> viable/strict/1763016840 2025-12-04T08:57:04.4157462Z * [new tag] viable/strict/1763020180 -> viable/strict/1763020180 2025-12-04T08:57:04.4158839Z * [new tag] viable/strict/1763027421 -> viable/strict/1763027421 2025-12-04T08:57:04.4160212Z * [new tag] viable/strict/1763031120 -> viable/strict/1763031120 2025-12-04T08:57:04.4161597Z * [new tag] viable/strict/1763036861 -> viable/strict/1763036861 2025-12-04T08:57:04.4162958Z * [new tag] viable/strict/1763038993 -> viable/strict/1763038993 2025-12-04T08:57:04.4164351Z * [new tag] viable/strict/1763054703 -> viable/strict/1763054703 2025-12-04T08:57:04.4165559Z * [new tag] viable/strict/1763067061 -> viable/strict/1763067061 2025-12-04T08:57:04.4166942Z * [new tag] viable/strict/1763070847 -> viable/strict/1763070847 2025-12-04T08:57:04.4168290Z * [new tag] viable/strict/1763072706 -> viable/strict/1763072706 2025-12-04T08:57:04.4169682Z * [new tag] viable/strict/1763076302 -> viable/strict/1763076302 2025-12-04T08:57:04.4171004Z * [new tag] viable/strict/1763080816 -> viable/strict/1763080816 2025-12-04T08:57:04.4172371Z * [new tag] viable/strict/1763082732 -> viable/strict/1763082732 2025-12-04T08:57:04.4174026Z * [new tag] viable/strict/1763085329 -> viable/strict/1763085329 2025-12-04T08:57:04.4175931Z * [new tag] viable/strict/1763088623 -> viable/strict/1763088623 2025-12-04T08:57:04.4177345Z * [new tag] viable/strict/1763091402 -> viable/strict/1763091402 2025-12-04T08:57:04.4178721Z * [new tag] viable/strict/1763092602 -> viable/strict/1763092602 2025-12-04T08:57:04.4180098Z * [new tag] viable/strict/1763094355 -> viable/strict/1763094355 2025-12-04T08:57:04.4181462Z * [new tag] viable/strict/1763099390 -> viable/strict/1763099390 2025-12-04T08:57:04.4182803Z * [new tag] viable/strict/1763101608 -> viable/strict/1763101608 2025-12-04T08:57:04.4184250Z * [new tag] viable/strict/1763105102 -> viable/strict/1763105102 2025-12-04T08:57:04.4185646Z * [new tag] viable/strict/1763112347 -> viable/strict/1763112347 2025-12-04T08:57:04.4187039Z * [new tag] viable/strict/1763119471 -> viable/strict/1763119471 2025-12-04T08:57:04.4188263Z * [new tag] viable/strict/1763126835 -> viable/strict/1763126835 2025-12-04T08:57:04.4189473Z * [new tag] viable/strict/1763149779 -> viable/strict/1763149779 2025-12-04T08:57:04.4190966Z * [new tag] viable/strict/1763164178 -> viable/strict/1763164178 2025-12-04T08:57:04.4192226Z * [new tag] viable/strict/1763167104 -> viable/strict/1763167104 2025-12-04T08:57:04.4193555Z * [new tag] viable/strict/1763169132 -> viable/strict/1763169132 2025-12-04T08:57:04.4194924Z * [new tag] viable/strict/1763171708 -> viable/strict/1763171708 2025-12-04T08:57:04.4196205Z * [new tag] viable/strict/1763174759 -> viable/strict/1763174759 2025-12-04T08:57:04.4197630Z * [new tag] viable/strict/1763180744 -> viable/strict/1763180744 2025-12-04T08:57:04.4198981Z * [new tag] viable/strict/1763182227 -> viable/strict/1763182227 2025-12-04T08:57:04.4200343Z * [new tag] viable/strict/1763184309 -> viable/strict/1763184309 2025-12-04T08:57:04.4202052Z * [new tag] viable/strict/1763187991 -> viable/strict/1763187991 2025-12-04T08:57:04.4203374Z * [new tag] viable/strict/1763191445 -> viable/strict/1763191445 2025-12-04T08:57:04.4205084Z * [new tag] viable/strict/1763195152 -> viable/strict/1763195152 2025-12-04T08:57:04.4207790Z * [new tag] viable/strict/1763205769 -> viable/strict/1763205769 2025-12-04T08:57:04.4209146Z * [new tag] viable/strict/1763246990 -> viable/strict/1763246990 2025-12-04T08:57:04.4210609Z * [new tag] viable/strict/1763261578 -> viable/strict/1763261578 2025-12-04T08:57:04.4211822Z * [new tag] viable/strict/1763286573 -> viable/strict/1763286573 2025-12-04T08:57:04.4213090Z * [new tag] viable/strict/1763292167 -> viable/strict/1763292167 2025-12-04T08:57:04.4214608Z * [new tag] viable/strict/1763333386 -> viable/strict/1763333386 2025-12-04T08:57:04.4215915Z * [new tag] viable/strict/1763340082 -> viable/strict/1763340082 2025-12-04T08:57:04.4217811Z * [new tag] viable/strict/1763364324 -> viable/strict/1763364324 2025-12-04T08:57:04.4219191Z * [new tag] viable/strict/1763371569 -> viable/strict/1763371569 2025-12-04T08:57:04.4220577Z * [new tag] viable/strict/1763373067 -> viable/strict/1763373067 2025-12-04T08:57:04.4221875Z * [new tag] viable/strict/1763375157 -> viable/strict/1763375157 2025-12-04T08:57:04.4223257Z * [new tag] viable/strict/1763382462 -> viable/strict/1763382462 2025-12-04T08:57:04.4224644Z * [new tag] viable/strict/1763394661 -> viable/strict/1763394661 2025-12-04T08:57:04.4226197Z * [new tag] viable/strict/1763396797 -> viable/strict/1763396797 2025-12-04T08:57:04.4227592Z * [new tag] viable/strict/1763398542 -> viable/strict/1763398542 2025-12-04T08:57:04.4228932Z * [new tag] viable/strict/1763401807 -> viable/strict/1763401807 2025-12-04T08:57:04.4230271Z * [new tag] viable/strict/1763414698 -> viable/strict/1763414698 2025-12-04T08:57:04.4231607Z * [new tag] viable/strict/1763419807 -> viable/strict/1763419807 2025-12-04T08:57:04.4232943Z * [new tag] viable/strict/1763426369 -> viable/strict/1763426369 2025-12-04T08:57:04.4234421Z * [new tag] viable/strict/1763428331 -> viable/strict/1763428331 2025-12-04T08:57:04.4235818Z * [new tag] viable/strict/1763430922 -> viable/strict/1763430922 2025-12-04T08:57:04.4237014Z * [new tag] viable/strict/1763434184 -> viable/strict/1763434184 2025-12-04T08:57:04.4238371Z * [new tag] viable/strict/1763439973 -> viable/strict/1763439973 2025-12-04T08:57:04.4239743Z * [new tag] viable/strict/1763444995 -> viable/strict/1763444995 2025-12-04T08:57:04.4241287Z * [new tag] viable/strict/1763447206 -> viable/strict/1763447206 2025-12-04T08:57:04.4242545Z * [new tag] viable/strict/1763448826 -> viable/strict/1763448826 2025-12-04T08:57:04.4243921Z * [new tag] viable/strict/1763450717 -> viable/strict/1763450717 2025-12-04T08:57:04.4245229Z * [new tag] viable/strict/1763452183 -> viable/strict/1763452183 2025-12-04T08:57:04.4246671Z * [new tag] viable/strict/1763457945 -> viable/strict/1763457945 2025-12-04T08:57:04.4248003Z * [new tag] viable/strict/1763459439 -> viable/strict/1763459439 2025-12-04T08:57:04.4249237Z * [new tag] viable/strict/1763461556 -> viable/strict/1763461556 2025-12-04T08:57:04.4250612Z * [new tag] viable/strict/1763463103 -> viable/strict/1763463103 2025-12-04T08:57:04.4252008Z * [new tag] viable/strict/1763465100 -> viable/strict/1763465100 2025-12-04T08:57:04.4253292Z * [new tag] viable/strict/1763468866 -> viable/strict/1763468866 2025-12-04T08:57:04.4254555Z * [new tag] viable/strict/1763493823 -> viable/strict/1763493823 2025-12-04T08:57:04.4255783Z * [new tag] viable/strict/1763496249 -> viable/strict/1763496249 2025-12-04T08:57:04.4257280Z * [new tag] viable/strict/1763502620 -> viable/strict/1763502620 2025-12-04T08:57:04.4258666Z * [new tag] viable/strict/1763504715 -> viable/strict/1763504715 2025-12-04T08:57:04.4260061Z * [new tag] viable/strict/1763506208 -> viable/strict/1763506208 2025-12-04T08:57:04.4261415Z * [new tag] viable/strict/1763520590 -> viable/strict/1763520590 2025-12-04T08:57:04.4262815Z * [new tag] viable/strict/1763523357 -> viable/strict/1763523357 2025-12-04T08:57:04.4264212Z * [new tag] viable/strict/1763529922 -> viable/strict/1763529922 2025-12-04T08:57:04.4266067Z * [new tag] viable/strict/1763531408 -> viable/strict/1763531408 2025-12-04T08:57:04.4267442Z * [new tag] viable/strict/1763533622 -> viable/strict/1763533622 2025-12-04T08:57:04.4268806Z * [new tag] viable/strict/1763538576 -> viable/strict/1763538576 2025-12-04T08:57:04.4270243Z * [new tag] viable/strict/1763545823 -> viable/strict/1763545823 2025-12-04T08:57:04.4271458Z * [new tag] viable/strict/1763547951 -> viable/strict/1763547951 2025-12-04T08:57:04.4272905Z * [new tag] viable/strict/1763551477 -> viable/strict/1763551477 2025-12-04T08:57:04.4274292Z * [new tag] viable/strict/1763552982 -> viable/strict/1763552982 2025-12-04T08:57:04.4275633Z * [new tag] viable/strict/1763594698 -> viable/strict/1763594698 2025-12-04T08:57:04.4277053Z * [new tag] viable/strict/1763596178 -> viable/strict/1763596178 2025-12-04T08:57:04.4278460Z * [new tag] viable/strict/1763599155 -> viable/strict/1763599155 2025-12-04T08:57:04.4279763Z * [new tag] viable/strict/1763603717 -> viable/strict/1763603717 2025-12-04T08:57:04.4281184Z * [new tag] viable/strict/1763606923 -> viable/strict/1763606923 2025-12-04T08:57:04.4282528Z * [new tag] viable/strict/1763609715 -> viable/strict/1763609715 2025-12-04T08:57:04.4283964Z * [new tag] viable/strict/1763612757 -> viable/strict/1763612757 2025-12-04T08:57:04.4285329Z * [new tag] viable/strict/1763616325 -> viable/strict/1763616325 2025-12-04T08:57:04.4286668Z * [new tag] viable/strict/1763623509 -> viable/strict/1763623509 2025-12-04T08:57:04.4288154Z * [new tag] viable/strict/1763624984 -> viable/strict/1763624984 2025-12-04T08:57:04.4289510Z * [new tag] viable/strict/1763628796 -> viable/strict/1763628796 2025-12-04T08:57:04.4290926Z * [new tag] viable/strict/1763634343 -> viable/strict/1763634343 2025-12-04T08:57:04.4292180Z * [new tag] viable/strict/1763635867 -> viable/strict/1763635867 2025-12-04T08:57:04.4293910Z * [new tag] viable/strict/1763639382 -> viable/strict/1763639382 2025-12-04T08:57:04.4295563Z * [new tag] viable/strict/1763646626 -> viable/strict/1763646626 2025-12-04T08:57:04.4297015Z * [new tag] viable/strict/1763655997 -> viable/strict/1763655997 2025-12-04T08:57:04.4298476Z * [new tag] viable/strict/1763659444 -> viable/strict/1763659444 2025-12-04T08:57:04.4299805Z * [new tag] viable/strict/1763660992 -> viable/strict/1763660992 2025-12-04T08:57:04.4301148Z * [new tag] viable/strict/1763663201 -> viable/strict/1763663201 2025-12-04T08:57:04.4302522Z * [new tag] viable/strict/1763670362 -> viable/strict/1763670362 2025-12-04T08:57:04.4303857Z * [new tag] viable/strict/1763675378 -> viable/strict/1763675378 2025-12-04T08:57:04.4305630Z * [new tag] viable/strict/1763693343 -> viable/strict/1763693343 2025-12-04T08:57:04.4307005Z * [new tag] viable/strict/1763696088 -> viable/strict/1763696088 2025-12-04T08:57:04.4308528Z * [new tag] viable/strict/1763697343 -> viable/strict/1763697343 2025-12-04T08:57:04.4309862Z * [new tag] viable/strict/1763699165 -> viable/strict/1763699165 2025-12-04T08:57:04.4311198Z * [new tag] viable/strict/1763700660 -> viable/strict/1763700660 2025-12-04T08:57:04.4312535Z * [new tag] viable/strict/1763704209 -> viable/strict/1763704209 2025-12-04T08:57:04.4313908Z * [new tag] viable/strict/1763706411 -> viable/strict/1763706411 2025-12-04T08:57:04.4315223Z * [new tag] viable/strict/1763708082 -> viable/strict/1763708082 2025-12-04T08:57:04.4316524Z * [new tag] viable/strict/1763711381 -> viable/strict/1763711381 2025-12-04T08:57:04.4317817Z * [new tag] viable/strict/1763713593 -> viable/strict/1763713593 2025-12-04T08:57:04.4319201Z * [new tag] viable/strict/1763715201 -> viable/strict/1763715201 2025-12-04T08:57:04.4320535Z * [new tag] viable/strict/1763733017 -> viable/strict/1763733017 2025-12-04T08:57:04.4321928Z * [new tag] viable/strict/1763735108 -> viable/strict/1763735108 2025-12-04T08:57:04.4323264Z * [new tag] viable/strict/1763749579 -> viable/strict/1763749579 2025-12-04T08:57:04.4324555Z * [new tag] viable/strict/1763751113 -> viable/strict/1763751113 2025-12-04T08:57:04.4325891Z * [new tag] viable/strict/1763753035 -> viable/strict/1763753035 2025-12-04T08:57:04.4327238Z * [new tag] viable/strict/1763754578 -> viable/strict/1763754578 2025-12-04T08:57:04.4328679Z * [new tag] viable/strict/1763756748 -> viable/strict/1763756748 2025-12-04T08:57:04.4329992Z * [new tag] viable/strict/1763758205 -> viable/strict/1763758205 2025-12-04T08:57:04.4331231Z * [new tag] viable/strict/1763764050 -> viable/strict/1763764050 2025-12-04T08:57:04.4332592Z * [new tag] viable/strict/1763771887 -> viable/strict/1763771887 2025-12-04T08:57:04.4334193Z * [new tag] viable/strict/1763773920 -> viable/strict/1763773920 2025-12-04T08:57:04.4335497Z * [new tag] viable/strict/1763776501 -> viable/strict/1763776501 2025-12-04T08:57:04.4336850Z * [new tag] viable/strict/1763779437 -> viable/strict/1763779437 2025-12-04T08:57:04.4338341Z * [new tag] viable/strict/1763781038 -> viable/strict/1763781038 2025-12-04T08:57:04.4339561Z * [new tag] viable/strict/1763782245 -> viable/strict/1763782245 2025-12-04T08:57:04.4341100Z * [new tag] viable/strict/1763785568 -> viable/strict/1763785568 2025-12-04T08:57:04.4342330Z * [new tag] viable/strict/1763787006 -> viable/strict/1763787006 2025-12-04T08:57:04.4343772Z * [new tag] viable/strict/1763789103 -> viable/strict/1763789103 2025-12-04T08:57:04.4345053Z * [new tag] viable/strict/1763790578 -> viable/strict/1763790578 2025-12-04T08:57:04.4346415Z * [new tag] viable/strict/1763796275 -> viable/strict/1763796275 2025-12-04T08:57:04.4347938Z * [new tag] viable/strict/1763801465 -> viable/strict/1763801465 2025-12-04T08:57:04.4349403Z * [new tag] viable/strict/1763803522 -> viable/strict/1763803522 2025-12-04T08:57:04.4350708Z * [new tag] viable/strict/1763808581 -> viable/strict/1763808581 2025-12-04T08:57:04.4352026Z * [new tag] viable/strict/1763840977 -> viable/strict/1763840977 2025-12-04T08:57:04.4353366Z * [new tag] viable/strict/1763846659 -> viable/strict/1763846659 2025-12-04T08:57:04.4355087Z * [new tag] viable/strict/1763872065 -> viable/strict/1763872065 2025-12-04T08:57:04.4356476Z * [new tag] viable/strict/1763873648 -> viable/strict/1763873648 2025-12-04T08:57:04.4357866Z * [new tag] viable/strict/1763875506 -> viable/strict/1763875506 2025-12-04T08:57:04.4359074Z * [new tag] viable/strict/1763889904 -> viable/strict/1763889904 2025-12-04T08:57:04.4360438Z * [new tag] viable/strict/1763930999 -> viable/strict/1763930999 2025-12-04T08:57:04.4361761Z * [new tag] viable/strict/1763944964 -> viable/strict/1763944964 2025-12-04T08:57:04.4363049Z * [new tag] viable/strict/1763958474 -> viable/strict/1763958474 2025-12-04T08:57:04.4364385Z * [new tag] viable/strict/1763967263 -> viable/strict/1763967263 2025-12-04T08:57:04.4365830Z * [new tag] viable/strict/1763972803 -> viable/strict/1763972803 2025-12-04T08:57:04.4367107Z * [new tag] viable/strict/1763976376 -> viable/strict/1763976376 2025-12-04T08:57:04.4368425Z * [new tag] viable/strict/1763989404 -> viable/strict/1763989404 2025-12-04T08:57:04.4369746Z * [new tag] viable/strict/1763990887 -> viable/strict/1763990887 2025-12-04T08:57:04.4371109Z * [new tag] viable/strict/1764019919 -> viable/strict/1764019919 2025-12-04T08:57:04.4372464Z * [new tag] viable/strict/1764023134 -> viable/strict/1764023134 2025-12-04T08:57:04.4373757Z * [new tag] viable/strict/1764024593 -> viable/strict/1764024593 2025-12-04T08:57:04.4375102Z * [new tag] viable/strict/1764026706 -> viable/strict/1764026706 2025-12-04T08:57:04.4376624Z * [new tag] viable/strict/1764031139 -> viable/strict/1764031139 2025-12-04T08:57:04.4377995Z * [new tag] viable/strict/1764033131 -> viable/strict/1764033131 2025-12-04T08:57:04.4379258Z * [new tag] viable/strict/1764035725 -> viable/strict/1764035725 2025-12-04T08:57:04.4380439Z * [new tag] viable/strict/1764624265 -> viable/strict/1764624265 2025-12-04T08:57:04.4381661Z * [new tag] viable/strict/1764631514 -> viable/strict/1764631514 2025-12-04T08:57:04.4382855Z * [new tag] viable/strict/1764632987 -> viable/strict/1764632987 2025-12-04T08:57:04.4384021Z * [new tag] viable/strict/1764636063 -> viable/strict/1764636063 2025-12-04T08:57:04.4385191Z * [new tag] viable/strict/1764643975 -> viable/strict/1764643975 2025-12-04T08:57:04.4386380Z * [new tag] viable/strict/1764646859 -> viable/strict/1764646859 2025-12-04T08:57:04.4387594Z * [new tag] viable/strict/1764653120 -> viable/strict/1764653120 2025-12-04T08:57:04.4388870Z * [new tag] viable/strict/1764654632 -> viable/strict/1764654632 2025-12-04T08:57:04.4390102Z * [new tag] viable/strict/1764656821 -> viable/strict/1764656821 2025-12-04T08:57:04.4391229Z * [new tag] viable/strict/1764658557 -> viable/strict/1764658557 2025-12-04T08:57:04.4392390Z * [new tag] viable/strict/1764660333 -> viable/strict/1764660333 2025-12-04T08:57:04.4393590Z * [new tag] viable/strict/1764661812 -> viable/strict/1764661812 2025-12-04T08:57:04.4394838Z * [new tag] viable/strict/1764664023 -> viable/strict/1764664023 2025-12-04T08:57:04.4396013Z * [new tag] viable/strict/1764669150 -> viable/strict/1764669150 2025-12-04T08:57:04.4397198Z * [new tag] viable/strict/1764680709 -> viable/strict/1764680709 2025-12-04T08:57:04.4398373Z * [new tag] viable/strict/1764687619 -> viable/strict/1764687619 2025-12-04T08:57:04.4399605Z * [new tag] viable/strict/1764696355 -> viable/strict/1764696355 2025-12-04T08:57:04.4400817Z * [new tag] viable/strict/1764701767 -> viable/strict/1764701767 2025-12-04T08:57:04.4401987Z * [new tag] viable/strict/1764710768 -> viable/strict/1764710768 2025-12-04T08:57:04.4403250Z * [new tag] viable/strict/1764716202 -> viable/strict/1764716202 2025-12-04T08:57:04.4404409Z * [new tag] viable/strict/1764793566 -> viable/strict/1764793566 2025-12-04T08:57:04.4405938Z * [new tag] viable/strict/1764797093 -> viable/strict/1764797093 2025-12-04T08:57:04.4407142Z * [new tag] viable/strict/1764800729 -> viable/strict/1764800729 2025-12-04T08:57:04.4408426Z * [new tag] whc_flight_1 -> whc_flight_1 2025-12-04T08:57:04.4409722Z * [new tag] whc_flight_2 -> whc_flight_2 2025-12-04T08:57:04.4411187Z * [new tag] whc_flight_4 -> whc_flight_4 2025-12-04T08:57:04.5249294Z [command]/usr/bin/git rev-parse --verify --quiet ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32^{object} 2025-12-04T08:57:04.5276664Z ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T08:57:04.5281041Z ##[endgroup] 2025-12-04T08:57:04.5281450Z ##[group]Determining the checkout info 2025-12-04T08:57:04.5282316Z ##[endgroup] 2025-12-04T08:57:04.5286138Z [command]/usr/bin/git sparse-checkout disable 2025-12-04T08:57:04.5319093Z [command]/usr/bin/git config --local --unset-all extensions.worktreeConfig 2025-12-04T08:57:04.5346015Z ##[group]Checking out the ref 2025-12-04T08:57:04.5349741Z [command]/usr/bin/git checkout --progress --force ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T08:57:05.5740117Z Updating files: 75% (15202/20121) 2025-12-04T08:57:05.5895064Z Updating files: 76% (15292/20121) 2025-12-04T08:57:05.6043725Z Updating files: 77% (15494/20121) 2025-12-04T08:57:05.6249477Z Updating files: 78% (15695/20121) 2025-12-04T08:57:05.6506633Z Updating files: 79% (15896/20121) 2025-12-04T08:57:05.6812465Z Updating files: 80% (16097/20121) 2025-12-04T08:57:05.7087296Z Updating files: 81% (16299/20121) 2025-12-04T08:57:05.7307108Z Updating files: 82% (16500/20121) 2025-12-04T08:57:05.7470095Z Updating files: 83% (16701/20121) 2025-12-04T08:57:05.7622253Z Updating files: 84% (16902/20121) 2025-12-04T08:57:05.7794941Z Updating files: 85% (17103/20121) 2025-12-04T08:57:05.7961892Z Updating files: 86% (17305/20121) 2025-12-04T08:57:05.8116025Z Updating files: 87% (17506/20121) 2025-12-04T08:57:05.8244956Z Updating files: 88% (17707/20121) 2025-12-04T08:57:05.8395400Z Updating files: 89% (17908/20121) 2025-12-04T08:57:05.8577216Z Updating files: 90% (18109/20121) 2025-12-04T08:57:05.8710202Z Updating files: 91% (18311/20121) 2025-12-04T08:57:05.8875120Z Updating files: 92% (18512/20121) 2025-12-04T08:57:05.9065443Z Updating files: 93% (18713/20121) 2025-12-04T08:57:05.9272103Z Updating files: 94% (18914/20121) 2025-12-04T08:57:05.9454758Z Updating files: 95% (19115/20121) 2025-12-04T08:57:05.9625011Z Updating files: 96% (19317/20121) 2025-12-04T08:57:05.9796591Z Updating files: 97% (19518/20121) 2025-12-04T08:57:06.0071841Z Updating files: 98% (19719/20121) 2025-12-04T08:57:06.0254047Z Updating files: 99% (19920/20121) 2025-12-04T08:57:06.0254340Z Updating files: 100% (20121/20121) 2025-12-04T08:57:06.0254630Z Updating files: 100% (20121/20121), done. 2025-12-04T08:57:06.0487096Z Note: switching to 'ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32'. 2025-12-04T08:57:06.0487397Z 2025-12-04T08:57:06.0487601Z You are in 'detached HEAD' state. You can look around, make experimental 2025-12-04T08:57:06.0488103Z changes and commit them, and you can discard any commits you make in this 2025-12-04T08:57:06.0488587Z state without impacting any branches by switching back to a branch. 2025-12-04T08:57:06.0488865Z 2025-12-04T08:57:06.0489048Z If you want to create a new branch to retain commits you create, you may 2025-12-04T08:57:06.0489524Z do so (now or later) by using -c with the switch command. Example: 2025-12-04T08:57:06.0489814Z 2025-12-04T08:57:06.0489934Z git switch -c 2025-12-04T08:57:06.0490109Z 2025-12-04T08:57:06.0490214Z Or undo this operation with: 2025-12-04T08:57:06.0490375Z 2025-12-04T08:57:06.0490458Z git switch - 2025-12-04T08:57:06.0490592Z 2025-12-04T08:57:06.0490808Z Turn off this advice by setting config variable advice.detachedHead to false 2025-12-04T08:57:06.0491120Z 2025-12-04T08:57:06.0491357Z HEAD is now at ffd9b0fb435 Resolve collective autotuning test failure on arm (#168919) 2025-12-04T08:57:06.0611469Z ##[endgroup] 2025-12-04T08:57:06.0611865Z ##[group]Setting up auth for fetching submodules 2025-12-04T08:57:06.0618312Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-12-04T08:57:06.0669194Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2025-12-04T08:57:06.0696637Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2025-12-04T08:57:06.0722683Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2025-12-04T08:57:06.0746684Z ##[endgroup] 2025-12-04T08:57:06.0747136Z ##[group]Fetching submodules 2025-12-04T08:57:06.0750261Z [command]/usr/bin/git submodule sync --recursive 2025-12-04T08:57:06.1094266Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --recursive 2025-12-04T08:57:06.1428012Z Submodule 'android/libs/fbjni' (https://github.com/facebookincubator/fbjni.git) registered for path 'android/libs/fbjni' 2025-12-04T08:57:06.1430412Z Submodule 'third_party/NNPACK_deps/FP16' (https://github.com/Maratyszcza/FP16.git) registered for path 'third_party/FP16' 2025-12-04T08:57:06.1433797Z Submodule 'third_party/NNPACK_deps/FXdiv' (https://github.com/Maratyszcza/FXdiv.git) registered for path 'third_party/FXdiv' 2025-12-04T08:57:06.1437370Z Submodule 'third_party/NNPACK' (https://github.com/Maratyszcza/NNPACK.git) registered for path 'third_party/NNPACK' 2025-12-04T08:57:06.1440664Z Submodule 'third_party/NVTX' (https://github.com/NVIDIA/NVTX.git) registered for path 'third_party/NVTX' 2025-12-04T08:57:06.1444316Z Submodule 'third_party/VulkanMemoryAllocator' (https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator.git) registered for path 'third_party/VulkanMemoryAllocator' 2025-12-04T08:57:06.1447886Z Submodule 'third_party/XNNPACK' (https://github.com/google/XNNPACK.git) registered for path 'third_party/XNNPACK' 2025-12-04T08:57:06.1451382Z Submodule 'third_party/aiter' (https://github.com/ROCm/aiter.git) registered for path 'third_party/aiter' 2025-12-04T08:57:06.1455236Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/benchmark' 2025-12-04T08:57:06.1459085Z Submodule 'third_party/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/composable_kernel' 2025-12-04T08:57:06.1463021Z Submodule 'third_party/cpp-httplib' (https://github.com/yhirose/cpp-httplib.git) registered for path 'third_party/cpp-httplib' 2025-12-04T08:57:06.1466746Z Submodule 'third_party/cpuinfo' (https://github.com/pytorch/cpuinfo.git) registered for path 'third_party/cpuinfo' 2025-12-04T08:57:06.1471561Z Submodule 'third_party/cudnn_frontend' (https://github.com/NVIDIA/cudnn-frontend.git) registered for path 'third_party/cudnn_frontend' 2025-12-04T08:57:06.1475458Z Submodule 'third_party/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/cutlass' 2025-12-04T08:57:06.1479581Z Submodule 'third_party/fbgemm' (https://github.com/pytorch/fbgemm) registered for path 'third_party/fbgemm' 2025-12-04T08:57:06.1483913Z Submodule 'third_party/flash-attention' (https://github.com/Dao-AILab/flash-attention.git) registered for path 'third_party/flash-attention' 2025-12-04T08:57:06.1491214Z Submodule 'third_party/flatbuffers' (https://github.com/google/flatbuffers.git) registered for path 'third_party/flatbuffers' 2025-12-04T08:57:06.1495441Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/fmt' 2025-12-04T08:57:06.1499892Z Submodule 'third_party/gemmlowp/gemmlowp' (https://github.com/google/gemmlowp.git) registered for path 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:57:06.1504431Z Submodule 'third_party/gloo' (https://github.com/pytorch/gloo) registered for path 'third_party/gloo' 2025-12-04T08:57:06.1509450Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/googletest' 2025-12-04T08:57:06.1513947Z Submodule 'third_party/ideep' (https://github.com/intel/ideep) registered for path 'third_party/ideep' 2025-12-04T08:57:06.1518647Z Submodule 'third_party/ittapi' (https://github.com/intel/ittapi.git) registered for path 'third_party/ittapi' 2025-12-04T08:57:06.1523405Z Submodule 'third_party/kineto' (https://github.com/pytorch/kineto) registered for path 'third_party/kineto' 2025-12-04T08:57:06.1528289Z Submodule 'third_party/kleidiai' (https://github.com/ARM-software/kleidiai.git) registered for path 'third_party/kleidiai' 2025-12-04T08:57:06.1533105Z Submodule 'third_party/mimalloc' (https://github.com/microsoft/mimalloc.git) registered for path 'third_party/mimalloc' 2025-12-04T08:57:06.1538065Z Submodule 'third_party/nlohmann' (https://github.com/nlohmann/json.git) registered for path 'third_party/nlohmann' 2025-12-04T08:57:06.1543013Z Submodule 'third_party/onnx' (https://github.com/onnx/onnx.git) registered for path 'third_party/onnx' 2025-12-04T08:57:06.1548334Z Submodule 'third_party/opentelemetry-cpp' (https://github.com/open-telemetry/opentelemetry-cpp.git) registered for path 'third_party/opentelemetry-cpp' 2025-12-04T08:57:06.1553425Z Submodule 'third_party/pocketfft' (https://github.com/mreineck/pocketfft) registered for path 'third_party/pocketfft' 2025-12-04T08:57:06.1558832Z Submodule 'third_party/protobuf' (https://github.com/protocolbuffers/protobuf.git) registered for path 'third_party/protobuf' 2025-12-04T08:57:06.1564214Z Submodule 'third_party/NNPACK_deps/psimd' (https://github.com/Maratyszcza/psimd.git) registered for path 'third_party/psimd' 2025-12-04T08:57:06.1569722Z Submodule 'third_party/NNPACK_deps/pthreadpool' (https://github.com/Maratyszcza/pthreadpool.git) registered for path 'third_party/pthreadpool' 2025-12-04T08:57:06.1578222Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/pybind11' 2025-12-04T08:57:06.1583876Z Submodule 'third_party/python-peachpy' (https://github.com/malfet/PeachPy.git) registered for path 'third_party/python-peachpy' 2025-12-04T08:57:06.1589786Z Submodule 'third_party/sleef' (https://github.com/shibatch/sleef) registered for path 'third_party/sleef' 2025-12-04T08:57:06.1596392Z Submodule 'third_party/tensorpipe' (https://github.com/pytorch/tensorpipe.git) registered for path 'third_party/tensorpipe' 2025-12-04T08:57:06.1628485Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/android/libs/fbjni'... 2025-12-04T08:57:06.3874374Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FXdiv'... 2025-12-04T08:57:06.3875296Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FP16'... 2025-12-04T08:57:06.3875821Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NNPACK'... 2025-12-04T08:57:06.3902028Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fmt'... 2025-12-04T08:57:09.0173036Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/benchmark'... 2025-12-04T08:57:09.0174490Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NVTX'... 2025-12-04T08:57:09.0175574Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gloo'... 2025-12-04T08:57:09.0176794Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gemmlowp/gemmlowp'... 2025-12-04T08:57:09.0178006Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpuinfo'... 2025-12-04T08:57:09.0179102Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention'... 2025-12-04T08:57:09.0180245Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpp-httplib'... 2025-12-04T08:57:09.0181295Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep'... 2025-12-04T08:57:09.0182407Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ittapi'... 2025-12-04T08:57:09.0183350Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kleidiai'... 2025-12-04T08:57:09.0184286Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pocketfft'... 2025-12-04T08:57:09.0185265Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cudnn_frontend'... 2025-12-04T08:57:09.0186133Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/psimd'... 2025-12-04T08:57:09.0186905Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/googletest'... 2025-12-04T08:57:09.0187864Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pthreadpool'... 2025-12-04T08:57:09.0242383Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flatbuffers'... 2025-12-04T08:57:09.2782065Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/mimalloc'... 2025-12-04T08:57:09.2783839Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/VulkanMemoryAllocator'... 2025-12-04T08:57:09.2785009Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/python-peachpy'... 2025-12-04T08:57:09.3783622Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto'... 2025-12-04T08:57:09.7482182Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe'... 2025-12-04T08:57:09.8483173Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx'... 2025-12-04T08:57:11.9545737Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pybind11'... 2025-12-04T08:57:11.9546614Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/sleef'... 2025-12-04T08:57:11.9547569Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm'... 2025-12-04T08:57:11.9548503Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cutlass'... 2025-12-04T08:57:12.0546833Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/XNNPACK'... 2025-12-04T08:57:27.4614427Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/composable_kernel'... 2025-12-04T08:57:27.4615159Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/nlohmann'... 2025-12-04T08:57:27.4615823Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp'... 2025-12-04T08:57:27.4616803Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter'... 2025-12-04T08:57:27.4618201Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf'... 2025-12-04T08:57:27.4796453Z Submodule path 'android/libs/fbjni': checked out '7e1e1fe3858c63c251c637ae41a20de425dde96f' 2025-12-04T08:57:27.4942623Z Submodule path 'third_party/FP16': checked out '4dfe081cf6bcd15db339cf2680b9281b8451eeb3' 2025-12-04T08:57:27.5057959Z Submodule path 'third_party/FXdiv': checked out 'b408327ac2a15ec3e43352421954f5b1967701d1' 2025-12-04T08:57:27.5352818Z Submodule path 'third_party/NNPACK': checked out 'c07e3a0400713d546e0dea2d5466dd22ea389c73' 2025-12-04T08:57:27.6269426Z Submodule path 'third_party/NVTX': checked out '3ebbc93ded7285963bff932c678fa367eb393ba6' 2025-12-04T08:57:27.6846306Z Submodule path 'third_party/VulkanMemoryAllocator': checked out '1d8f600fd424278486eade7ed3e877c99f0846b1' 2025-12-04T08:57:28.5648935Z Submodule path 'third_party/XNNPACK': checked out '51a0103656eff6fc9bfd39a4597923c4b542c883' 2025-12-04T08:57:28.7573846Z Submodule path 'third_party/aiter': checked out '01aae101b9e5e94d6c16a9514c9fb8df99c93150' 2025-12-04T08:57:28.7596449Z Submodule '3rdparty/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:57:28.7627060Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter/3rdparty/composable_kernel'... 2025-12-04T08:57:33.0712714Z Submodule path 'third_party/aiter/3rdparty/composable_kernel': checked out 'cffe8fa2a442ac8e80dd236a1a5d24fe3d7e0cbf' 2025-12-04T08:57:33.0992040Z Submodule path 'third_party/benchmark': checked out '299e5928955cc62af9968370293b916f5130916f' 2025-12-04T08:57:33.5071363Z Submodule path 'third_party/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-12-04T08:57:33.5607384Z Submodule path 'third_party/cpp-httplib': checked out '89c932f313c6437c38f2982869beacc89c2f2246' 2025-12-04T08:57:33.6601057Z Submodule path 'third_party/cpuinfo': checked out 'f858c30bcb16f8effd5ff46996f0514539e17abc' 2025-12-04T08:57:33.7111281Z Submodule path 'third_party/cudnn_frontend': checked out '0b1577c8c83401237d601d0d0db5210506705396' 2025-12-04T08:57:34.4300558Z Submodule path 'third_party/cutlass': checked out 'f88806b1e31dfa579842638740216dd41fc6c588' 2025-12-04T08:57:34.6052487Z Submodule path 'third_party/fbgemm': checked out 'c0b988d39a9e47c794d699f29930ed4d7c7e13a4' 2025-12-04T08:57:34.6077357Z Submodule 'external/asmjit' (https://github.com/asmjit/asmjit.git) registered for path 'third_party/fbgemm/external/asmjit' 2025-12-04T08:57:34.6080917Z Submodule 'external/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:57:34.6084154Z Submodule 'external/cpuinfo' (https://github.com/pytorch/cpuinfo) registered for path 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:57:34.6087411Z Submodule 'external/cutlass' (https://github.com/jwfromm/cutlass) registered for path 'third_party/fbgemm/external/cutlass' 2025-12-04T08:57:34.6090892Z Submodule 'external/googletest' (https://github.com/google/googletest) registered for path 'third_party/fbgemm/external/googletest' 2025-12-04T08:57:34.6094680Z Submodule 'external/hipify_torch' (https://github.com/ROCmSoftwarePlatform/hipify_torch.git) registered for path 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:57:34.6098315Z Submodule 'external/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/fbgemm/external/json' 2025-12-04T08:57:34.6128956Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/asmjit'... 2025-12-04T08:57:35.7347279Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/hipify_torch'... 2025-12-04T08:57:35.7348407Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cpuinfo'... 2025-12-04T08:57:35.7349381Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/googletest'... 2025-12-04T08:57:35.8347702Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/composable_kernel'... 2025-12-04T08:57:38.7330130Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cutlass'... 2025-12-04T08:57:38.8330820Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/json'... 2025-12-04T08:57:40.8782097Z Submodule path 'third_party/fbgemm/external/asmjit': checked out 'a3199e8857792cd10b7589ff5d58343d2c9008ea' 2025-12-04T08:57:41.2872093Z Submodule path 'third_party/fbgemm/external/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-12-04T08:57:41.3896317Z Submodule path 'third_party/fbgemm/external/cpuinfo': checked out '6543fec09b2f04ac4a666882998b534afc9c1349' 2025-12-04T08:57:42.0922182Z Submodule path 'third_party/fbgemm/external/cutlass': checked out '98125ce499b0fdf7ffbe0e3052f5b8709f4840f8' 2025-12-04T08:57:42.1408888Z Submodule path 'third_party/fbgemm/external/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T08:57:42.1542586Z Submodule path 'third_party/fbgemm/external/hipify_torch': checked out '63b6a7b541fa7f08f8475ca7d74054db36ff2691' 2025-12-04T08:57:42.2745131Z Submodule path 'third_party/fbgemm/external/json': checked out '9cca280a4d0ccf0c08f47a99aa71d1b0e52f8d03' 2025-12-04T08:57:42.3565140Z Submodule path 'third_party/flash-attention': checked out '979702c87a8713a8e0a5e9fee122b90d2ef13be5' 2025-12-04T08:57:42.3586680Z Submodule 'csrc/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:57:42.3589809Z Submodule 'csrc/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:57:42.3619547Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/composable_kernel'... 2025-12-04T08:57:46.2817769Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/cutlass'... 2025-12-04T08:57:46.5654670Z Submodule path 'third_party/flash-attention/csrc/composable_kernel': checked out '888317e698e9803c62bd38568abc9e05d7709f33' 2025-12-04T08:57:47.1877090Z Submodule path 'third_party/flash-attention/csrc/cutlass': checked out 'c506e16788cb08416a4a57e11a9067beeee29420' 2025-12-04T08:57:47.3485906Z Submodule path 'third_party/flatbuffers': checked out 'a2cd1ea3b6d3fee220106b5fed3f7ce8da9eb757' 2025-12-04T08:57:47.3805940Z Submodule path 'third_party/fmt': checked out '407c905e45ad75fc29bf0f9bb7c5c2fd3475976f' 2025-12-04T08:57:47.4218555Z Submodule path 'third_party/gemmlowp/gemmlowp': checked out '3fb5c176c17c765a3492cd2f0321b0dab712f350' 2025-12-04T08:57:47.4528640Z Submodule path 'third_party/gloo': checked out '54cbae0d3a67fa890b4c3d9ee162b7860315e341' 2025-12-04T08:57:47.4990222Z Submodule path 'third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T08:57:47.5136200Z Submodule path 'third_party/ideep': checked out '719d8e6cd7f7a0e01b155657526d693acf97c2b3' 2025-12-04T08:57:47.5156437Z Submodule 'mkl-dnn' (https://github.com/intel/mkl-dnn.git) registered for path 'third_party/ideep/mkl-dnn' 2025-12-04T08:57:47.5185066Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep/mkl-dnn'... 2025-12-04T08:58:00.8904432Z Submodule path 'third_party/ideep/mkl-dnn': checked out '8d263e693366ef8db40acc569cc7d8edf644556d' 2025-12-04T08:58:00.9144993Z Submodule path 'third_party/ittapi': checked out 'dec1d23ca65ab069d225dfe40dea14f455170959' 2025-12-04T08:58:00.9997074Z Submodule path 'third_party/kineto': checked out '31f85df8fbd89c188f14ef10f1ec65379786b943' 2025-12-04T08:58:01.0018273Z Submodule 'libkineto/third_party/dynolog' (https://github.com/facebookincubator/dynolog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:58:01.0021096Z Submodule 'libkineto/third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:58:01.0024416Z Submodule 'libkineto/third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:58:01.0055330Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog'... 2025-12-04T08:58:01.6652439Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/fmt'... 2025-12-04T08:58:02.0947359Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/googletest'... 2025-12-04T08:58:02.1871898Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog': checked out 'd2ffe0a4e3acace628db49974246b66fc3e85fb1' 2025-12-04T08:58:02.1892487Z Submodule 'third_party/DCGM' (https://github.com/NVIDIA/DCGM.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:58:02.1896065Z Submodule 'third_party/cpr' (https://github.com/libcpr/cpr.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:58:02.1899531Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:58:02.1903066Z Submodule 'third_party/gflags' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:58:02.1906782Z Submodule 'third_party/glog' (https://github.com/google/glog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:58:02.1910472Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:58:02.1914386Z Submodule 'third_party/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:58:02.1918171Z Submodule 'third_party/pfs' (https://github.com/dtrugman/pfs.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:58:02.1922140Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:58:02.1953196Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM'... 2025-12-04T08:58:04.1762644Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/pfs'... 2025-12-04T08:58:04.1763764Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags'... 2025-12-04T08:58:04.1764652Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp'... 2025-12-04T08:58:04.1765472Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/cpr'... 2025-12-04T08:58:04.1766242Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/glog'... 2025-12-04T08:58:04.1767180Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/googletest'... 2025-12-04T08:58:04.1767987Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/fmt'... 2025-12-04T08:58:04.2763744Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/json'... 2025-12-04T08:58:08.7965757Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM': checked out 'ffde4e54bc7249a6039a5e6b45b395141e1217f9' 2025-12-04T08:58:08.8199034Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr': checked out '871ed52d350214a034f6ef8a3b8f51c5ce1bd400' 2025-12-04T08:58:08.8590301Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt': checked out 'cd4af11efc9c622896a3e4cb599fa28668ca3d05' 2025-12-04T08:58:08.8747838Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags': checked out 'e171aa2d15ed9eb17054558e0b3a6a413bb01067' 2025-12-04T08:58:08.8767243Z Submodule 'doc' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:58:08.8805505Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc'... 2025-12-04T08:58:09.1097010Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc': checked out '8411df715cf522606e3b1aca386ddfc0b63d34b4' 2025-12-04T08:58:09.1317403Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog': checked out 'b33e3bad4c46c8a6345525fd822af355e5ef9446' 2025-12-04T08:58:09.1786968Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T08:58:09.2883718Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json': checked out '4f8fba14066156b73f1189a2b8bd568bde5284c5' 2025-12-04T08:58:09.3083337Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs': checked out 'f68a2fa8ea36c783bdd760371411fcb495aa3150' 2025-12-04T08:58:09.3288206Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp': checked out 'b1234816facfdda29845c46696a02998a4af115a' 2025-12-04T08:58:09.3307366Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:58:09.3310788Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:58:09.3341259Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-12-04T08:58:11.2243649Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest'... 2025-12-04T08:58:11.4716124Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'd7ba35bbb649209c66e582d5a0244ba988a15159' 2025-12-04T08:58:11.5211855Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-12-04T08:58:11.5545875Z Submodule path 'third_party/kineto/libkineto/third_party/fmt': checked out '40626af88bd7df9a5fb80be7b25ac85b122d6c21' 2025-12-04T08:58:11.6017339Z Submodule path 'third_party/kineto/libkineto/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T08:58:11.6639451Z Submodule path 'third_party/kleidiai': checked out 'd7770c89632329a9914ef1a90289917597639cbe' 2025-12-04T08:58:11.7059764Z Submodule path 'third_party/mimalloc': checked out 'fbd8b99c2b828428947d70fdc046bb55609be93e' 2025-12-04T08:58:11.8206191Z Submodule path 'third_party/nlohmann': checked out '55f93686c01528224f448c19128836e7df245f72' 2025-12-04T08:58:12.3448851Z Submodule path 'third_party/onnx': checked out 'e709452ef2bbc1d113faf678c24e6d3467696e83' 2025-12-04T08:58:12.3485454Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/onnx/third_party/pybind11' 2025-12-04T08:58:12.3516054Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx/third_party/pybind11'... 2025-12-04T08:58:13.1293831Z Submodule path 'third_party/onnx/third_party/pybind11': checked out 'a2e59f0e7065404b44dfe92a28aca47ba1378dc4' 2025-12-04T08:58:13.2156039Z Submodule path 'third_party/opentelemetry-cpp': checked out 'a799f4aed9c94b765dcdaabaeab7d5e7e2310878' 2025-12-04T08:58:13.2176816Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark) registered for path 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:58:13.2179865Z Submodule 'third_party/googletest' (https://github.com/google/googletest) registered for path 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:58:13.2183232Z Submodule 'third_party/ms-gsl' (https://github.com/microsoft/GSL) registered for path 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:58:13.2186690Z Submodule 'third_party/nlohmann-json' (https://github.com/nlohmann/json) registered for path 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:58:13.2190222Z Submodule 'third_party/opentelemetry-proto' (https://github.com/open-telemetry/opentelemetry-proto) registered for path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:58:13.2193692Z Submodule 'third_party/opentracing-cpp' (https://github.com/opentracing/opentracing-cpp.git) registered for path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:58:13.2197295Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:58:13.2200896Z Submodule 'tools/vcpkg' (https://github.com/Microsoft/vcpkg) registered for path 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:58:13.2231871Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/benchmark'... 2025-12-04T08:58:13.6135929Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentracing-cpp'... 2025-12-04T08:58:13.6137098Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentelemetry-proto'... 2025-12-04T08:58:13.6138102Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/ms-gsl'... 2025-12-04T08:58:13.6138926Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp'... 2025-12-04T08:58:13.7137180Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/googletest'... 2025-12-04T08:58:14.2357151Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/nlohmann-json'... 2025-12-04T08:58:20.8948331Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/tools/vcpkg'... 2025-12-04T08:58:21.2746654Z Submodule path 'third_party/opentelemetry-cpp/third_party/benchmark': checked out 'd572f4777349d43653b21d6c2fc63020ab326db2' 2025-12-04T08:58:21.3250681Z Submodule path 'third_party/opentelemetry-cpp/third_party/googletest': checked out 'b796f7d44681514f58a683a3a71ff17c94edb0c1' 2025-12-04T08:58:21.3438679Z Submodule path 'third_party/opentelemetry-cpp/third_party/ms-gsl': checked out '6f4529395c5b7c2d661812257cd6780c67e54afa' 2025-12-04T08:58:21.4601294Z Submodule path 'third_party/opentelemetry-cpp/third_party/nlohmann-json': checked out 'bc889afb4c5bf1c0d8ee29ef35eaaf4c8bef8a5d' 2025-12-04T08:58:21.4770328Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto': checked out '4ca4f0335c63cda7ab31ea7ed70d6553aee14dce' 2025-12-04T08:58:21.4955332Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp': checked out '06b57f48ded1fa3bdd3d4346f6ef29e40e08eaf5' 2025-12-04T08:58:21.5151905Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp': checked out 'c9ffcdda9086ffd9e1283ea7a0276d831f3c8a8d' 2025-12-04T08:58:21.5170942Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:58:21.5174392Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:58:21.5204038Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-12-04T08:58:23.4646359Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest'... 2025-12-04T08:58:23.7110610Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'eefb26f82b233268fc98577d265352720d477ba4' 2025-12-04T08:58:23.7597112Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-12-04T08:58:24.3979210Z Submodule path 'third_party/opentelemetry-cpp/tools/vcpkg': checked out '8eb57355a4ffb410a2e94c07b4dca2dffbee8e50' 2025-12-04T08:58:24.4116099Z Submodule path 'third_party/pocketfft': checked out '0fa0ef591e38c2758e3184c6c23e497b9f732ffa' 2025-12-04T08:58:24.7043730Z Submodule path 'third_party/protobuf': checked out 'd1eca4e4b421cd2997495c4b4e65cea6be4e9b8a' 2025-12-04T08:58:24.7068824Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:58:24.7072446Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/protobuf/third_party/googletest' 2025-12-04T08:58:24.7103711Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/benchmark'... 2025-12-04T08:58:25.1874815Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/googletest'... 2025-12-04T08:58:25.5382058Z Submodule path 'third_party/protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' 2025-12-04T08:58:25.6113939Z Submodule path 'third_party/protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' 2025-12-04T08:58:25.6224780Z Submodule path 'third_party/psimd': checked out '072586a71b55b7f8c584153d223e95687148a900' 2025-12-04T08:58:25.6367997Z Submodule path 'third_party/pthreadpool': checked out '4fe0e1e183925bf8cfa6aae24237e724a96479b8' 2025-12-04T08:58:25.6837233Z Submodule path 'third_party/pybind11': checked out 'f5fbe867d2d26e4a0a9177a51f6e568868ad3dc8' 2025-12-04T08:58:25.7139375Z Submodule path 'third_party/python-peachpy': checked out 'f45429b087dd7d5bc78bb40dc7cf06425c252d67' 2025-12-04T08:58:25.7597325Z Submodule path 'third_party/sleef': checked out '5a1d179df9cf652951b59010a2d2075372d67f68' 2025-12-04T08:58:25.7921886Z Submodule path 'third_party/tensorpipe': checked out '2b4cd91092d335a697416b2a3cb398283246849d' 2025-12-04T08:58:25.7942218Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:58:25.7945563Z Submodule 'third_party/libnop' (https://github.com/google/libnop.git) registered for path 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:58:25.7948935Z Submodule 'third_party/libuv' (https://github.com/libuv/libuv.git) registered for path 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:58:25.7952461Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:58:25.7983221Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/googletest'... 2025-12-04T08:58:26.7853771Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libnop'... 2025-12-04T08:58:26.7854714Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11'... 2025-12-04T08:58:26.7855542Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libuv'... 2025-12-04T08:58:26.8422730Z Submodule path 'third_party/tensorpipe/third_party/googletest': checked out 'aee0f9d9b5b87796ee8a0ab26b7587ec30e8858e' 2025-12-04T08:58:26.8612361Z Submodule path 'third_party/tensorpipe/third_party/libnop': checked out '910b55815be16109f04f4180e9adee14fb4ce281' 2025-12-04T08:58:26.9365315Z Submodule path 'third_party/tensorpipe/third_party/libuv': checked out '5152db2cbfeb5582e9c27c5ea1dba2cd9e10759b' 2025-12-04T08:58:26.9682689Z Submodule path 'third_party/tensorpipe/third_party/pybind11': checked out 'a23996fce38ff6ccfbcdc09f1e63f2c4be5ea2ef' 2025-12-04T08:58:26.9702450Z Submodule 'tools/clang' (https://github.com/wjakob/clang-cindex-python3) registered for path 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:58:26.9731461Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11/tools/clang'... 2025-12-04T08:58:27.1674149Z Submodule path 'third_party/tensorpipe/third_party/pybind11/tools/clang': checked out '6a00cbc4a9b8e68b71caf7f774b3f9c753ae84d5' 2025-12-04T08:58:27.1719048Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2025-12-04T08:58:27.2045704Z Entering 'android/libs/fbjni' 2025-12-04T08:58:27.2093494Z Entering 'third_party/FP16' 2025-12-04T08:58:27.2141831Z Entering 'third_party/FXdiv' 2025-12-04T08:58:27.2190439Z Entering 'third_party/NNPACK' 2025-12-04T08:58:27.2239932Z Entering 'third_party/NVTX' 2025-12-04T08:58:27.2289698Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:58:27.2339166Z Entering 'third_party/XNNPACK' 2025-12-04T08:58:27.2402455Z Entering 'third_party/aiter' 2025-12-04T08:58:27.2450744Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:58:27.2508512Z Entering 'third_party/benchmark' 2025-12-04T08:58:27.2561022Z Entering 'third_party/composable_kernel' 2025-12-04T08:58:27.2615815Z Entering 'third_party/cpp-httplib' 2025-12-04T08:58:27.2662538Z Entering 'third_party/cpuinfo' 2025-12-04T08:58:27.2711249Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:58:27.2760343Z Entering 'third_party/cutlass' 2025-12-04T08:58:27.2818252Z Entering 'third_party/fbgemm' 2025-12-04T08:58:27.2865749Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:58:27.2923731Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:58:27.2964599Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:58:27.3010446Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:58:27.3063069Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:58:27.3109475Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:58:27.3155336Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:58:27.3206352Z Entering 'third_party/flash-attention' 2025-12-04T08:58:27.3254262Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:58:27.3307033Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:58:27.3362676Z Entering 'third_party/flatbuffers' 2025-12-04T08:58:27.3414273Z Entering 'third_party/fmt' 2025-12-04T08:58:27.3461802Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:58:27.3511037Z Entering 'third_party/gloo' 2025-12-04T08:58:27.3559749Z Entering 'third_party/googletest' 2025-12-04T08:58:27.3609526Z Entering 'third_party/ideep' 2025-12-04T08:58:27.3656928Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:58:27.3711297Z Entering 'third_party/ittapi' 2025-12-04T08:58:27.3762158Z Entering 'third_party/kineto' 2025-12-04T08:58:27.3810297Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:58:27.3857840Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:58:27.3906620Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:58:27.3953215Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:58:27.4001741Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:58:27.4052358Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:58:27.4103409Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:58:27.4151360Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:58:27.4200129Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:58:27.4249648Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:58:27.4300306Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:58:27.4347707Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:58:27.4396056Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:58:27.4447669Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:58:27.4493557Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:58:27.4543017Z Entering 'third_party/kleidiai' 2025-12-04T08:58:27.4591143Z Entering 'third_party/mimalloc' 2025-12-04T08:58:27.4639920Z Entering 'third_party/nlohmann' 2025-12-04T08:58:27.4694341Z Entering 'third_party/onnx' 2025-12-04T08:58:27.4756047Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:58:27.4807095Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:58:27.4854043Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:58:27.4900880Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:58:27.4947426Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:58:27.4992650Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:58:27.5040997Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:58:27.5090584Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:58:27.5141138Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:58:27.5189818Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:58:27.5243413Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:58:27.5292840Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:58:27.5358317Z Entering 'third_party/pocketfft' 2025-12-04T08:58:27.5411074Z Entering 'third_party/protobuf' 2025-12-04T08:58:27.5461235Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:58:27.5512544Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:58:27.5562213Z Entering 'third_party/psimd' 2025-12-04T08:58:27.5611260Z Entering 'third_party/pthreadpool' 2025-12-04T08:58:27.5659211Z Entering 'third_party/pybind11' 2025-12-04T08:58:27.5709505Z Entering 'third_party/python-peachpy' 2025-12-04T08:58:27.5757741Z Entering 'third_party/sleef' 2025-12-04T08:58:27.5803937Z Entering 'third_party/tensorpipe' 2025-12-04T08:58:27.5850214Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:58:27.5898451Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:58:27.5944254Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:58:27.5990674Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:58:27.6040316Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:58:27.6115099Z ##[endgroup] 2025-12-04T08:58:27.6115741Z ##[group]Persisting credentials for submodules 2025-12-04T08:58:27.6121544Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || :" 2025-12-04T08:58:27.6443218Z Entering 'android/libs/fbjni' 2025-12-04T08:58:27.6511093Z Entering 'third_party/FP16' 2025-12-04T08:58:27.6572821Z Entering 'third_party/FXdiv' 2025-12-04T08:58:27.6643062Z Entering 'third_party/NNPACK' 2025-12-04T08:58:27.6710883Z Entering 'third_party/NVTX' 2025-12-04T08:58:27.6774458Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:58:27.6836264Z Entering 'third_party/XNNPACK' 2025-12-04T08:58:27.6912128Z Entering 'third_party/aiter' 2025-12-04T08:58:27.6973402Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:58:27.7045500Z Entering 'third_party/benchmark' 2025-12-04T08:58:27.7115347Z Entering 'third_party/composable_kernel' 2025-12-04T08:58:27.7186130Z Entering 'third_party/cpp-httplib' 2025-12-04T08:58:27.7249074Z Entering 'third_party/cpuinfo' 2025-12-04T08:58:27.7315343Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:58:27.7378580Z Entering 'third_party/cutlass' 2025-12-04T08:58:27.7447678Z Entering 'third_party/fbgemm' 2025-12-04T08:58:27.7511335Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:58:27.7572716Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:58:27.7639436Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:58:27.7705515Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:58:27.7776281Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:58:27.7839577Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:58:27.7907199Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:58:27.7977043Z Entering 'third_party/flash-attention' 2025-12-04T08:58:27.8038959Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:58:27.8110322Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:58:27.8181352Z Entering 'third_party/flatbuffers' 2025-12-04T08:58:27.8246914Z Entering 'third_party/fmt' 2025-12-04T08:58:27.8310724Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:58:27.8371624Z Entering 'third_party/gloo' 2025-12-04T08:58:27.8435241Z Entering 'third_party/googletest' 2025-12-04T08:58:27.8499447Z Entering 'third_party/ideep' 2025-12-04T08:58:27.8559273Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:58:27.8635712Z Entering 'third_party/ittapi' 2025-12-04T08:58:27.8705129Z Entering 'third_party/kineto' 2025-12-04T08:58:27.8770981Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:58:27.8832297Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:58:27.8895458Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:58:27.8960862Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:58:27.9023625Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:58:27.9089323Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:58:27.9155324Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:58:27.9219944Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:58:27.9283479Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:58:27.9344807Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:58:27.9413984Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:58:27.9475562Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:58:27.9545907Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:58:27.9619356Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:58:27.9681613Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:58:27.9752028Z Entering 'third_party/kleidiai' 2025-12-04T08:58:27.9821508Z Entering 'third_party/mimalloc' 2025-12-04T08:58:27.9893903Z Entering 'third_party/nlohmann' 2025-12-04T08:58:27.9959704Z Entering 'third_party/onnx' 2025-12-04T08:58:28.0036264Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:58:28.0103077Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:58:28.0171675Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:58:28.0232705Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:58:28.0293117Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:58:28.0360050Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:58:28.0423607Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:58:28.0483595Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:58:28.0542750Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:58:28.0613726Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:58:28.0682498Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:58:28.0747280Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:58:28.0831933Z Entering 'third_party/pocketfft' 2025-12-04T08:58:28.0899699Z Entering 'third_party/protobuf' 2025-12-04T08:58:28.0964467Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:58:28.1025929Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:58:28.1096546Z Entering 'third_party/psimd' 2025-12-04T08:58:28.1160074Z Entering 'third_party/pthreadpool' 2025-12-04T08:58:28.1224300Z Entering 'third_party/pybind11' 2025-12-04T08:58:28.1289781Z Entering 'third_party/python-peachpy' 2025-12-04T08:58:28.1355271Z Entering 'third_party/sleef' 2025-12-04T08:58:28.1425571Z Entering 'third_party/tensorpipe' 2025-12-04T08:58:28.1487647Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:58:28.1547968Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:58:28.1610040Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:58:28.1672053Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:58:28.1738267Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:58:28.1822909Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url" 2025-12-04T08:58:28.2145486Z Entering 'android/libs/fbjni' 2025-12-04T08:58:28.2205412Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-12-04T08:58:28.2224756Z Entering 'third_party/FP16' 2025-12-04T08:58:28.2283126Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-12-04T08:58:28.2302737Z Entering 'third_party/FXdiv' 2025-12-04T08:58:28.2361285Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-12-04T08:58:28.2381676Z Entering 'third_party/NNPACK' 2025-12-04T08:58:28.2440172Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-12-04T08:58:28.2460334Z Entering 'third_party/NVTX' 2025-12-04T08:58:28.2524418Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-12-04T08:58:28.2544398Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:58:28.2603378Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-12-04T08:58:28.2621812Z Entering 'third_party/XNNPACK' 2025-12-04T08:58:28.2677950Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-12-04T08:58:28.2715735Z Entering 'third_party/aiter' 2025-12-04T08:58:28.2772621Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-12-04T08:58:28.2791868Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:58:28.2849864Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-12-04T08:58:28.2883021Z Entering 'third_party/benchmark' 2025-12-04T08:58:28.2941120Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-12-04T08:58:28.2960830Z Entering 'third_party/composable_kernel' 2025-12-04T08:58:28.3019605Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-12-04T08:58:28.3046956Z Entering 'third_party/cpp-httplib' 2025-12-04T08:58:28.3106261Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-12-04T08:58:28.3125147Z Entering 'third_party/cpuinfo' 2025-12-04T08:58:28.3181807Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-12-04T08:58:28.3201796Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:58:28.3266034Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-12-04T08:58:28.3285225Z Entering 'third_party/cutlass' 2025-12-04T08:58:28.3349175Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-12-04T08:58:28.3378819Z Entering 'third_party/fbgemm' 2025-12-04T08:58:28.3435352Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-12-04T08:58:28.3456164Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:58:28.3513924Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-12-04T08:58:28.3532569Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:58:28.3590320Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-12-04T08:58:28.3618140Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:58:28.3673456Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-12-04T08:58:28.3693000Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:58:28.3750705Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-12-04T08:58:28.3778892Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:58:28.3835543Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-12-04T08:58:28.3854777Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:58:28.3912389Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-12-04T08:58:28.3929846Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:58:28.3989036Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-12-04T08:58:28.4012240Z Entering 'third_party/flash-attention' 2025-12-04T08:58:28.4070535Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-12-04T08:58:28.4090163Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:58:28.4149161Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-12-04T08:58:28.4174466Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:58:28.4240598Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-12-04T08:58:28.4269568Z Entering 'third_party/flatbuffers' 2025-12-04T08:58:28.4331635Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-12-04T08:58:28.4353704Z Entering 'third_party/fmt' 2025-12-04T08:58:28.4411253Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-12-04T08:58:28.4429386Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:58:28.4489653Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-12-04T08:58:28.4510177Z Entering 'third_party/gloo' 2025-12-04T08:58:28.4568170Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-12-04T08:58:28.4588867Z Entering 'third_party/googletest' 2025-12-04T08:58:28.4648819Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:58:28.4669230Z Entering 'third_party/ideep' 2025-12-04T08:58:28.4730487Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-12-04T08:58:28.4748684Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:58:28.4809858Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-12-04T08:58:28.4836937Z Entering 'third_party/ittapi' 2025-12-04T08:58:28.4895768Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-12-04T08:58:28.4913874Z Entering 'third_party/kineto' 2025-12-04T08:58:28.4973025Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-12-04T08:58:28.4991607Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:58:28.5049758Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-12-04T08:58:28.5068525Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:58:28.5125782Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-12-04T08:58:28.5145945Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:58:28.5203234Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-12-04T08:58:28.5221994Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:58:28.5279682Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-12-04T08:58:28.5299697Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:58:28.5360108Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-12-04T08:58:28.5378561Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:58:28.5438484Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-12-04T08:58:28.5461352Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:58:28.5521216Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-12-04T08:58:28.5541017Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:58:28.5598743Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:58:28.5617829Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:58:28.5674510Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-12-04T08:58:28.5694604Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:58:28.5754482Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-12-04T08:58:28.5773934Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:58:28.5833529Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T08:58:28.5851313Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:58:28.5913961Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T08:58:28.5935315Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:58:28.5994669Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T08:58:28.6019118Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:58:28.6083141Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-12-04T08:58:28.6102406Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:58:28.6160491Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-12-04T08:58:28.6182758Z Entering 'third_party/kleidiai' 2025-12-04T08:58:28.6240236Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-12-04T08:58:28.6260984Z Entering 'third_party/mimalloc' 2025-12-04T08:58:28.6319219Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-12-04T08:58:28.6339770Z Entering 'third_party/nlohmann' 2025-12-04T08:58:28.6398090Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-12-04T08:58:28.6421825Z Entering 'third_party/onnx' 2025-12-04T08:58:28.6482241Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-12-04T08:58:28.6516560Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:58:28.6573145Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-12-04T08:58:28.6596576Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:58:28.6660005Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-12-04T08:58:28.6684151Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:58:28.6739621Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-12-04T08:58:28.6759168Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:58:28.6817479Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:58:28.6835744Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:58:28.6892478Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-12-04T08:58:28.6911599Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:58:28.6969312Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-12-04T08:58:28.6990207Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:58:28.7053329Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-12-04T08:58:28.7072061Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:58:28.7130652Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-12-04T08:58:28.7150047Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:58:28.7213658Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T08:58:28.7230777Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:58:28.7289447Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T08:58:28.7312082Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:58:28.7369902Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T08:58:28.7392826Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:58:28.7448788Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-12-04T08:58:28.7491329Z Entering 'third_party/pocketfft' 2025-12-04T08:58:28.7549986Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-12-04T08:58:28.7569912Z Entering 'third_party/protobuf' 2025-12-04T08:58:28.7630181Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-12-04T08:58:28.7651936Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:58:28.7710053Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-12-04T08:58:28.7729854Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:58:28.7792182Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:58:28.7815755Z Entering 'third_party/psimd' 2025-12-04T08:58:28.7873154Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-12-04T08:58:28.7892777Z Entering 'third_party/pthreadpool' 2025-12-04T08:58:28.7951787Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-12-04T08:58:28.7971455Z Entering 'third_party/pybind11' 2025-12-04T08:58:28.8026018Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-12-04T08:58:28.8045621Z Entering 'third_party/python-peachpy' 2025-12-04T08:58:28.8103384Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-12-04T08:58:28.8121487Z Entering 'third_party/sleef' 2025-12-04T08:58:28.8179005Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-12-04T08:58:28.8199503Z Entering 'third_party/tensorpipe' 2025-12-04T08:58:28.8257296Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-12-04T08:58:28.8279999Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:58:28.8340732Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:58:28.8360060Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:58:28.8418173Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-12-04T08:58:28.8438102Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:58:28.8496790Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-12-04T08:58:28.8514839Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:58:28.8571112Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-12-04T08:58:28.8589468Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:58:28.8648180Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-12-04T08:58:28.9310336Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2025-12-04T08:58:28.9645209Z Entering 'android/libs/fbjni' 2025-12-04T08:58:28.9692806Z Entering 'third_party/FP16' 2025-12-04T08:58:28.9741642Z Entering 'third_party/FXdiv' 2025-12-04T08:58:28.9790557Z Entering 'third_party/NNPACK' 2025-12-04T08:58:28.9839917Z Entering 'third_party/NVTX' 2025-12-04T08:58:28.9890973Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:58:28.9940040Z Entering 'third_party/XNNPACK' 2025-12-04T08:58:28.9998597Z Entering 'third_party/aiter' 2025-12-04T08:58:29.0044883Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:58:29.0099982Z Entering 'third_party/benchmark' 2025-12-04T08:58:29.0151126Z Entering 'third_party/composable_kernel' 2025-12-04T08:58:29.0208171Z Entering 'third_party/cpp-httplib' 2025-12-04T08:58:29.0254767Z Entering 'third_party/cpuinfo' 2025-12-04T08:58:29.0303097Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:58:29.0354319Z Entering 'third_party/cutlass' 2025-12-04T08:58:29.0410237Z Entering 'third_party/fbgemm' 2025-12-04T08:58:29.0460989Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:58:29.0514364Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:58:29.0568574Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:58:29.0615647Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:58:29.0669600Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:58:29.0715028Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:58:29.0761618Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:58:29.0818085Z Entering 'third_party/flash-attention' 2025-12-04T08:58:29.0863300Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:58:29.0916128Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:58:29.0971903Z Entering 'third_party/flatbuffers' 2025-12-04T08:58:29.1026482Z Entering 'third_party/fmt' 2025-12-04T08:58:29.1072529Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:58:29.1122581Z Entering 'third_party/gloo' 2025-12-04T08:58:29.1170955Z Entering 'third_party/googletest' 2025-12-04T08:58:29.1222091Z Entering 'third_party/ideep' 2025-12-04T08:58:29.1268768Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:58:29.1322627Z Entering 'third_party/ittapi' 2025-12-04T08:58:29.1372819Z Entering 'third_party/kineto' 2025-12-04T08:58:29.1424702Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:58:29.1470362Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:58:29.1521619Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:58:29.1574436Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:58:29.1622801Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:58:29.1669445Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:58:29.1719983Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:58:29.1772962Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:58:29.1822009Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:58:29.1875290Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:58:29.1923721Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:58:29.1970232Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:58:29.2026475Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:58:29.2079307Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:58:29.2129198Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:58:29.2180712Z Entering 'third_party/kleidiai' 2025-12-04T08:58:29.2230921Z Entering 'third_party/mimalloc' 2025-12-04T08:58:29.2280064Z Entering 'third_party/nlohmann' 2025-12-04T08:58:29.2331462Z Entering 'third_party/onnx' 2025-12-04T08:58:29.2396186Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:58:29.2447123Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:58:29.2494598Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:58:29.2542399Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:58:29.2590476Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:58:29.2639094Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:58:29.2688648Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:58:29.2734364Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:58:29.2780840Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:58:29.2827731Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:58:29.2874735Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:58:29.2925752Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:58:29.2992383Z Entering 'third_party/pocketfft' 2025-12-04T08:58:29.3043241Z Entering 'third_party/protobuf' 2025-12-04T08:58:29.3093447Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:58:29.3141524Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:58:29.3192248Z Entering 'third_party/psimd' 2025-12-04T08:58:29.3243246Z Entering 'third_party/pthreadpool' 2025-12-04T08:58:29.3291328Z Entering 'third_party/pybind11' 2025-12-04T08:58:29.3340672Z Entering 'third_party/python-peachpy' 2025-12-04T08:58:29.3392173Z Entering 'third_party/sleef' 2025-12-04T08:58:29.3444796Z Entering 'third_party/tensorpipe' 2025-12-04T08:58:29.3491351Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:58:29.3542919Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:58:29.3590211Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:58:29.3639290Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:58:29.3686572Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:58:29.3755708Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2025-12-04T08:58:29.4085892Z Entering 'android/libs/fbjni' 2025-12-04T08:58:29.4134504Z Entering 'third_party/FP16' 2025-12-04T08:58:29.4181748Z Entering 'third_party/FXdiv' 2025-12-04T08:58:29.4230834Z Entering 'third_party/NNPACK' 2025-12-04T08:58:29.4279932Z Entering 'third_party/NVTX' 2025-12-04T08:58:29.4330344Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:58:29.4379855Z Entering 'third_party/XNNPACK' 2025-12-04T08:58:29.4442100Z Entering 'third_party/aiter' 2025-12-04T08:58:29.4492820Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:58:29.4548545Z Entering 'third_party/benchmark' 2025-12-04T08:58:29.4600313Z Entering 'third_party/composable_kernel' 2025-12-04T08:58:29.4660740Z Entering 'third_party/cpp-httplib' 2025-12-04T08:58:29.4714698Z Entering 'third_party/cpuinfo' 2025-12-04T08:58:29.4762713Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:58:29.4812338Z Entering 'third_party/cutlass' 2025-12-04T08:58:29.4870167Z Entering 'third_party/fbgemm' 2025-12-04T08:58:29.4921035Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:58:29.4972347Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:58:29.5028302Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:58:29.5074331Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:58:29.5129822Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:58:29.5181700Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:58:29.5228220Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:58:29.5277805Z Entering 'third_party/flash-attention' 2025-12-04T08:58:29.5324100Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:58:29.5376887Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:58:29.5433491Z Entering 'third_party/flatbuffers' 2025-12-04T08:58:29.5485457Z Entering 'third_party/fmt' 2025-12-04T08:58:29.5533650Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:58:29.5582161Z Entering 'third_party/gloo' 2025-12-04T08:58:29.5629912Z Entering 'third_party/googletest' 2025-12-04T08:58:29.5680193Z Entering 'third_party/ideep' 2025-12-04T08:58:29.5728013Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:58:29.5781650Z Entering 'third_party/ittapi' 2025-12-04T08:58:29.5833672Z Entering 'third_party/kineto' 2025-12-04T08:58:29.5880714Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:58:29.5929612Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:58:29.5983299Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:58:29.6031970Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:58:29.6084349Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:58:29.6130643Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:58:29.6187136Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:58:29.6234237Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:58:29.6282220Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:58:29.6336169Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:58:29.6382848Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:58:29.6430045Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:58:29.6481757Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:58:29.6536835Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:58:29.6585051Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:58:29.6637437Z Entering 'third_party/kleidiai' 2025-12-04T08:58:29.6693485Z Entering 'third_party/mimalloc' 2025-12-04T08:58:29.6741824Z Entering 'third_party/nlohmann' 2025-12-04T08:58:29.6792186Z Entering 'third_party/onnx' 2025-12-04T08:58:29.6855012Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:58:29.6906278Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:58:29.6952779Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:58:29.7000109Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:58:29.7052662Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:58:29.7103638Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:58:29.7152562Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:58:29.7202676Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:58:29.7250268Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:58:29.7296919Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:58:29.7344758Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:58:29.7394615Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:58:29.7459619Z Entering 'third_party/pocketfft' 2025-12-04T08:58:29.7512404Z Entering 'third_party/protobuf' 2025-12-04T08:58:29.7562351Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:58:29.7614961Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:58:29.7664544Z Entering 'third_party/psimd' 2025-12-04T08:58:29.7713365Z Entering 'third_party/pthreadpool' 2025-12-04T08:58:29.7762085Z Entering 'third_party/pybind11' 2025-12-04T08:58:29.7814259Z Entering 'third_party/python-peachpy' 2025-12-04T08:58:29.7863053Z Entering 'third_party/sleef' 2025-12-04T08:58:29.7912133Z Entering 'third_party/tensorpipe' 2025-12-04T08:58:29.7959279Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:58:29.8008906Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:58:29.8053773Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:58:29.8101550Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:58:29.8152642Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:58:29.8225184Z ##[endgroup] 2025-12-04T08:58:29.8262442Z [command]/usr/bin/git log -1 --format=%H 2025-12-04T08:58:29.8285214Z ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T08:58:29.8401637Z ##[group]Run cd "${GITHUB_WORKSPACE}" 2025-12-04T08:58:29.8401909Z cd "${GITHUB_WORKSPACE}" 2025-12-04T08:58:29.8402135Z # Clean stale submodule dirs 2025-12-04T08:58:29.8402363Z if [ -z "${NO_SUDO}" ]; then 2025-12-04T08:58:29.8402638Z  sudo git submodule foreach --recursive git clean -ffdx 2025-12-04T08:58:29.8403059Z else 2025-12-04T08:58:29.8403405Z  git submodule foreach --recursive git clean -ffdx 2025-12-04T08:58:29.8403663Z fi 2025-12-04T08:58:29.8413493Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:58:29.8413792Z env: 2025-12-04T08:58:29.8413959Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:58:29.8414139Z NO_SUDO: true 2025-12-04T08:58:29.8414313Z ##[endgroup] 2025-12-04T08:58:29.8763054Z Entering 'android/libs/fbjni' 2025-12-04T08:58:29.8801410Z Entering 'third_party/FP16' 2025-12-04T08:58:29.8838368Z Entering 'third_party/FXdiv' 2025-12-04T08:58:29.8872355Z Entering 'third_party/NNPACK' 2025-12-04T08:58:29.8912355Z Entering 'third_party/NVTX' 2025-12-04T08:58:29.8955333Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:58:29.8990958Z Entering 'third_party/XNNPACK' 2025-12-04T08:58:29.9114815Z Entering 'third_party/aiter' 2025-12-04T08:58:29.9161050Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:58:29.9277480Z Entering 'third_party/benchmark' 2025-12-04T08:58:29.9315630Z Entering 'third_party/composable_kernel' 2025-12-04T08:58:29.9437543Z Entering 'third_party/cpp-httplib' 2025-12-04T08:58:29.9473535Z Entering 'third_party/cpuinfo' 2025-12-04T08:58:29.9513758Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:58:29.9552748Z Entering 'third_party/cutlass' 2025-12-04T08:58:29.9660798Z Entering 'third_party/fbgemm' 2025-12-04T08:58:29.9726550Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:58:29.9762698Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:58:29.9889541Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:58:29.9929077Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:58:30.0030822Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:58:30.0069695Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:58:30.0103224Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:58:30.0153244Z Entering 'third_party/flash-attention' 2025-12-04T08:58:30.0201222Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:58:30.0307583Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:58:30.0401008Z Entering 'third_party/flatbuffers' 2025-12-04T08:58:30.0477550Z Entering 'third_party/fmt' 2025-12-04T08:58:30.0514569Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:58:30.0552236Z Entering 'third_party/gloo' 2025-12-04T08:58:30.0595001Z Entering 'third_party/googletest' 2025-12-04T08:58:30.0633541Z Entering 'third_party/ideep' 2025-12-04T08:58:30.0672321Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:58:30.0760014Z Entering 'third_party/ittapi' 2025-12-04T08:58:30.0799554Z Entering 'third_party/kineto' 2025-12-04T08:58:30.0840806Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:58:30.0882124Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:58:30.0932696Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:58:30.0970395Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:58:30.1013796Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:58:30.1048309Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:58:30.1084914Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:58:30.1121604Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:58:30.1160513Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:58:30.1206710Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:58:30.1241728Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:58:30.1281872Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:58:30.1337278Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:58:30.1380549Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:58:30.1415684Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:58:30.1455838Z Entering 'third_party/kleidiai' 2025-12-04T08:58:30.1503580Z Entering 'third_party/mimalloc' 2025-12-04T08:58:30.1542758Z Entering 'third_party/nlohmann' 2025-12-04T08:58:30.1592339Z Entering 'third_party/onnx' 2025-12-04T08:58:30.1933059Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:58:30.1977146Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:58:30.2039924Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:58:30.2081418Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:58:30.2120828Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:58:30.2157127Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:58:30.2203429Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:58:30.2240112Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:58:30.2277664Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:58:30.2312654Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:58:30.2365200Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:58:30.2406146Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:58:30.2672202Z Entering 'third_party/pocketfft' 2025-12-04T08:58:30.2711646Z Entering 'third_party/protobuf' 2025-12-04T08:58:30.2794904Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:58:30.2831105Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:58:30.2873618Z Entering 'third_party/psimd' 2025-12-04T08:58:30.2914677Z Entering 'third_party/pthreadpool' 2025-12-04T08:58:30.2949970Z Entering 'third_party/pybind11' 2025-12-04T08:58:30.2990398Z Entering 'third_party/python-peachpy' 2025-12-04T08:58:30.3029058Z Entering 'third_party/sleef' 2025-12-04T08:58:30.3071992Z Entering 'third_party/tensorpipe' 2025-12-04T08:58:30.3111826Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:58:30.3149752Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:58:30.3190415Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:58:30.3231898Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:58:30.3270653Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:58:30.3431084Z Prepare all required actions 2025-12-04T08:58:30.3431549Z Getting action download info 2025-12-04T08:58:30.5211700Z ##[group]Run ./.github/actions/setup-linux 2025-12-04T08:58:30.5211924Z env: 2025-12-04T08:58:30.5212089Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:58:30.5212278Z ##[endgroup] 2025-12-04T08:58:30.5243286Z ##[group]Run set -euo pipefail 2025-12-04T08:58:30.5243549Z set -euo pipefail 2025-12-04T08:58:30.5243761Z function get_ec2_metadata() { 2025-12-04T08:58:30.5244036Z  # Pulled from instance metadata endpoint for EC2 2025-12-04T08:58:30.5244501Z  # see https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/instancedata-data-retrieval.html 2025-12-04T08:58:30.5244888Z  category=$1 2025-12-04T08:58:30.5245144Z  # If it is GCP runner (runner name contains gcp), do not run this 2025-12-04T08:58:30.5245451Z  runner_name_str=i-077765b42bd7d5290 2025-12-04T08:58:30.5245713Z  if [[ -f /.inarc ]]; then 2025-12-04T08:58:30.5245957Z  echo "ARC Runner, no info on ec2 metadata" 2025-12-04T08:58:30.5246250Z  elif [[ $runner_name_str == *"gcp"* ]]; then 2025-12-04T08:58:30.5246584Z  echo "Runner is from Google Cloud Platform, No info on ec2 metadata" 2025-12-04T08:58:30.5246881Z  else 2025-12-04T08:58:30.5247652Z  curl -H "X-aws-ec2-metadata-token: $(curl -s -X PUT "http://169.254.169.254/latest/api/token" -H "X-aws-ec2-metadata-token-ttl-seconds: 30")" -fsSL "http://169.254.169.254/latest/meta-data/${category}" 2025-12-04T08:58:30.5248274Z  fi 2025-12-04T08:58:30.5248421Z } 2025-12-04T08:58:30.5248604Z echo "ami-id: $(get_ec2_metadata ami-id)" 2025-12-04T08:58:30.5248894Z echo "instance-id: $(get_ec2_metadata instance-id)" 2025-12-04T08:58:30.5249226Z echo "instance-type: $(get_ec2_metadata instance-type)" 2025-12-04T08:58:30.5249512Z echo "system info $(uname -a)" 2025-12-04T08:58:30.5257105Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:58:30.5257379Z env: 2025-12-04T08:58:30.5257534Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:58:30.5257711Z ##[endgroup] 2025-12-04T08:58:30.5394308Z ami-id: ami-08982f1c5bf93d976 2025-12-04T08:58:30.5503247Z instance-id: i-077765b42bd7d5290 2025-12-04T08:58:30.5598663Z instance-type: g6.4xlarge 2025-12-04T08:58:30.5611550Z system info Linux ip-10-1-22-180.ec2.internal 6.1.150-174.273.amzn2023.x86_64 #1 SMP PREEMPT_DYNAMIC Tue Sep 9 12:21:26 UTC 2025 x86_64 x86_64 x86_64 GNU/Linux 2025-12-04T08:58:30.5630612Z ##[group]Run if [ -f /usr/bin/nvidia-smi ]; then nvidia-smi; fi 2025-12-04T08:58:30.5630961Z if [ -f /usr/bin/nvidia-smi ]; then nvidia-smi; fi 2025-12-04T08:58:30.5638244Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:58:30.5638528Z env: 2025-12-04T08:58:30.5638685Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:58:30.5638876Z ##[endgroup] 2025-12-04T08:58:31.9874392Z Thu Dec 4 08:58:31 2025 2025-12-04T08:58:31.9874810Z +-----------------------------------------------------------------------------------------+ 2025-12-04T08:58:31.9875319Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-12-04T08:58:31.9875805Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T08:58:31.9876298Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T08:58:31.9876798Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-12-04T08:58:31.9877190Z | | | MIG M. | 2025-12-04T08:58:31.9877493Z |=========================================+========================+======================| 2025-12-04T08:58:31.9942685Z | 0 NVIDIA L4 Off | 00000000:35:00.0 Off | 0 | 2025-12-04T08:58:31.9943521Z | N/A 41C P0 29W / 72W | 0MiB / 23034MiB | 4% Default | 2025-12-04T08:58:31.9943901Z | | | N/A | 2025-12-04T08:58:31.9944259Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T08:58:31.9944547Z 2025-12-04T08:58:31.9944700Z +-----------------------------------------------------------------------------------------+ 2025-12-04T08:58:31.9945089Z | Processes: | 2025-12-04T08:58:31.9945500Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T08:58:31.9945865Z | ID ID Usage | 2025-12-04T08:58:31.9946166Z |=========================================================================================| 2025-12-04T08:58:31.9947514Z | No running processes found | 2025-12-04T08:58:31.9947970Z +-----------------------------------------------------------------------------------------+ 2025-12-04T08:58:32.3159303Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T08:58:32.3160140Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T08:58:32.3170931Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:58:32.3171218Z env: 2025-12-04T08:58:32.3171388Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:58:32.3171570Z ##[endgroup] 2025-12-04T08:58:32.3232239Z ##[group]Run if systemctl is-active --quiet docker; then 2025-12-04T08:58:32.3232777Z if systemctl is-active --quiet docker; then 2025-12-04T08:58:32.3233272Z  echo "Docker daemon is running..."; 2025-12-04T08:58:32.3233747Z else 2025-12-04T08:58:32.3234180Z  echo "Starting docker daemon..." && sudo systemctl start docker; 2025-12-04T08:58:32.3234704Z fi 2025-12-04T08:58:32.3244613Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:58:32.3245099Z env: 2025-12-04T08:58:32.3245353Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:58:32.3245682Z ##[endgroup] 2025-12-04T08:58:32.3336009Z Docker daemon is running... 2025-12-04T08:58:32.3369771Z ##[group]Run nick-fields/retry@v3.0.0 2025-12-04T08:58:32.3369988Z with: 2025-12-04T08:58:32.3370139Z shell: bash 2025-12-04T08:58:32.3370304Z timeout_minutes: 5 2025-12-04T08:58:32.3370477Z max_attempts: 3 2025-12-04T08:58:32.3370641Z retry_wait_seconds: 30 2025-12-04T08:58:32.3372269Z command: AWS_ACCOUNT_ID=$(aws sts get-caller-identity|grep Account|cut -f4 -d\") aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" # For LF Runners we need to make sure we also login to Meta's ECR docker registry too. META_AWS_ACCOUNT_ID=308535385114 if [ "$AWS_ACCOUNT_ID" != "$META_AWS_ACCOUNT_ID" ] ; then aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$META_AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" fi 2025-12-04T08:58:32.3374026Z polling_interval_seconds: 1 2025-12-04T08:58:32.3374226Z warning_on_retry: true 2025-12-04T08:58:32.3374409Z continue_on_error: false 2025-12-04T08:58:32.3374580Z env: 2025-12-04T08:58:32.3374729Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:58:32.3374921Z AWS_RETRY_MODE: standard 2025-12-04T08:58:32.3375101Z AWS_MAX_ATTEMPTS: 5 2025-12-04T08:58:32.3375289Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T08:58:32.3375483Z ##[endgroup] 2025-12-04T08:58:33.3307540Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T08:58:33.3308471Z Configure a credential helper to remove this warning. See 2025-12-04T08:58:33.3309036Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T08:58:33.3309400Z 2025-12-04T08:58:33.3309489Z Login Succeeded 2025-12-04T08:58:33.7749448Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T08:58:33.7750210Z Configure a credential helper to remove this warning. See 2025-12-04T08:58:33.7750862Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T08:58:33.7751200Z 2025-12-04T08:58:33.7751304Z Login Succeeded 2025-12-04T08:58:34.4135500Z Command completed after 1 attempt(s). 2025-12-04T08:58:34.4195467Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T08:58:34.4195844Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T08:58:34.4196165Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T08:58:34.4205545Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:58:34.4205847Z env: 2025-12-04T08:58:34.4206017Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:58:34.4206210Z ##[endgroup] 2025-12-04T08:58:34.4298842Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T08:58:34.4299249Z # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T08:58:34.4299755Z # shellcheck disable=SC2046 2025-12-04T08:58:34.4299998Z docker stop $(docker ps -q) || true 2025-12-04T08:58:34.4300245Z # Prune all of the docker images 2025-12-04T08:58:34.4300480Z docker system prune -af 2025-12-04T08:58:34.4307692Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:58:34.4307981Z env: 2025-12-04T08:58:34.4308141Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:58:34.4308334Z ##[endgroup] 2025-12-04T08:58:34.4542200Z "docker stop" requires at least 1 argument. 2025-12-04T08:58:34.4542580Z See 'docker stop --help'. 2025-12-04T08:58:34.4542759Z 2025-12-04T08:58:34.4542936Z Usage: docker stop [OPTIONS] CONTAINER [CONTAINER...] 2025-12-04T08:58:34.4543186Z 2025-12-04T08:58:34.4543286Z Stop one or more running containers 2025-12-04T08:58:34.4764456Z Total reclaimed space: 0B 2025-12-04T08:58:34.4912250Z ##[group]Run pytorch/test-infra/.github/actions/calculate-docker-image@main 2025-12-04T08:58:34.4912636Z with: 2025-12-04T08:58:34.4913209Z docker-image-name: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:34.4913847Z use-custom-docker-registry: true 2025-12-04T08:58:34.4914087Z docker-build-dir: .ci/docker 2025-12-04T08:58:34.4914297Z docker-build-script: ./build.sh 2025-12-04T08:58:34.4914508Z working-directory: . 2025-12-04T08:58:34.4914751Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:58:34.4915042Z force-push: false 2025-12-04T08:58:34.4915202Z env: 2025-12-04T08:58:34.4915346Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:58:34.4915526Z ##[endgroup] 2025-12-04T08:58:34.4931670Z ##[group]Run set -ex 2025-12-04T08:58:34.4931895Z set -ex 2025-12-04T08:58:34.4932065Z  2025-12-04T08:58:34.4932391Z # If the docker build directory or the build script doesn't exist, the action will 2025-12-04T08:58:34.4932878Z # gracefully return the docker image name as it is. Pulling docker image in Linux 2025-12-04T08:58:34.4933392Z # job could then download the pre-built image as usual 2025-12-04T08:58:34.4933886Z if [[ -d "${DOCKER_BUILD_DIR}" ]] && [[ -f "${DOCKER_BUILD_DIR}/${DOCKER_BUILD_SCRIPT}" ]] && [[ "${USE_CUSTOM_DOCKER_REGISTRY}" == "true" ]]; then 2025-12-04T08:58:34.4934336Z  echo "skip=false" >> "${GITHUB_OUTPUT}" 2025-12-04T08:58:34.4934589Z else 2025-12-04T08:58:34.4934788Z  echo "skip=true" >> "${GITHUB_OUTPUT}" 2025-12-04T08:58:34.4935112Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-12-04T08:58:34.4935404Z  2025-12-04T08:58:34.4935799Z  echo "Not using custom ECR registry. Either it was not requested or there is no Docker build script in the ${REPO_NAME} repo..." 2025-12-04T08:58:34.4936258Z  exit 0 2025-12-04T08:58:34.4936417Z fi 2025-12-04T08:58:34.4936578Z  2025-12-04T08:58:34.4936830Z if [[ "${DOCKER_IMAGE_NAME}" == *"${DOCKER_REGISTRY}/${REPO_NAME}"* ]]; then 2025-12-04T08:58:34.4937264Z  # The docker image name already includes the ECR prefix and tag, so we can just 2025-12-04T08:58:34.4937639Z  # use it as it is, but first let's extract the tag 2025-12-04T08:58:34.4937985Z  DOCKER_TAG=$(echo "${DOCKER_IMAGE_NAME}" | awk -F '[:,]' '{print $2}') 2025-12-04T08:58:34.4938349Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T08:58:34.4938693Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-12-04T08:58:34.4938978Z else 2025-12-04T08:58:34.4939171Z  if [[ "${DOCKER_IMAGE_NAME}" == *:* ]]; then 2025-12-04T08:58:34.4939446Z  CUSTOM_TAG_PREFIX=${DOCKER_IMAGE_NAME#*:} 2025-12-04T08:58:34.4939735Z  DOCKER_IMAGE_NAME=${DOCKER_IMAGE_NAME%%:*} 2025-12-04T08:58:34.4939976Z  fi 2025-12-04T08:58:34.4940462Z  DOCKER_TAG=${CUSTOM_TAG_PREFIX:+${CUSTOM_TAG_PREFIX}-}$(git rev-parse HEAD:"${DOCKER_BUILD_DIR}") 2025-12-04T08:58:34.4940892Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T08:58:34.4941342Z  echo "docker-image=${DOCKER_REGISTRY}/${REPO_NAME}/${DOCKER_IMAGE_NAME}:${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T08:58:34.4941835Z  echo "custom-tag-prefix=${CUSTOM_TAG_PREFIX}" >> "${GITHUB_OUTPUT}" 2025-12-04T08:58:34.4942137Z fi 2025-12-04T08:58:34.4949659Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:58:34.4949937Z env: 2025-12-04T08:58:34.4950101Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:58:34.4950294Z REPO_NAME: pytorch 2025-12-04T08:58:34.4951001Z DOCKER_IMAGE_NAME: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:34.4951648Z DOCKER_BUILD_DIR: .ci/docker 2025-12-04T08:58:34.4951869Z DOCKER_BUILD_SCRIPT: ./build.sh 2025-12-04T08:58:34.4952147Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:58:34.4952447Z USE_CUSTOM_DOCKER_REGISTRY: true 2025-12-04T08:58:34.4952665Z CUSTOM_TAG_PREFIX: 2025-12-04T08:58:34.4952841Z ##[endgroup] 2025-12-04T08:58:34.4979279Z + [[ -d .ci/docker ]] 2025-12-04T08:58:34.4979571Z + [[ -f .ci/docker/./build.sh ]] 2025-12-04T08:58:34.4979829Z + [[ true == \t\r\u\e ]] 2025-12-04T08:58:34.4980054Z + echo skip=false 2025-12-04T08:58:34.4981008Z + [[ 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a == *\3\0\8\5\3\5\3\8\5\1\1\4\.\d\k\r\.\e\c\r\.\u\s\-\e\a\s\t\-\1\.\a\m\a\z\o\n\a\w\s\.\c\o\m\/\p\y\t\o\r\c\h* ]] 2025-12-04T08:58:34.4986943Z ++ echo 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:34.4987869Z ++ awk -F '[:,]' '{print $2}' 2025-12-04T08:58:34.5011152Z + DOCKER_TAG=pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:34.5011990Z + echo docker-tag=pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:34.5013020Z + echo docker-image=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:34.5036933Z ##[group]Run set +e 2025-12-04T08:58:34.5037140Z set +e 2025-12-04T08:58:34.5037294Z set -x 2025-12-04T08:58:34.5037446Z  2025-12-04T08:58:34.5037592Z login() { 2025-12-04T08:58:34.5037933Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-12-04T08:58:34.5038304Z } 2025-12-04T08:58:34.5038459Z  2025-12-04T08:58:34.5038599Z retry () { 2025-12-04T08:58:34.5038783Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-12-04T08:58:34.5039005Z } 2025-12-04T08:58:34.5039140Z  2025-12-04T08:58:34.5039308Z retry login "${DOCKER_REGISTRY}" 2025-12-04T08:58:34.5039527Z  2025-12-04T08:58:34.5039677Z START_TIME=$(date +%s) 2025-12-04T08:58:34.5039877Z # Wait up to 120 minutes 2025-12-04T08:58:34.5040136Z while [[ $(( $(date +%s) - 7200 )) -lt $START_TIME ]]; do 2025-12-04T08:58:34.5040478Z  # Check if image already exists, if it does then skip building it 2025-12-04T08:58:34.5040818Z  if docker manifest inspect "${DOCKER_IMAGE}"; then 2025-12-04T08:58:34.5041077Z  exit 0 2025-12-04T08:58:34.5041240Z  fi 2025-12-04T08:58:34.5041382Z  2025-12-04T08:58:34.5041653Z  # NB: This flag is used by Docker build workflow to push the image to ECR, so we can 2025-12-04T08:58:34.5042115Z  # use this to differentiate between the Docker build and regular build jobs. For the 2025-12-04T08:58:34.5042701Z  # latter, it will wait for the Docker images to become available before continuing 2025-12-04T08:58:34.5043057Z  if [ "${DOCKER_PUSH:-false}" == "true" ]; then 2025-12-04T08:58:34.5043341Z  # It's a Docker build job, let's build the image 2025-12-04T08:58:34.5043590Z  break 2025-12-04T08:58:34.5043755Z  else 2025-12-04T08:58:34.5043991Z  # It's a regular build job, wait for the image to become available 2025-12-04T08:58:34.5044282Z  sleep 300 2025-12-04T08:58:34.5044464Z  fi 2025-12-04T08:58:34.5044607Z done 2025-12-04T08:58:34.5044754Z  2025-12-04T08:58:34.5045002Z # NB: This part requires a full checkout. Otherwise, the merge base will 2025-12-04T08:58:34.5045525Z # be empty. The default action would be to continue rebuild the image 2025-12-04T08:58:34.5045917Z if [[ "$BASE_REVISION" = "$(git rev-parse HEAD)" ]]; then 2025-12-04T08:58:34.5046255Z  # if we're on the base branch then use the parent commit 2025-12-04T08:58:34.5046538Z  MERGE_BASE=$(git rev-parse HEAD~) 2025-12-04T08:58:34.5046750Z else 2025-12-04T08:58:34.5046975Z  # otherwise we're on a PR, so use the most recent base commit 2025-12-04T08:58:34.5047305Z  MERGE_BASE=$(git merge-base HEAD "$BASE_REVISION") 2025-12-04T08:58:34.5047546Z fi 2025-12-04T08:58:34.5047689Z  2025-12-04T08:58:34.5047849Z if [[ -z "${MERGE_BASE}" ]]; then 2025-12-04T08:58:34.5048096Z  echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-12-04T08:58:34.5048323Z  2025-12-04T08:58:34.5048655Z  echo "Finding merge base only works with full checkout, please set fetch-depth to 0, continuing ..." 2025-12-04T08:58:34.5049043Z  exit 0 2025-12-04T08:58:34.5049197Z fi 2025-12-04T08:58:34.5049337Z  2025-12-04T08:58:34.5049562Z if ! git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}"; then 2025-12-04T08:58:34.5050038Z  echo "Directory '${DOCKER_BUILD_DIR}' not found in commit $MERGE_BASE, you should rebase onto a more recent commit" 2025-12-04T08:58:34.5050435Z  exit 1 2025-12-04T08:58:34.5050587Z fi 2025-12-04T08:58:34.5050731Z  2025-12-04T08:58:34.5050973Z PREVIOUS_DOCKER_TAG=$(git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}") 2025-12-04T08:58:34.5051433Z # If no image exists but the hash is the same as the previous hash then we should error out here 2025-12-04T08:58:34.5051840Z if [[ "${PREVIOUS_DOCKER_TAG}" == "${DOCKER_TAG}" ]]; then 2025-12-04T08:58:34.5052327Z  echo "WARNING: Something has gone wrong and the previous image isn't available for the merge-base of your branch" 2025-12-04T08:58:34.5052862Z  echo " Will re-build docker image to store in local cache, TTS may be longer" 2025-12-04T08:58:34.5053295Z fi 2025-12-04T08:58:34.5053450Z  2025-12-04T08:58:34.5053627Z echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-12-04T08:58:34.5060413Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:58:34.5060700Z env: 2025-12-04T08:58:34.5060861Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:58:34.5061049Z DOCKER_BUILD_DIR: .ci/docker 2025-12-04T08:58:34.5061295Z BASE_REVISION: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T08:58:34.5061948Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:34.5062735Z DOCKER_TAG: pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:34.5063209Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:58:34.5063488Z DOCKER_PUSH: 2025-12-04T08:58:34.5063659Z ##[endgroup] 2025-12-04T08:58:34.5087902Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:58:34.5088651Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:58:34.5090727Z + aws ecr get-login-password --region us-east-1 2025-12-04T08:58:34.5091778Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:58:34.9601754Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T08:58:34.9602452Z Configure a credential helper to remove this warning. See 2025-12-04T08:58:34.9603065Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T08:58:34.9603485Z 2025-12-04T08:58:34.9606096Z Login Succeeded 2025-12-04T08:58:34.9626119Z ++ date +%s 2025-12-04T08:58:34.9636552Z + START_TIME=1764838714 2025-12-04T08:58:34.9640304Z ++ date +%s 2025-12-04T08:58:34.9656654Z + [[ 1764831514 -lt 1764838714 ]] 2025-12-04T08:58:34.9657521Z + docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:35.1485566Z { 2025-12-04T08:58:35.1485989Z "schemaVersion": 2, 2025-12-04T08:58:35.1486335Z "mediaType": "application/vnd.docker.distribution.manifest.v2+json", 2025-12-04T08:58:35.1486660Z "config": { 2025-12-04T08:58:35.1486919Z "mediaType": "application/vnd.docker.container.image.v1+json", 2025-12-04T08:58:35.1487244Z "size": 34864, 2025-12-04T08:58:35.1487538Z "digest": "sha256:add7313791033822205cdb3cf32096534b2cfaa4855bd48119b59000bfe00301" 2025-12-04T08:58:35.1487867Z }, 2025-12-04T08:58:35.1488011Z "layers": [ 2025-12-04T08:58:35.1488161Z { 2025-12-04T08:58:35.1488398Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1488701Z "size": 30447951, 2025-12-04T08:58:35.1489024Z "digest": "sha256:63e5bc7682b85ae57a1221210f64d62e7a90b0a30f19af4ca734b8242ae49d63" 2025-12-04T08:58:35.1489347Z }, 2025-12-04T08:58:35.1489486Z { 2025-12-04T08:58:35.1489731Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1490028Z "size": 1554, 2025-12-04T08:58:35.1490310Z "digest": "sha256:0678d56345c994444b77bb70b1177189d23e794748b1d75ffc45d227c7dea94a" 2025-12-04T08:58:35.1490625Z }, 2025-12-04T08:58:35.1490750Z { 2025-12-04T08:58:35.1490974Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1491259Z "size": 313275661, 2025-12-04T08:58:35.1491553Z "digest": "sha256:45f5c9ddfce78349dff3d5edfbaa0310ae17311f66abdcd7e00fa21b500e801c" 2025-12-04T08:58:35.1491877Z }, 2025-12-04T08:58:35.1492009Z { 2025-12-04T08:58:35.1492218Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1492501Z "size": 787, 2025-12-04T08:58:35.1492778Z "digest": "sha256:086b1df51ac1162d9c45698e9dfaf91c6c222c8bd9ab01797ac8f9344bc8044f" 2025-12-04T08:58:35.1493094Z }, 2025-12-04T08:58:35.1493358Z { 2025-12-04T08:58:35.1493578Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1493857Z "size": 106, 2025-12-04T08:58:35.1494162Z "digest": "sha256:fe8a7b64bf98352f89057bcba66beef2fb44cc05fbd3606abccd8e86cf476234" 2025-12-04T08:58:35.1494669Z }, 2025-12-04T08:58:35.1494887Z { 2025-12-04T08:58:35.1495118Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1495397Z "size": 703, 2025-12-04T08:58:35.1495665Z "digest": "sha256:7680723e9a578033dd106b45784c639f06cc8adb1f5239ec513d9de01087c1af" 2025-12-04T08:58:35.1495970Z }, 2025-12-04T08:58:35.1496094Z { 2025-12-04T08:58:35.1496311Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1496583Z "size": 1216, 2025-12-04T08:58:35.1496858Z "digest": "sha256:9c5027aeeb4e3101f48c1d2e400c387110e1009e42497ee801f1b4b7f7edb5c0" 2025-12-04T08:58:35.1497185Z }, 2025-12-04T08:58:35.1497316Z { 2025-12-04T08:58:35.1497534Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1498052Z "size": 483, 2025-12-04T08:58:35.1498316Z "digest": "sha256:9a56521103600bd37a1e7c1191b5136c2d738c092f8a6701499f7068a32c2628" 2025-12-04T08:58:35.1498627Z }, 2025-12-04T08:58:35.1498753Z { 2025-12-04T08:58:35.1498968Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1499248Z "size": 110361875, 2025-12-04T08:58:35.1499527Z "digest": "sha256:375c4427e9141269458333b1463fdb219e736fd6231ec1c56c625c48437ace77" 2025-12-04T08:58:35.1499834Z }, 2025-12-04T08:58:35.1499958Z { 2025-12-04T08:58:35.1500178Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1500458Z "size": 4961, 2025-12-04T08:58:35.1500726Z "digest": "sha256:a86faaa7dbdd70e678e5ea20072637ee42618921ca8f80ca089f789325d4b0c2" 2025-12-04T08:58:35.1501040Z }, 2025-12-04T08:58:35.1501175Z { 2025-12-04T08:58:35.1501520Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1501807Z "size": 1755, 2025-12-04T08:58:35.1502087Z "digest": "sha256:fb7848686804957915d98f8655ef6da0fe4c521b50a82aefdebf475983505a15" 2025-12-04T08:58:35.1502410Z }, 2025-12-04T08:58:35.1502535Z { 2025-12-04T08:58:35.1502765Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1503058Z "size": 724, 2025-12-04T08:58:35.1503328Z "digest": "sha256:3541df015cdb7e8925273399d28e56c31b3c9196f00439ac2925537b173b1f84" 2025-12-04T08:58:35.1503647Z }, 2025-12-04T08:58:35.1503773Z { 2025-12-04T08:58:35.1503985Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1504271Z "size": 543, 2025-12-04T08:58:35.1504898Z "digest": "sha256:79dc80f426b29d4ae9157b967050b03e66aa0c4b1295b944a1dd70106be87066" 2025-12-04T08:58:35.1505224Z }, 2025-12-04T08:58:35.1505360Z { 2025-12-04T08:58:35.1505585Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1505864Z "size": 3185190117, 2025-12-04T08:58:35.1506155Z "digest": "sha256:a13fcc1b90bb9c251ebe7ef2a03c4cb3afa1c8bdafe84f5f85136773059a3735" 2025-12-04T08:58:35.1506481Z }, 2025-12-04T08:58:35.1506605Z { 2025-12-04T08:58:35.1506809Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1507089Z "size": 32, 2025-12-04T08:58:35.1507365Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T08:58:35.1507678Z }, 2025-12-04T08:58:35.1507802Z { 2025-12-04T08:58:35.1508014Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1508283Z "size": 396, 2025-12-04T08:58:35.1508550Z "digest": "sha256:549db4d6c618ecd9534658a233e3c90508f82d8735f965c2786b2eaa078869e5" 2025-12-04T08:58:35.1508874Z }, 2025-12-04T08:58:35.1508998Z { 2025-12-04T08:58:35.1509207Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1509488Z "size": 236860, 2025-12-04T08:58:35.1509764Z "digest": "sha256:5c63528cb580001e65104f4cb0809bf0673a00f989a7db42fd6d86aa1ec27cee" 2025-12-04T08:58:35.1510074Z }, 2025-12-04T08:58:35.1510200Z { 2025-12-04T08:58:35.1510466Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1510756Z "size": 231, 2025-12-04T08:58:35.1511029Z "digest": "sha256:75bd83b989a44e4d4119a3f972891025eb0e9ce95cfbe4a0ca5cdbe7130028d6" 2025-12-04T08:58:35.1511497Z }, 2025-12-04T08:58:35.1511632Z { 2025-12-04T08:58:35.1511859Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1512145Z "size": 3043497, 2025-12-04T08:58:35.1512427Z "digest": "sha256:de6e78970f517178cb91f36cd02bd9ca7b72a08fb82a0f9007516026f258c035" 2025-12-04T08:58:35.1512743Z }, 2025-12-04T08:58:35.1512872Z { 2025-12-04T08:58:35.1513089Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1513368Z "size": 1472, 2025-12-04T08:58:35.1513660Z "digest": "sha256:e13ed7c7e4736e81dc21af755b3363eb26e4d3b2f1ca988dfe65effa47d8fa42" 2025-12-04T08:58:35.1514143Z }, 2025-12-04T08:58:35.1514280Z { 2025-12-04T08:58:35.1514504Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1514801Z "size": 481, 2025-12-04T08:58:35.1515081Z "digest": "sha256:6e2949bcb74152577a0f20c38bcb6dd80f5e68427e3e531a80e08c9ecc73a979" 2025-12-04T08:58:35.1515392Z }, 2025-12-04T08:58:35.1515521Z { 2025-12-04T08:58:35.1515741Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1516013Z "size": 202, 2025-12-04T08:58:35.1516288Z "digest": "sha256:14d69d9aaec70287efd2fd35c4f93e43a29a4098458cc9fca1c93f02ad7356cb" 2025-12-04T08:58:35.1516610Z }, 2025-12-04T08:58:35.1516731Z { 2025-12-04T08:58:35.1516944Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1517224Z "size": 607, 2025-12-04T08:58:35.1517617Z "digest": "sha256:5c02769dd8e5bba2f7f5fd84bde9595fcb3bdbffcae497503fa846f9b5e78bf5" 2025-12-04T08:58:35.1517944Z }, 2025-12-04T08:58:35.1518083Z { 2025-12-04T08:58:35.1518303Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1518578Z "size": 7889619584, 2025-12-04T08:58:35.1518867Z "digest": "sha256:35041ce524ac4afec40ecd73b1393c830614f1f79d43a6439767a6c7d5b7027b" 2025-12-04T08:58:35.1519179Z }, 2025-12-04T08:58:35.1519298Z { 2025-12-04T08:58:35.1519511Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1519784Z "size": 830, 2025-12-04T08:58:35.1520046Z "digest": "sha256:2fa92dc5885e080e049ceb4139288b6c0e39fab34256945708b08ea55a1f7a0b" 2025-12-04T08:58:35.1520354Z }, 2025-12-04T08:58:35.1520482Z { 2025-12-04T08:58:35.1520688Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1520964Z "size": 33451739, 2025-12-04T08:58:35.1521259Z "digest": "sha256:2b85eafbd92a0e70a0a70154ad8bf4584095e576d95873368f30373f5966714a" 2025-12-04T08:58:35.1521580Z }, 2025-12-04T08:58:35.1521705Z { 2025-12-04T08:58:35.1521930Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1522212Z "size": 104, 2025-12-04T08:58:35.1522487Z "digest": "sha256:ff755a4ddad7880f23c6b767d432d6f1eafdb62b3ea18f8a98e22c441c099fcb" 2025-12-04T08:58:35.1522808Z }, 2025-12-04T08:58:35.1522937Z { 2025-12-04T08:58:35.1523146Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1523438Z "size": 1496, 2025-12-04T08:58:35.1523712Z "digest": "sha256:09eb41bdf42d8605b57b2363348154140904dec914b34a67298b82122bfce2b3" 2025-12-04T08:58:35.1524014Z }, 2025-12-04T08:58:35.1524140Z { 2025-12-04T08:58:35.1524359Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1524637Z "size": 458787828, 2025-12-04T08:58:35.1524920Z "digest": "sha256:11ede4d59e935e62f41b33220fe871794ab5e57ce724173b713368977683bcf6" 2025-12-04T08:58:35.1525235Z }, 2025-12-04T08:58:35.1525362Z { 2025-12-04T08:58:35.1525571Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1525857Z "size": 164, 2025-12-04T08:58:35.1526134Z "digest": "sha256:1283cd8f801a142172f3ab76fd472df8583223d9437de3e4d18d8cf98ea3fa98" 2025-12-04T08:58:35.1526446Z }, 2025-12-04T08:58:35.1526576Z { 2025-12-04T08:58:35.1526797Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1527083Z "size": 346, 2025-12-04T08:58:35.1527420Z "digest": "sha256:024fa855425fa524ad4500660cf61d53be62b99556d31b8b280d14caba434a35" 2025-12-04T08:58:35.1527815Z }, 2025-12-04T08:58:35.1527938Z { 2025-12-04T08:58:35.1528155Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1528440Z "size": 32, 2025-12-04T08:58:35.1528713Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T08:58:35.1529035Z }, 2025-12-04T08:58:35.1529170Z { 2025-12-04T08:58:35.1529394Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1529754Z "size": 106, 2025-12-04T08:58:35.1530023Z "digest": "sha256:303e6747a62efecf5efa1f97d0e66b40a3b39da8d79a51f75b89f4c92ae7ec52" 2025-12-04T08:58:35.1530339Z }, 2025-12-04T08:58:35.1530458Z { 2025-12-04T08:58:35.1530666Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1530943Z "size": 424, 2025-12-04T08:58:35.1531214Z "digest": "sha256:3017cdf4838bcc9a33daebc07487f8ae1f6bd6e7ce8322c14f5480e8db9ef90e" 2025-12-04T08:58:35.1531532Z }, 2025-12-04T08:58:35.1531656Z { 2025-12-04T08:58:35.1531860Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1532140Z "size": 19309374, 2025-12-04T08:58:35.1532427Z "digest": "sha256:6b6cd1c358e886dc6ed7fd46ac4bcc1a0a73b7b1301739ea1953478ee5d83f50" 2025-12-04T08:58:35.1532744Z }, 2025-12-04T08:58:35.1532865Z { 2025-12-04T08:58:35.1533291Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1533584Z "size": 108, 2025-12-04T08:58:35.1533854Z "digest": "sha256:b2dd045011241d1cf8889e2a7369d9fe4844dfe15529b520ccd6a59bd3c1532e" 2025-12-04T08:58:35.1534166Z }, 2025-12-04T08:58:35.1534291Z { 2025-12-04T08:58:35.1534497Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1534776Z "size": 827, 2025-12-04T08:58:35.1535059Z "digest": "sha256:55adc51fe5897031d4cf2f2b8fd162213f6e46a52848630c616606271b97952e" 2025-12-04T08:58:35.1535375Z }, 2025-12-04T08:58:35.1535505Z { 2025-12-04T08:58:35.1535722Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1535991Z "size": 724, 2025-12-04T08:58:35.1536255Z "digest": "sha256:3541df015cdb7e8925273399d28e56c31b3c9196f00439ac2925537b173b1f84" 2025-12-04T08:58:35.1536565Z }, 2025-12-04T08:58:35.1536691Z { 2025-12-04T08:58:35.1536903Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1537181Z "size": 149, 2025-12-04T08:58:35.1537446Z "digest": "sha256:a43ca0e4b837964b12b7469194cfe939c26de027298040028975324dce25938a" 2025-12-04T08:58:35.1537760Z }, 2025-12-04T08:58:35.1537890Z { 2025-12-04T08:58:35.1538104Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1538376Z "size": 138, 2025-12-04T08:58:35.1538648Z "digest": "sha256:b7212f17fd1404837fcfdd086dd0e2667931e4db377d45d8d89a44390c84e11d" 2025-12-04T08:58:35.1538969Z }, 2025-12-04T08:58:35.1539091Z { 2025-12-04T08:58:35.1539303Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1539583Z "size": 141, 2025-12-04T08:58:35.1539851Z "digest": "sha256:083e42cac090e6486c35f392b64ee54448f5e4aa947003aeb3e1f92c8ea5c099" 2025-12-04T08:58:35.1540160Z }, 2025-12-04T08:58:35.1540285Z { 2025-12-04T08:58:35.1540496Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1540771Z "size": 32, 2025-12-04T08:58:35.1541045Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T08:58:35.1541366Z }, 2025-12-04T08:58:35.1541497Z { 2025-12-04T08:58:35.1541712Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1541986Z "size": 223, 2025-12-04T08:58:35.1542250Z "digest": "sha256:0a00b784a4aac341795729b254f7edd09e811b7f51d0c58e0e6bfeeee6940503" 2025-12-04T08:58:35.1542563Z }, 2025-12-04T08:58:35.1542688Z { 2025-12-04T08:58:35.1542894Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1543169Z "size": 255, 2025-12-04T08:58:35.1543445Z "digest": "sha256:c6173c779f7ba143a21214ea5f032b141863a37ceb4c0ac01d3248c216ce5241" 2025-12-04T08:58:35.1543759Z }, 2025-12-04T08:58:35.1543878Z { 2025-12-04T08:58:35.1544096Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1544377Z "size": 145520672, 2025-12-04T08:58:35.1544664Z "digest": "sha256:ed3d1e3387b924585c332bf1bc252fa159cd0d25256a874043ff0141b1ab5ff7" 2025-12-04T08:58:35.1545047Z }, 2025-12-04T08:58:35.1545333Z { 2025-12-04T08:58:35.1545617Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1546114Z "size": 106, 2025-12-04T08:58:35.1546476Z "digest": "sha256:b29343478586aeee19d2a622661716f6f1591280c890f49b727a8da13a610784" 2025-12-04T08:58:35.1546852Z }, 2025-12-04T08:58:35.1573320Z { 2025-12-04T08:58:35.1573651Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1574000Z "size": 312293530, 2025-12-04T08:58:35.1574300Z "digest": "sha256:c6f0520487fb506bc4601fd84d5f28d8a76b203e004731e4b2067c2ab1a14e0b" 2025-12-04T08:58:35.1574698Z }, 2025-12-04T08:58:35.1574855Z { 2025-12-04T08:58:35.1575130Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1575461Z "size": 3058011133, 2025-12-04T08:58:35.1575965Z "digest": "sha256:148171691cd4c4d20310d490d4b4dd903490d04ea07fb8f7e668a28768683e9a" 2025-12-04T08:58:35.1576356Z }, 2025-12-04T08:58:35.1576521Z { 2025-12-04T08:58:35.1576776Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1577055Z "size": 129, 2025-12-04T08:58:35.1577353Z "digest": "sha256:2c666d30ed77fff9ff1167d41cd645dad98280fcbe941f5bc3828c7ae66b1287" 2025-12-04T08:58:35.1577687Z }, 2025-12-04T08:58:35.1577820Z { 2025-12-04T08:58:35.1578051Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1578349Z "size": 880, 2025-12-04T08:58:35.1578641Z "digest": "sha256:5d8d3a0a98e012c5068e0f3bae5a03e3148ecf2d063634eee4c9241a1e3fdfb5" 2025-12-04T08:58:35.1578965Z }, 2025-12-04T08:58:35.1579085Z { 2025-12-04T08:58:35.1579306Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1579587Z "size": 724, 2025-12-04T08:58:35.1579868Z "digest": "sha256:3541df015cdb7e8925273399d28e56c31b3c9196f00439ac2925537b173b1f84" 2025-12-04T08:58:35.1580183Z }, 2025-12-04T08:58:35.1580309Z { 2025-12-04T08:58:35.1580531Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1580815Z "size": 139, 2025-12-04T08:58:35.1581088Z "digest": "sha256:b06bafce9e817295d8127207747c80aa18e04392ff0875844fc30a1e794a8a0c" 2025-12-04T08:58:35.1581398Z }, 2025-12-04T08:58:35.1581517Z { 2025-12-04T08:58:35.1581727Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1582001Z "size": 32, 2025-12-04T08:58:35.1582285Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T08:58:35.1582612Z }, 2025-12-04T08:58:35.1582750Z { 2025-12-04T08:58:35.1582962Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1583246Z "size": 159, 2025-12-04T08:58:35.1583534Z "digest": "sha256:15e0d7e4590d3d8f598d05aec3a92f891bf8b4605bcc38cc2de852b6014ef8f3" 2025-12-04T08:58:35.1583860Z }, 2025-12-04T08:58:35.1583986Z { 2025-12-04T08:58:35.1584205Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1584539Z "size": 1011, 2025-12-04T08:58:35.1584879Z "digest": "sha256:a514bd1add3164d8d7ca99aa19294c4ed8b97b074635d98714c4f598a959f4cd" 2025-12-04T08:58:35.1585261Z }, 2025-12-04T08:58:35.1585413Z { 2025-12-04T08:58:35.1585666Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1585998Z "size": 724, 2025-12-04T08:58:35.1586284Z "digest": "sha256:3541df015cdb7e8925273399d28e56c31b3c9196f00439ac2925537b173b1f84" 2025-12-04T08:58:35.1586592Z }, 2025-12-04T08:58:35.1586725Z { 2025-12-04T08:58:35.1586956Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1587232Z "size": 134, 2025-12-04T08:58:35.1587508Z "digest": "sha256:57b84ee6000204f27a1d9bca199b19be4c86ecd324540dbdf239c56a6c3b34ea" 2025-12-04T08:58:35.1587824Z }, 2025-12-04T08:58:35.1587952Z { 2025-12-04T08:58:35.1588164Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1588442Z "size": 32, 2025-12-04T08:58:35.1588805Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T08:58:35.1589120Z }, 2025-12-04T08:58:35.1589247Z { 2025-12-04T08:58:35.1589470Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1589752Z "size": 157, 2025-12-04T08:58:35.1590035Z "digest": "sha256:b8babeff6d817a5961dddc15c6bdfdbd05da187fae75d5804015f99fd7c066d8" 2025-12-04T08:58:35.1590376Z }, 2025-12-04T08:58:35.1590498Z { 2025-12-04T08:58:35.1590714Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1590992Z "size": 602, 2025-12-04T08:58:35.1591262Z "digest": "sha256:83779ddf6a85ab387f64a45f274cba245b69e4fd1931ff0b5d7d3efd4b7a43bc" 2025-12-04T08:58:35.1591579Z }, 2025-12-04T08:58:35.1591707Z { 2025-12-04T08:58:35.1592008Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1592287Z "size": 724, 2025-12-04T08:58:35.1592562Z "digest": "sha256:3541df015cdb7e8925273399d28e56c31b3c9196f00439ac2925537b173b1f84" 2025-12-04T08:58:35.1592878Z }, 2025-12-04T08:58:35.1593002Z { 2025-12-04T08:58:35.1593218Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1593496Z "size": 155, 2025-12-04T08:58:35.1593763Z "digest": "sha256:8b7620c0d736cc79381207ce5afe2af90f0cd7f0cd394577d2c9520d7f74762f" 2025-12-04T08:58:35.1594081Z }, 2025-12-04T08:58:35.1594209Z { 2025-12-04T08:58:35.1594420Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1594705Z "size": 32, 2025-12-04T08:58:35.1594981Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T08:58:35.1595303Z }, 2025-12-04T08:58:35.1595426Z { 2025-12-04T08:58:35.1595654Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1595937Z "size": 188, 2025-12-04T08:58:35.1596211Z "digest": "sha256:3bcfa090e4efd3677425f76baea9f1e0c50a75d8c6b5713ec05310f1dff24539" 2025-12-04T08:58:35.1596539Z }, 2025-12-04T08:58:35.1596674Z { 2025-12-04T08:58:35.1596887Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1597173Z "size": 1370, 2025-12-04T08:58:35.1597461Z "digest": "sha256:eb0504ec4d9218a79896b604f73dc0ea5a0f96266ad9c2cdbbbe5f0f18222694" 2025-12-04T08:58:35.1597776Z }, 2025-12-04T08:58:35.1597910Z { 2025-12-04T08:58:35.1598128Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1598401Z "size": 32, 2025-12-04T08:58:35.1598674Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T08:58:35.1598996Z }, 2025-12-04T08:58:35.1599127Z { 2025-12-04T08:58:35.1599335Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1599614Z "size": 136, 2025-12-04T08:58:35.1599890Z "digest": "sha256:15d0fec09d7b196a1462d51516ee90fc3443ba178d3e56d59cacf32146b4321d" 2025-12-04T08:58:35.1600205Z }, 2025-12-04T08:58:35.1600331Z { 2025-12-04T08:58:35.1600547Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1600820Z "size": 528, 2025-12-04T08:58:35.1601093Z "digest": "sha256:cca81fcc62a949959ca4dd3c9056fb293d548ef8607127eeeef6cfd3a8897ca8" 2025-12-04T08:58:35.1601423Z }, 2025-12-04T08:58:35.1601549Z { 2025-12-04T08:58:35.1601760Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1602034Z "size": 32, 2025-12-04T08:58:35.1602298Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T08:58:35.1602617Z }, 2025-12-04T08:58:35.1602742Z { 2025-12-04T08:58:35.1602960Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1603236Z "size": 104, 2025-12-04T08:58:35.1603525Z "digest": "sha256:b0b8f9b5c6ab98db9cd830dc584e1b6aec9add139e4cc48d8c243d36691e25b4" 2025-12-04T08:58:35.1603856Z }, 2025-12-04T08:58:35.1604059Z { 2025-12-04T08:58:35.1604272Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1604809Z "size": 435, 2025-12-04T08:58:35.1605087Z "digest": "sha256:0606ca4d47a8a70e91e92b03ca51a85e731641b09342136a54ef2f2a6d9dfb44" 2025-12-04T08:58:35.1605403Z }, 2025-12-04T08:58:35.1605532Z { 2025-12-04T08:58:35.1605739Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1606021Z "size": 32, 2025-12-04T08:58:35.1606295Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T08:58:35.1606615Z }, 2025-12-04T08:58:35.1606739Z { 2025-12-04T08:58:35.1606954Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1607235Z "size": 109, 2025-12-04T08:58:35.1607652Z "digest": "sha256:2f80a4e1b3b95ed67bb781ea787e8a63e46de79117d9d8e65c257072b38afa2d" 2025-12-04T08:58:35.1607979Z }, 2025-12-04T08:58:35.1608110Z { 2025-12-04T08:58:35.1608324Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1608617Z "size": 1896, 2025-12-04T08:58:35.1608899Z "digest": "sha256:35c916fb1bd057e517dcab78c3a2a018e68096d8993892ad84f47562d37ae352" 2025-12-04T08:58:35.1609213Z }, 2025-12-04T08:58:35.1609346Z { 2025-12-04T08:58:35.1609582Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1609863Z "size": 197526165, 2025-12-04T08:58:35.1610155Z "digest": "sha256:195537b7dafc96192f768323b1a8cc2a914d41959849b73198579576b0872a44" 2025-12-04T08:58:35.1610478Z }, 2025-12-04T08:58:35.1610611Z { 2025-12-04T08:58:35.1610822Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1611103Z "size": 106, 2025-12-04T08:58:35.1611375Z "digest": "sha256:dc454fd3967e5735b2498b7f1d958a2c626987d5e4ce225ca98da3cd945b59f3" 2025-12-04T08:58:35.1611686Z }, 2025-12-04T08:58:35.1611814Z { 2025-12-04T08:58:35.1612027Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1612303Z "size": 165, 2025-12-04T08:58:35.1612576Z "digest": "sha256:701b34f115fa897181c046dc37288e87cbc3ad74c36a9e2224b5bfe7c5703afb" 2025-12-04T08:58:35.1612898Z }, 2025-12-04T08:58:35.1613021Z { 2025-12-04T08:58:35.1613318Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1613604Z "size": 7944, 2025-12-04T08:58:35.1613880Z "digest": "sha256:39cefc00ffedebc9098261c798408b87a20c95a88fccb110594077f48dadf760" 2025-12-04T08:58:35.1614198Z }, 2025-12-04T08:58:35.1614327Z { 2025-12-04T08:58:35.1614544Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1614825Z "size": 8071, 2025-12-04T08:58:35.1615114Z "digest": "sha256:6ae51eb61a325b2c2995a5088c81aa20821b75be65b5aa722c7c40556b5d03ea" 2025-12-04T08:58:35.1615436Z }, 2025-12-04T08:58:35.1615563Z { 2025-12-04T08:58:35.1615786Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1616068Z "size": 304, 2025-12-04T08:58:35.1616336Z "digest": "sha256:1fd5341e66dfc0c1ae23af014641a92a6fd02640c528fe6d4dc55921ed659a26" 2025-12-04T08:58:35.1616655Z }, 2025-12-04T08:58:35.1616780Z { 2025-12-04T08:58:35.1617001Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1617280Z "size": 13364291, 2025-12-04T08:58:35.1617569Z "digest": "sha256:72a7c87e35e40ab796f90aee1b51add7902f0cdc44406d2505b6c6a1f55a8da6" 2025-12-04T08:58:35.1617888Z }, 2025-12-04T08:58:35.1618010Z { 2025-12-04T08:58:35.1618222Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1618499Z "size": 108, 2025-12-04T08:58:35.1618774Z "digest": "sha256:ec36862ac98ebaac52ee1a8b1d162d45bd0e3bf59ae7e19c8f80ad3960b4c600" 2025-12-04T08:58:35.1619107Z }, 2025-12-04T08:58:35.1619240Z { 2025-12-04T08:58:35.1619456Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1619737Z "size": 54145699, 2025-12-04T08:58:35.1620168Z "digest": "sha256:05ddbf246e8add0e293474dbf88bb028d5a295a25ac59e8648a18db644377773" 2025-12-04T08:58:35.1620491Z }, 2025-12-04T08:58:35.1620628Z { 2025-12-04T08:58:35.1620861Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:58:35.1621148Z "size": 32, 2025-12-04T08:58:35.1621439Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T08:58:35.1621760Z } 2025-12-04T08:58:35.1621892Z ] 2025-12-04T08:58:35.1622027Z } 2025-12-04T08:58:35.1622169Z + exit 0 2025-12-04T08:58:35.1645130Z ##[group]Run set -eux 2025-12-04T08:58:35.1645347Z set -eux 2025-12-04T08:58:35.1645651Z # It's ok if this steps fails, it would then be an anonymous user like what we used to have 2025-12-04T08:58:35.1646574Z aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token | jq --raw-output '.SecretString' | jq -r .docker_hub_readonly_token | docker login --username pytorchbot --password-stdin || true 2025-12-04T08:58:35.1654679Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:58:35.1654982Z env: 2025-12-04T08:58:35.1655148Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:58:35.1655344Z ##[endgroup] 2025-12-04T08:58:35.1684419Z + aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token 2025-12-04T08:58:35.1685260Z + jq --raw-output .SecretString 2025-12-04T08:58:35.1686263Z + jq -r .docker_hub_readonly_token 2025-12-04T08:58:35.1687392Z + docker login --username pytorchbot --password-stdin 2025-12-04T08:58:35.6582080Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T08:58:35.6582945Z Configure a credential helper to remove this warning. See 2025-12-04T08:58:35.6583716Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T08:58:35.6584273Z 2025-12-04T08:58:35.6585706Z Login Succeeded 2025-12-04T08:58:35.6662174Z ##[group]Run tag=${ECR_DOCKER_IMAGE##*:} 2025-12-04T08:58:35.6662515Z tag=${ECR_DOCKER_IMAGE##*:} 2025-12-04T08:58:35.6662836Z echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}" 2025-12-04T08:58:35.6670473Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:58:35.6670771Z env: 2025-12-04T08:58:35.6670945Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:58:35.6671558Z ECR_DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:35.6672185Z ##[endgroup] 2025-12-04T08:58:35.6700319Z docker pull ghcr.io/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:35.6737777Z ##[group]Run pytorch/test-infra/.github/actions/pull-docker-image@main 2025-12-04T08:58:35.6738126Z with: 2025-12-04T08:58:35.6738689Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:35.6739395Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:58:35.6739677Z env: 2025-12-04T08:58:35.6739835Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:58:35.6740032Z ##[endgroup] 2025-12-04T08:58:35.6753176Z ##[group]Run set -x 2025-12-04T08:58:35.6753379Z set -x 2025-12-04T08:58:35.6753542Z set +e 2025-12-04T08:58:35.6753709Z  2025-12-04T08:58:35.6753870Z login() { 2025-12-04T08:58:35.6754246Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-12-04T08:58:35.6754628Z } 2025-12-04T08:58:35.6754799Z  2025-12-04T08:58:35.6754986Z retry () { 2025-12-04T08:58:35.6755188Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-12-04T08:58:35.6755425Z } 2025-12-04T08:58:35.6755585Z  2025-12-04T08:58:35.6755763Z retry login "${DOCKER_REGISTRY}" 2025-12-04T08:58:35.6755984Z  2025-12-04T08:58:35.6756502Z IMAGE_SIZE=$(docker manifest inspect "${DOCKER_IMAGE}" | jq '[.layers[].size, .config.size] | add / 1024 / 1024') 2025-12-04T08:58:35.6756987Z echo "Compressed size of image in MB: ${IMAGE_SIZE}" 2025-12-04T08:58:35.6757256Z  2025-12-04T08:58:35.6757412Z set -e 2025-12-04T08:58:35.6757671Z # ignore output since only exit code is used for conditional 2025-12-04T08:58:35.6758035Z # only pull docker image if it's not available locally 2025-12-04T08:58:35.6758433Z if ! docker inspect --type=image "${DOCKER_IMAGE}" >/dev/null 2>/dev/null; then 2025-12-04T08:58:35.6758813Z  retry docker pull "${DOCKER_IMAGE}" 2025-12-04T08:58:35.6759054Z fi 2025-12-04T08:58:35.6765699Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:58:35.6765978Z env: 2025-12-04T08:58:35.6766144Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:58:35.6766727Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:35.6767408Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:58:35.6767691Z ##[endgroup] 2025-12-04T08:58:35.6791435Z + set +e 2025-12-04T08:58:35.6791788Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:58:35.6792122Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:58:35.6794874Z + aws ecr get-login-password --region us-east-1 2025-12-04T08:58:35.6796000Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:58:36.1349806Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T08:58:36.1350378Z Configure a credential helper to remove this warning. See 2025-12-04T08:58:36.1350905Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T08:58:36.1351250Z 2025-12-04T08:58:36.1351721Z Login Succeeded 2025-12-04T08:58:36.1374022Z ++ docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:36.1375377Z ++ jq '[.layers[].size, .config.size] | add / 1024 / 1024' 2025-12-04T08:58:36.3424417Z + IMAGE_SIZE=15091.581844329834 2025-12-04T08:58:36.3424964Z + echo 'Compressed size of image in MB: 15091.581844329834' 2025-12-04T08:58:36.3425508Z + set -e 2025-12-04T08:58:36.3425895Z Compressed size of image in MB: 15091.581844329834 2025-12-04T08:58:36.3427222Z + docker inspect --type=image 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:36.3542371Z + retry docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:36.3543507Z + docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:58:36.5345774Z pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a: Pulling from pytorch/ci-image 2025-12-04T08:58:36.5347182Z 63e5bc7682b8: Pulling fs layer 2025-12-04T08:58:36.5347754Z 0678d56345c9: Pulling fs layer 2025-12-04T08:58:36.5348196Z 45f5c9ddfce7: Pulling fs layer 2025-12-04T08:58:36.5348463Z 086b1df51ac1: Pulling fs layer 2025-12-04T08:58:36.5348726Z fe8a7b64bf98: Pulling fs layer 2025-12-04T08:58:36.5349101Z 7680723e9a57: Pulling fs layer 2025-12-04T08:58:36.5349397Z 9c5027aeeb4e: Pulling fs layer 2025-12-04T08:58:36.5349662Z 9a5652110360: Pulling fs layer 2025-12-04T08:58:36.5350024Z 375c4427e914: Pulling fs layer 2025-12-04T08:58:36.5350266Z a86faaa7dbdd: Pulling fs layer 2025-12-04T08:58:36.5350503Z fb7848686804: Pulling fs layer 2025-12-04T08:58:36.5350858Z 3541df015cdb: Pulling fs layer 2025-12-04T08:58:36.5351403Z 79dc80f426b2: Pulling fs layer 2025-12-04T08:58:36.5351801Z a13fcc1b90bb: Pulling fs layer 2025-12-04T08:58:36.5352097Z 4f4fb700ef54: Pulling fs layer 2025-12-04T08:58:36.5352396Z 549db4d6c618: Pulling fs layer 2025-12-04T08:58:36.5352833Z 5c63528cb580: Pulling fs layer 2025-12-04T08:58:36.5353154Z 75bd83b989a4: Pulling fs layer 2025-12-04T08:58:36.5353385Z de6e78970f51: Pulling fs layer 2025-12-04T08:58:36.5353810Z e13ed7c7e473: Pulling fs layer 2025-12-04T08:58:36.5354192Z 6e2949bcb741: Pulling fs layer 2025-12-04T08:58:36.5354633Z 14d69d9aaec7: Pulling fs layer 2025-12-04T08:58:36.5355062Z 5c02769dd8e5: Pulling fs layer 2025-12-04T08:58:36.5355512Z 35041ce524ac: Pulling fs layer 2025-12-04T08:58:36.5355873Z 2fa92dc5885e: Pulling fs layer 2025-12-04T08:58:36.5356133Z a86faaa7dbdd: Waiting 2025-12-04T08:58:36.5356405Z 2b85eafbd92a: Pulling fs layer 2025-12-04T08:58:36.5356593Z fb7848686804: Waiting 2025-12-04T08:58:36.5356781Z ff755a4ddad7: Pulling fs layer 2025-12-04T08:58:36.5357031Z 9a5652110360: Waiting 2025-12-04T08:58:36.5357200Z 3541df015cdb: Waiting 2025-12-04T08:58:36.5357364Z 09eb41bdf42d: Pulling fs layer 2025-12-04T08:58:36.5357649Z 11ede4d59e93: Pulling fs layer 2025-12-04T08:58:36.5357844Z 375c4427e914: Waiting 2025-12-04T08:58:36.5358006Z 7680723e9a57: Waiting 2025-12-04T08:58:36.5358168Z a13fcc1b90bb: Waiting 2025-12-04T08:58:36.5358332Z 4f4fb700ef54: Waiting 2025-12-04T08:58:36.5358484Z 75bd83b989a4: Waiting 2025-12-04T08:58:36.5358640Z e13ed7c7e473: Waiting 2025-12-04T08:58:36.5358807Z 1283cd8f801a: Pulling fs layer 2025-12-04T08:58:36.5358994Z 024fa855425f: Pulling fs layer 2025-12-04T08:58:36.5359174Z 79dc80f426b2: Waiting 2025-12-04T08:58:36.5359342Z 303e6747a62e: Pulling fs layer 2025-12-04T08:58:36.5359522Z 6e2949bcb741: Waiting 2025-12-04T08:58:36.5359676Z 2b85eafbd92a: Waiting 2025-12-04T08:58:36.5359841Z 3017cdf4838b: Pulling fs layer 2025-12-04T08:58:36.5360032Z 14d69d9aaec7: Waiting 2025-12-04T08:58:36.5360190Z ff755a4ddad7: Waiting 2025-12-04T08:58:36.5360359Z 6b6cd1c358e8: Pulling fs layer 2025-12-04T08:58:36.5360547Z 086b1df51ac1: Waiting 2025-12-04T08:58:36.5360712Z b2dd04501124: Pulling fs layer 2025-12-04T08:58:36.5360902Z 35041ce524ac: Waiting 2025-12-04T08:58:36.5361074Z 55adc51fe589: Pulling fs layer 2025-12-04T08:58:36.5361255Z 2fa92dc5885e: Waiting 2025-12-04T08:58:36.5361423Z a43ca0e4b837: Pulling fs layer 2025-12-04T08:58:36.5361604Z 11ede4d59e93: Waiting 2025-12-04T08:58:36.5361767Z b7212f17fd14: Pulling fs layer 2025-12-04T08:58:36.5361966Z 083e42cac090: Pulling fs layer 2025-12-04T08:58:36.5362155Z 0a00b784a4aa: Pulling fs layer 2025-12-04T08:58:36.5362329Z 303e6747a62e: Waiting 2025-12-04T08:58:36.5362486Z 3017cdf4838b: Waiting 2025-12-04T08:58:36.5362858Z 6b6cd1c358e8: Waiting 2025-12-04T08:58:36.5363081Z c6173c779f7b: Pulling fs layer 2025-12-04T08:58:36.5363449Z fe8a7b64bf98: Waiting 2025-12-04T08:58:36.5363755Z ed3d1e3387b9: Pulling fs layer 2025-12-04T08:58:36.5364073Z 024fa855425f: Waiting 2025-12-04T08:58:36.5364339Z b29343478586: Pulling fs layer 2025-12-04T08:58:36.5364663Z c6f0520487fb: Pulling fs layer 2025-12-04T08:58:36.5364986Z b2dd04501124: Waiting 2025-12-04T08:58:36.5365294Z 083e42cac090: Waiting 2025-12-04T08:58:36.5365485Z 0a00b784a4aa: Waiting 2025-12-04T08:58:36.5365642Z 55adc51fe589: Waiting 2025-12-04T08:58:36.5365830Z b7212f17fd14: Waiting 2025-12-04T08:58:36.5366104Z 148171691cd4: Pulling fs layer 2025-12-04T08:58:36.5366289Z ed3d1e3387b9: Waiting 2025-12-04T08:58:36.5366451Z 2c666d30ed77: Pulling fs layer 2025-12-04T08:58:36.5366622Z c6173c779f7b: Waiting 2025-12-04T08:58:36.5366784Z 5d8d3a0a98e0: Pulling fs layer 2025-12-04T08:58:36.5366962Z 1283cd8f801a: Waiting 2025-12-04T08:58:36.5367113Z a43ca0e4b837: Waiting 2025-12-04T08:58:36.5367284Z b06bafce9e81: Pulling fs layer 2025-12-04T08:58:36.5367470Z 2c666d30ed77: Waiting 2025-12-04T08:58:36.5367621Z 5d8d3a0a98e0: Waiting 2025-12-04T08:58:36.5367779Z c6f0520487fb: Waiting 2025-12-04T08:58:36.5367943Z 15e0d7e4590d: Pulling fs layer 2025-12-04T08:58:36.5368124Z a514bd1add31: Pulling fs layer 2025-12-04T08:58:36.5368426Z 148171691cd4: Waiting 2025-12-04T08:58:36.5368595Z b06bafce9e81: Waiting 2025-12-04T08:58:36.5368756Z 57b84ee60002: Pulling fs layer 2025-12-04T08:58:36.5368932Z b29343478586: Waiting 2025-12-04T08:58:36.5369096Z 15e0d7e4590d: Waiting 2025-12-04T08:58:36.5369253Z de6e78970f51: Waiting 2025-12-04T08:58:36.5369406Z 09eb41bdf42d: Waiting 2025-12-04T08:58:36.5369576Z b8babeff6d81: Pulling fs layer 2025-12-04T08:58:36.5369760Z 57b84ee60002: Waiting 2025-12-04T08:58:36.5369919Z 83779ddf6a85: Pulling fs layer 2025-12-04T08:58:36.5370100Z a514bd1add31: Waiting 2025-12-04T08:58:36.5370262Z b8babeff6d81: Waiting 2025-12-04T08:58:36.5370412Z 83779ddf6a85: Waiting 2025-12-04T08:58:36.5370605Z 8b7620c0d736: Pulling fs layer 2025-12-04T08:58:36.5370889Z 3bcfa090e4ef: Pulling fs layer 2025-12-04T08:58:36.5371071Z 5c02769dd8e5: Waiting 2025-12-04T08:58:36.5371232Z 8b7620c0d736: Waiting 2025-12-04T08:58:36.5371403Z eb0504ec4d92: Pulling fs layer 2025-12-04T08:58:36.5371578Z 3bcfa090e4ef: Waiting 2025-12-04T08:58:36.5371748Z 15d0fec09d7b: Pulling fs layer 2025-12-04T08:58:36.5371926Z eb0504ec4d92: Waiting 2025-12-04T08:58:36.5372084Z cca81fcc62a9: Pulling fs layer 2025-12-04T08:58:36.5372267Z 15d0fec09d7b: Waiting 2025-12-04T08:58:36.5372424Z cca81fcc62a9: Waiting 2025-12-04T08:58:36.5372583Z b0b8f9b5c6ab: Pulling fs layer 2025-12-04T08:58:36.5372760Z 5c63528cb580: Waiting 2025-12-04T08:58:36.5372921Z 0606ca4d47a8: Pulling fs layer 2025-12-04T08:58:36.5373100Z 549db4d6c618: Waiting 2025-12-04T08:58:36.5373358Z b0b8f9b5c6ab: Waiting 2025-12-04T08:58:36.5373585Z 2f80a4e1b3b9: Pulling fs layer 2025-12-04T08:58:36.5373909Z 0606ca4d47a8: Waiting 2025-12-04T08:58:36.5374209Z 35c916fb1bd0: Pulling fs layer 2025-12-04T08:58:36.5374562Z 195537b7dafc: Pulling fs layer 2025-12-04T08:58:36.5374870Z 2f80a4e1b3b9: Waiting 2025-12-04T08:58:36.5375155Z 35c916fb1bd0: Waiting 2025-12-04T08:58:36.5375465Z dc454fd3967e: Pulling fs layer 2025-12-04T08:58:36.5375722Z 195537b7dafc: Waiting 2025-12-04T08:58:36.5375906Z 701b34f115fa: Pulling fs layer 2025-12-04T08:58:36.5376087Z dc454fd3967e: Waiting 2025-12-04T08:58:36.5376253Z 39cefc00ffed: Pulling fs layer 2025-12-04T08:58:36.5388237Z 6ae51eb61a32: Pulling fs layer 2025-12-04T08:58:36.5388607Z 1fd5341e66df: Pulling fs layer 2025-12-04T08:58:36.5388965Z 72a7c87e35e4: Pulling fs layer 2025-12-04T08:58:36.5389310Z 39cefc00ffed: Waiting 2025-12-04T08:58:36.5389620Z 1fd5341e66df: Waiting 2025-12-04T08:58:36.5389898Z 72a7c87e35e4: Waiting 2025-12-04T08:58:36.5390195Z ec36862ac98e: Pulling fs layer 2025-12-04T08:58:36.5390541Z 05ddbf246e8a: Pulling fs layer 2025-12-04T08:58:36.5390858Z 6ae51eb61a32: Waiting 2025-12-04T08:58:36.5391136Z ec36862ac98e: Waiting 2025-12-04T08:58:36.5391562Z 701b34f115fa: Waiting 2025-12-04T08:58:36.6252359Z 0678d56345c9: Verifying Checksum 2025-12-04T08:58:36.6252684Z 0678d56345c9: Download complete 2025-12-04T08:58:36.6916708Z 086b1df51ac1: Verifying Checksum 2025-12-04T08:58:36.6917164Z 086b1df51ac1: Download complete 2025-12-04T08:58:36.7669577Z fe8a7b64bf98: Download complete 2025-12-04T08:58:36.8301191Z 7680723e9a57: Verifying Checksum 2025-12-04T08:58:36.8301538Z 7680723e9a57: Download complete 2025-12-04T08:58:36.8832485Z 63e5bc7682b8: Verifying Checksum 2025-12-04T08:58:36.8832830Z 63e5bc7682b8: Download complete 2025-12-04T08:58:36.8890845Z 9c5027aeeb4e: Verifying Checksum 2025-12-04T08:58:36.8891271Z 9c5027aeeb4e: Download complete 2025-12-04T08:58:36.9473675Z 9a5652110360: Download complete 2025-12-04T08:58:37.0349129Z a86faaa7dbdd: Verifying Checksum 2025-12-04T08:58:37.0349486Z a86faaa7dbdd: Download complete 2025-12-04T08:58:37.0903474Z fb7848686804: Download complete 2025-12-04T08:58:37.1780953Z 3541df015cdb: Verifying Checksum 2025-12-04T08:58:37.1781514Z 3541df015cdb: Download complete 2025-12-04T08:58:37.2443525Z 79dc80f426b2: Download complete 2025-12-04T08:58:37.7253938Z 63e5bc7682b8: Pull complete 2025-12-04T08:58:37.7365959Z 0678d56345c9: Pull complete 2025-12-04T08:58:38.0553004Z 375c4427e914: Verifying Checksum 2025-12-04T08:58:38.0553840Z 375c4427e914: Download complete 2025-12-04T08:58:38.0660205Z 4f4fb700ef54: Download complete 2025-12-04T08:58:38.1442005Z 549db4d6c618: Verifying Checksum 2025-12-04T08:58:38.1442612Z 549db4d6c618: Download complete 2025-12-04T08:58:38.2104041Z 5c63528cb580: Verifying Checksum 2025-12-04T08:58:38.2104741Z 5c63528cb580: Download complete 2025-12-04T08:58:38.2846319Z 75bd83b989a4: Verifying Checksum 2025-12-04T08:58:38.2846785Z 75bd83b989a4: Download complete 2025-12-04T08:58:38.3574189Z de6e78970f51: Verifying Checksum 2025-12-04T08:58:38.3574661Z de6e78970f51: Download complete 2025-12-04T08:58:38.4373163Z e13ed7c7e473: Verifying Checksum 2025-12-04T08:58:38.4373806Z e13ed7c7e473: Download complete 2025-12-04T08:58:38.5014295Z 6e2949bcb741: Verifying Checksum 2025-12-04T08:58:38.5014739Z 6e2949bcb741: Download complete 2025-12-04T08:58:38.5702283Z 14d69d9aaec7: Verifying Checksum 2025-12-04T08:58:38.5702776Z 14d69d9aaec7: Download complete 2025-12-04T08:58:38.6609172Z 5c02769dd8e5: Verifying Checksum 2025-12-04T08:58:38.6609664Z 5c02769dd8e5: Download complete 2025-12-04T08:58:39.7438230Z 45f5c9ddfce7: Verifying Checksum 2025-12-04T08:58:39.8236514Z 45f5c9ddfce7: Download complete 2025-12-04T08:58:39.8237053Z 2fa92dc5885e: Verifying Checksum 2025-12-04T08:58:39.8237501Z 2fa92dc5885e: Download complete 2025-12-04T08:58:40.2131269Z 2b85eafbd92a: Verifying Checksum 2025-12-04T08:58:40.2131707Z 2b85eafbd92a: Download complete 2025-12-04T08:58:40.2903967Z ff755a4ddad7: Verifying Checksum 2025-12-04T08:58:40.2904415Z ff755a4ddad7: Download complete 2025-12-04T08:58:40.3614297Z 09eb41bdf42d: Download complete 2025-12-04T08:58:45.0076767Z 11ede4d59e93: Verifying Checksum 2025-12-04T08:58:45.0077109Z 11ede4d59e93: Download complete 2025-12-04T08:58:45.1292879Z 1283cd8f801a: Verifying Checksum 2025-12-04T08:58:45.1293466Z 1283cd8f801a: Download complete 2025-12-04T08:58:45.2240820Z 024fa855425f: Download complete 2025-12-04T08:58:45.2969036Z 303e6747a62e: Verifying Checksum 2025-12-04T08:58:45.2969371Z 303e6747a62e: Download complete 2025-12-04T08:58:45.3716506Z 3017cdf4838b: Verifying Checksum 2025-12-04T08:58:45.3716825Z 3017cdf4838b: Download complete 2025-12-04T08:58:45.6070654Z 6b6cd1c358e8: Verifying Checksum 2025-12-04T08:58:45.6070957Z 6b6cd1c358e8: Download complete 2025-12-04T08:58:45.6740433Z b2dd04501124: Verifying Checksum 2025-12-04T08:58:45.6740967Z b2dd04501124: Download complete 2025-12-04T08:58:45.7535609Z 55adc51fe589: Verifying Checksum 2025-12-04T08:58:45.7535920Z 55adc51fe589: Download complete 2025-12-04T08:58:45.9198002Z a43ca0e4b837: Verifying Checksum 2025-12-04T08:58:45.9198361Z a43ca0e4b837: Download complete 2025-12-04T08:58:45.9865264Z b7212f17fd14: Verifying Checksum 2025-12-04T08:58:45.9865816Z b7212f17fd14: Download complete 2025-12-04T08:58:46.0606934Z 083e42cac090: Verifying Checksum 2025-12-04T08:58:46.0607340Z 083e42cac090: Download complete 2025-12-04T08:58:46.1385392Z 0a00b784a4aa: Verifying Checksum 2025-12-04T08:58:46.1385872Z 0a00b784a4aa: Download complete 2025-12-04T08:58:46.2134879Z c6173c779f7b: Download complete 2025-12-04T08:58:46.6997087Z 45f5c9ddfce7: Pull complete 2025-12-04T08:58:46.7215018Z 086b1df51ac1: Pull complete 2025-12-04T08:58:46.7424965Z fe8a7b64bf98: Pull complete 2025-12-04T08:58:46.7638796Z 7680723e9a57: Pull complete 2025-12-04T08:58:46.7860889Z 9c5027aeeb4e: Pull complete 2025-12-04T08:58:46.8073235Z 9a5652110360: Pull complete 2025-12-04T08:58:47.7194215Z ed3d1e3387b9: Verifying Checksum 2025-12-04T08:58:47.7194581Z ed3d1e3387b9: Download complete 2025-12-04T08:58:47.8085251Z b29343478586: Verifying Checksum 2025-12-04T08:58:47.8085724Z b29343478586: Download complete 2025-12-04T08:58:48.7496608Z 375c4427e914: Pull complete 2025-12-04T08:58:48.9087385Z a86faaa7dbdd: Pull complete 2025-12-04T08:58:49.1294150Z fb7848686804: Pull complete 2025-12-04T08:58:49.3161708Z 3541df015cdb: Pull complete 2025-12-04T08:58:49.4555368Z 79dc80f426b2: Pull complete 2025-12-04T08:58:51.0572837Z c6f0520487fb: Verifying Checksum 2025-12-04T08:58:51.0573139Z c6f0520487fb: Download complete 2025-12-04T08:59:09.1409038Z a13fcc1b90bb: Download complete 2025-12-04T08:59:09.2070316Z 2c666d30ed77: Verifying Checksum 2025-12-04T08:59:09.2070666Z 2c666d30ed77: Download complete 2025-12-04T08:59:09.2956926Z 5d8d3a0a98e0: Verifying Checksum 2025-12-04T08:59:09.2957236Z 5d8d3a0a98e0: Download complete 2025-12-04T08:59:09.3776246Z b06bafce9e81: Verifying Checksum 2025-12-04T08:59:09.3776542Z b06bafce9e81: Download complete 2025-12-04T08:59:09.4477840Z 15e0d7e4590d: Verifying Checksum 2025-12-04T08:59:09.4478135Z 15e0d7e4590d: Download complete 2025-12-04T08:59:09.5268811Z a514bd1add31: Verifying Checksum 2025-12-04T08:59:09.5269172Z a514bd1add31: Download complete 2025-12-04T08:59:09.5826388Z 57b84ee60002: Verifying Checksum 2025-12-04T08:59:09.5826833Z 57b84ee60002: Download complete 2025-12-04T08:59:09.6586154Z b8babeff6d81: Verifying Checksum 2025-12-04T08:59:09.6586565Z b8babeff6d81: Download complete 2025-12-04T08:59:09.7519363Z 83779ddf6a85: Verifying Checksum 2025-12-04T08:59:09.7519735Z 83779ddf6a85: Download complete 2025-12-04T08:59:09.8269947Z 8b7620c0d736: Verifying Checksum 2025-12-04T08:59:09.8270274Z 8b7620c0d736: Download complete 2025-12-04T08:59:09.9009221Z 3bcfa090e4ef: Download complete 2025-12-04T08:59:09.9859323Z eb0504ec4d92: Download complete 2025-12-04T08:59:10.1000212Z 15d0fec09d7b: Verifying Checksum 2025-12-04T08:59:10.1000579Z 15d0fec09d7b: Download complete 2025-12-04T08:59:10.1835076Z cca81fcc62a9: Verifying Checksum 2025-12-04T08:59:10.1835623Z cca81fcc62a9: Download complete 2025-12-04T08:59:10.2485085Z b0b8f9b5c6ab: Verifying Checksum 2025-12-04T08:59:10.2485484Z b0b8f9b5c6ab: Download complete 2025-12-04T08:59:10.3180402Z 0606ca4d47a8: Verifying Checksum 2025-12-04T08:59:10.3180832Z 0606ca4d47a8: Download complete 2025-12-04T08:59:10.3903458Z 2f80a4e1b3b9: Verifying Checksum 2025-12-04T08:59:10.3903804Z 2f80a4e1b3b9: Download complete 2025-12-04T08:59:10.4721310Z 35c916fb1bd0: Verifying Checksum 2025-12-04T08:59:10.4721780Z 35c916fb1bd0: Download complete 2025-12-04T08:59:12.4955957Z 195537b7dafc: Verifying Checksum 2025-12-04T08:59:12.4956253Z 195537b7dafc: Download complete 2025-12-04T08:59:12.5801040Z dc454fd3967e: Verifying Checksum 2025-12-04T08:59:12.5801441Z dc454fd3967e: Download complete 2025-12-04T08:59:12.6617761Z 701b34f115fa: Download complete 2025-12-04T08:59:12.7738510Z 39cefc00ffed: Download complete 2025-12-04T08:59:12.8455813Z 6ae51eb61a32: Verifying Checksum 2025-12-04T08:59:12.8456278Z 6ae51eb61a32: Download complete 2025-12-04T08:59:12.9188686Z 1fd5341e66df: Download complete 2025-12-04T08:59:13.1092155Z 72a7c87e35e4: Download complete 2025-12-04T08:59:13.1788472Z ec36862ac98e: Verifying Checksum 2025-12-04T08:59:13.1789243Z ec36862ac98e: Download complete 2025-12-04T08:59:13.7660443Z 05ddbf246e8a: Verifying Checksum 2025-12-04T08:59:13.7660802Z 05ddbf246e8a: Download complete 2025-12-04T08:59:21.6823877Z 148171691cd4: Verifying Checksum 2025-12-04T08:59:21.6824240Z 148171691cd4: Download complete 2025-12-04T08:59:59.4234176Z 35041ce524ac: Verifying Checksum 2025-12-04T08:59:59.4234580Z 35041ce524ac: Download complete 2025-12-04T09:00:30.5244726Z a13fcc1b90bb: Pull complete 2025-12-04T09:00:30.7346313Z 4f4fb700ef54: Pull complete 2025-12-04T09:00:30.8794224Z 549db4d6c618: Pull complete 2025-12-04T09:00:31.0323822Z 5c63528cb580: Pull complete 2025-12-04T09:00:31.1116128Z 75bd83b989a4: Pull complete 2025-12-04T09:00:31.3711601Z de6e78970f51: Pull complete 2025-12-04T09:00:31.5877857Z e13ed7c7e473: Pull complete 2025-12-04T09:00:31.7956829Z 6e2949bcb741: Pull complete 2025-12-04T09:00:31.9964349Z 14d69d9aaec7: Pull complete 2025-12-04T09:00:32.2104221Z 5c02769dd8e5: Pull complete 2025-12-04T09:02:05.2253787Z 35041ce524ac: Pull complete 2025-12-04T09:02:05.4439296Z 2fa92dc5885e: Pull complete 2025-12-04T09:02:06.0179757Z 2b85eafbd92a: Pull complete 2025-12-04T09:02:06.1084871Z ff755a4ddad7: Pull complete 2025-12-04T09:02:06.1331717Z 09eb41bdf42d: Pull complete 2025-12-04T09:02:12.5007824Z 11ede4d59e93: Pull complete 2025-12-04T09:02:12.7117305Z 1283cd8f801a: Pull complete 2025-12-04T09:02:12.9319799Z 024fa855425f: Pull complete 2025-12-04T09:02:13.3282990Z 303e6747a62e: Pull complete 2025-12-04T09:02:13.5661963Z 3017cdf4838b: Pull complete 2025-12-04T09:02:13.9275053Z 6b6cd1c358e8: Pull complete 2025-12-04T09:02:14.1547383Z b2dd04501124: Pull complete 2025-12-04T09:02:14.3862020Z 55adc51fe589: Pull complete 2025-12-04T09:02:14.7424257Z a43ca0e4b837: Pull complete 2025-12-04T09:02:14.9571635Z b7212f17fd14: Pull complete 2025-12-04T09:02:15.1555137Z 083e42cac090: Pull complete 2025-12-04T09:02:15.5703925Z 0a00b784a4aa: Pull complete 2025-12-04T09:02:15.7987355Z c6173c779f7b: Pull complete 2025-12-04T09:02:18.5141586Z ed3d1e3387b9: Pull complete 2025-12-04T09:02:18.7241547Z b29343478586: Pull complete 2025-12-04T09:02:19.7941611Z c6f0520487fb: Pull complete 2025-12-04T09:03:02.8294689Z 148171691cd4: Pull complete 2025-12-04T09:03:03.0275711Z 2c666d30ed77: Pull complete 2025-12-04T09:03:03.2337270Z 5d8d3a0a98e0: Pull complete 2025-12-04T09:03:03.5906208Z b06bafce9e81: Pull complete 2025-12-04T09:03:03.9893915Z 15e0d7e4590d: Pull complete 2025-12-04T09:03:04.1883089Z a514bd1add31: Pull complete 2025-12-04T09:03:04.5290807Z 57b84ee60002: Pull complete 2025-12-04T09:03:04.7693070Z b8babeff6d81: Pull complete 2025-12-04T09:03:04.8449026Z 83779ddf6a85: Pull complete 2025-12-04T09:03:05.0393751Z 8b7620c0d736: Pull complete 2025-12-04T09:03:05.4034457Z 3bcfa090e4ef: Pull complete 2025-12-04T09:03:05.6027028Z eb0504ec4d92: Pull complete 2025-12-04T09:03:05.9000819Z 15d0fec09d7b: Pull complete 2025-12-04T09:03:06.0957099Z cca81fcc62a9: Pull complete 2025-12-04T09:03:06.4198229Z b0b8f9b5c6ab: Pull complete 2025-12-04T09:03:06.6389723Z 0606ca4d47a8: Pull complete 2025-12-04T09:03:06.9921312Z 2f80a4e1b3b9: Pull complete 2025-12-04T09:03:07.1911161Z 35c916fb1bd0: Pull complete 2025-12-04T09:03:12.1679367Z 195537b7dafc: Pull complete 2025-12-04T09:03:12.3842632Z dc454fd3967e: Pull complete 2025-12-04T09:03:12.6056244Z 701b34f115fa: Pull complete 2025-12-04T09:03:12.8258911Z 39cefc00ffed: Pull complete 2025-12-04T09:03:13.0447990Z 6ae51eb61a32: Pull complete 2025-12-04T09:03:13.2615508Z 1fd5341e66df: Pull complete 2025-12-04T09:03:14.5989326Z 72a7c87e35e4: Pull complete 2025-12-04T09:03:14.8123541Z ec36862ac98e: Pull complete 2025-12-04T09:03:16.0431743Z 05ddbf246e8a: Pull complete 2025-12-04T09:03:16.2959706Z Digest: sha256:ba21003510dba4bdeed83df81a56fa468e0ee1b612a9445ae1f402a280804f97 2025-12-04T09:03:16.3286002Z Status: Downloaded newer image for 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:03:16.3453479Z 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:03:16.3535615Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:03:16.3536370Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:03:16.3545191Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:03:16.3545466Z env: 2025-12-04T09:03:16.3545773Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:03:16.3545971Z ##[endgroup] 2025-12-04T09:03:16.3693347Z ##[group]Run pytorch/test-infra/.github/actions/setup-nvidia@main 2025-12-04T09:03:16.3693679Z with: 2025-12-04T09:03:16.3693843Z driver-version: 580.82.07 2025-12-04T09:03:16.3694029Z env: 2025-12-04T09:03:16.3694185Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:03:16.3694375Z ##[endgroup] 2025-12-04T09:03:16.3723851Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:03:16.3724520Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:03:16.3731661Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:03:16.3731948Z env: 2025-12-04T09:03:16.3732108Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:03:16.3732285Z ##[endgroup] 2025-12-04T09:03:16.3785067Z ##[group]Run set -euo pipefail 2025-12-04T09:03:16.3785326Z set -euo pipefail 2025-12-04T09:03:16.3785544Z  2025-12-04T09:03:16.3785695Z has_gpu=false 2025-12-04T09:03:16.3785875Z devices="" 2025-12-04T09:03:16.3786040Z  2025-12-04T09:03:16.3786233Z if command -v nvidia-smi >/dev/null 2>&1; then 2025-12-04T09:03:16.3786574Z  if nvidia-smi -L >/tmp/nvidia_devices 2>/dev/null; then 2025-12-04T09:03:16.3786854Z  has_gpu=true 2025-12-04T09:03:16.3787069Z  devices=$(cat /tmp/nvidia_devices) 2025-12-04T09:03:16.3787288Z  fi 2025-12-04T09:03:16.3787437Z fi 2025-12-04T09:03:16.3787579Z  2025-12-04T09:03:16.3787742Z if [ "$has_gpu" = false ]; then 2025-12-04T09:03:16.3788029Z  if ls /dev/nvidia* >/tmp/nvidia_devices 2>/dev/null; then 2025-12-04T09:03:16.3788298Z  has_gpu=true 2025-12-04T09:03:16.3788500Z  devices=$(cat /tmp/nvidia_devices) 2025-12-04T09:03:16.3788717Z  fi 2025-12-04T09:03:16.3788870Z fi 2025-12-04T09:03:16.3789013Z  2025-12-04T09:03:16.3789238Z if [ "$has_gpu" = false ] && command -v lspci >/dev/null 2>&1; then 2025-12-04T09:03:16.3789599Z  if lspci | grep -i 'nvidia' >/tmp/nvidia_devices 2>/dev/null; then 2025-12-04T09:03:16.3789886Z  has_gpu=true 2025-12-04T09:03:16.3790085Z  devices=$(cat /tmp/nvidia_devices) 2025-12-04T09:03:16.3790303Z  fi 2025-12-04T09:03:16.3790462Z fi 2025-12-04T09:03:16.3790601Z  2025-12-04T09:03:16.3790818Z printf 'HAS_NVIDIA=%s\n' "$has_gpu" >> "$GITHUB_OUTPUT" 2025-12-04T09:03:16.3791193Z printf 'DETECTED_DEVICES<> "$GITHUB_OUTPUT" 2025-12-04T09:03:16.3797987Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:03:16.3798258Z env: 2025-12-04T09:03:16.3798412Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:03:16.3798592Z ##[endgroup] 2025-12-04T09:03:18.0123158Z ##[group]Run if [ "${HAS_NVIDIA}" = "true" ]; then 2025-12-04T09:03:18.0123471Z if [ "${HAS_NVIDIA}" = "true" ]; then 2025-12-04T09:03:18.0123761Z  echo "HAS_NVIDIA_GPU=true" >> "${GITHUB_ENV}" 2025-12-04T09:03:18.0124131Z  echo "GPU_FLAG=--gpus all -e NVIDIA_DRIVER_CAPABILITIES=all" >> "${GITHUB_ENV}" 2025-12-04T09:03:18.0124457Z else 2025-12-04T09:03:18.0124661Z  echo "HAS_NVIDIA_GPU=false" >> "${GITHUB_ENV}" 2025-12-04T09:03:18.0124907Z fi 2025-12-04T09:03:18.0132524Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:03:18.0132810Z env: 2025-12-04T09:03:18.0132972Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:03:18.0133158Z HAS_NVIDIA: true 2025-12-04T09:03:18.0133464Z ##[endgroup] 2025-12-04T09:03:18.0221965Z ##[group]Run nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482 2025-12-04T09:03:18.0222268Z with: 2025-12-04T09:03:18.0222418Z timeout_minutes: 10 2025-12-04T09:03:18.0222600Z max_attempts: 3 2025-12-04T09:03:18.0241575Z command: # Is it disgusting to have a full shell script here in this github action? Sure # But is it the best way to make it so that this action relies on nothing else? Absolutely set -eou pipefail DISTRIBUTION=$(. /etc/os-release;echo $ID$VERSION_ID) DRIVER_FN="NVIDIA-Linux-x86_64-${DRIVER_VERSION}.run" install_nvidia_docker2_amzn2() { ( set -x # Needed for yum-config-manager sudo yum install -y yum-utils if [[ "${DISTRIBUTION}" == "amzn2023" ]] ; then YUM_REPO_URL="https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo" else # Amazon Linux 2 YUM_REPO_URL="https://nvidia.github.io/nvidia-docker/${DISTRIBUTION}/nvidia-docker.repo" fi sudo yum-config-manager --add-repo "${YUM_REPO_URL}" sudo yum install -y \ nvidia-container-toolkit-1.17.8 \ libnvidia-container-tools-1.17.8 \ libnvidia-container1-1.17.8 \ nvidia-container-toolkit-base-1.17.8 sudo systemctl restart docker ) } install_nvidia_docker2_ubuntu20() { ( set -x # Install nvidia-driver package if not installed status="$(dpkg-query -W --showformat='${db:Status-Status}' nvidia-docker2 2>&1)" if [ ! $? = 0 ] || [ ! "$status" = installed ]; then sudo apt-get install -y nvidia-container-toolkit-1.17.8 sudo systemctl restart docker fi ) } pre_install_nvidia_driver_amzn2() { ( # Purge any nvidia driver installed from RHEL repo sudo yum remove -y nvidia-driver-latest-dkms ) } install_nvidia_driver_common() { ( # Try to gather more information about the runner and its existing NVIDIA driver if any echo "Before installing NVIDIA driver" lspci lsmod modinfo nvidia || true HAS_NVIDIA_DRIVER=0 # Check if NVIDIA driver has already been installed if [ -x "$(command -v nvidia-smi)" ]; then set +e # The driver exists, check its version next. Also check only the first GPU if there are more than one of them # so that the same driver version is not print over multiple lines INSTALLED_DRIVER_VERSION=$(nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0) NVIDIA_SMI_STATUS=$? if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then echo "Failed to get NVIDIA driver version ($INSTALLED_DRIVER_VERSION). Continuing" elif [ "$INSTALLED_DRIVER_VERSION" != "$DRIVER_VERSION" ]; then echo "NVIDIA driver ($INSTALLED_DRIVER_VERSION) has been installed, but we expect to have $DRIVER_VERSION instead. Continuing" # Turn off persistent mode so that the installation script can unload the kernel module sudo killall nvidia-persistenced || true else HAS_NVIDIA_DRIVER=1 echo "NVIDIA driver ($INSTALLED_DRIVER_VERSION) has already been installed. Skipping NVIDIA driver installation" fi set -e fi if [ "$HAS_NVIDIA_DRIVER" -eq 0 ]; then # CAUTION: this may need to be updated in future if [ "${DISTRIBUTION}" != ubuntu20.04 ]; then sudo yum groupinstall -y "Development Tools" # ensure our kernel install is the same as our underlying kernel, # groupinstall "Development Tools" has a habit of mismatching kernel headers sudo yum install -y "kernel-devel-uname-r == $(uname -r)" sudo modprobe backlight fi sudo curl -fsL -o /tmp/nvidia_driver "https://s3.amazonaws.com/ossci-linux/nvidia_driver/$DRIVER_FN" set +e sudo /bin/bash /tmp/nvidia_driver -s --no-drm NVIDIA_INSTALLATION_STATUS=$? RESET_GPU=0 if [ "$NVIDIA_INSTALLATION_STATUS" -ne 0 ]; then sudo cat /var/log/nvidia-installer.log # Fail to install NVIDIA driver, try to reset the GPU RESET_GPU=1 elif [ -x "$(command -v nvidia-smi)" ]; then # Check again if nvidia-smi works even if the driver installation completes successfully INSTALLED_DRIVER_VERSION=$(nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0) NVIDIA_SMI_STATUS=$? if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then RESET_GPU=1 fi fi if [ "$RESET_GPU" -eq 1 ]; then NVIDIA_DEVICES=$(lspci -D | grep -i NVIDIA | cut -d' ' -f1) # The GPU can get stuck in a failure state if somehow the test crashs the GPU microcode. When this # happens, we'll try to reset all NVIDIA devices https://github.com/pytorch/pytorch/issues/88388 for PCI_ID in $NVIDIA_DEVICES; do DEVICE_ENABLED=$(cat /sys/bus/pci/devices/$PCI_ID/enable) echo "Reseting $PCI_ID (enabled state: $DEVICE_ENABLED)" # This requires sudo permission of course echo "1" | sudo tee /sys/bus/pci/devices/$PCI_ID/reset sleep 1 done fi sudo rm -fv /tmp/nvidia_driver set -e fi ) } post_install_nvidia_driver_common() { ( sudo modprobe nvidia || true echo "After installing NVIDIA driver" lspci lsmod modinfo nvidia || true ( set +e nvidia-smi # NB: Annoyingly, nvidia-smi command returns successfully with return code 0 even in # the case where the driver has already crashed as it still can get the driver version # and some basic information like the bus ID. However, the rest of the information # would be missing (ERR!), for example: # # +-----------------------------------------------------------------------------+ # | NVIDIA-SMI 525.89.02 Driver Version: 525.89.02 CUDA Version: 12.0 | # |-------------------------------+----------------------+----------------------+ # | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | # | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | # | | | MIG M. | # |===============================+======================+======================| # | 0 ERR! Off | 00000000:00:1E.0 Off | ERR! | # |ERR! ERR! ERR! ERR! / ERR! | 4184MiB / 23028MiB | ERR! Default | # | | | ERR! | # +-------------------------------+----------------------+----------------------+ # # +-----------------------------------------------------------------------------+ # | Processes: | # | GPU GI CI PID Type Process name GPU Memory | # | ID ID Usage | # |=============================================================================| # +-----------------------------------------------------------------------------+ # # This should be reported as a failure instead as it will guarantee to fail when # Docker tries to run with --gpus all # # So, the correct check here is to query one of the missing piece of info like # GPU name, so that the command can fail accordingly nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 NVIDIA_SMI_STATUS=$? # Allowable exit statuses for nvidia-smi, see: https://github.com/NVIDIA/gpu-operator/issues/285 if [ "$NVIDIA_SMI_STATUS" -eq 0 ] || [ "$NVIDIA_SMI_STATUS" -eq 14 ]; then echo "INFO: Ignoring allowed status ${NVIDIA_SMI_STATUS}" else echo "ERROR: nvidia-smi exited with unresolved status ${NVIDIA_SMI_STATUS}" exit ${NVIDIA_SMI_STATUS} fi set -e ) ) } install_nvidia_driver_amzn2() { ( set -x pre_install_nvidia_driver_amzn2 install_nvidia_driver_common post_install_nvidia_driver_common ) } install_nvidia_driver_ubuntu20() { ( set -x install_nvidia_driver_common post_install_nvidia_driver_common ) } echo "== Installing nvidia driver ${DRIVER_FN} ==" case "${DISTRIBUTION}" in amzn*) install_nvidia_driver_amzn2 ;; ubuntu20.04) install_nvidia_driver_ubuntu20 ;; *) echo "ERROR: Unknown distribution ${DISTRIBUTION}" exit 1 ;; esac # Install container toolkit based on distribution echo "== Installing nvidia container toolkit for ${DISTRIBUTION} ==" case "${DISTRIBUTION}" in amzn*) install_nvidia_docker2_amzn2 ;; ubuntu20.04) install_nvidia_docker2_ubuntu20 ;; *) echo "ERROR: Unknown distribution ${DISTRIBUTION}" exit 1 ;; esac # Fix https://github.com/NVIDIA/nvidia-docker/issues/1648 on runners with # more than one GPUs. This just needs to be run once. The command fails # on subsequent runs and complains that the mode is already on, but that's # ok sudo nvidia-persistenced || true # This should show persistence mode ON nvidia-smi # check if the container-toolkit is correctly installed and CUDA is available inside a container docker run --rm -t --gpus=all public.ecr.aws/docker/library/python:3.13 nvidia-smi 2025-12-04T09:03:18.0260775Z retry_wait_seconds: 10 2025-12-04T09:03:18.0260983Z polling_interval_seconds: 1 2025-12-04T09:03:18.0261188Z warning_on_retry: true 2025-12-04T09:03:18.0261380Z continue_on_error: false 2025-12-04T09:03:18.0261560Z env: 2025-12-04T09:03:18.0261705Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:03:18.0261890Z HAS_NVIDIA_GPU: true 2025-12-04T09:03:18.0262113Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:03:18.0262370Z DRIVER_VERSION: 580.82.07 2025-12-04T09:03:18.0262554Z ##[endgroup] 2025-12-04T09:03:18.0909902Z == Installing nvidia driver NVIDIA-Linux-x86_64-580.82.07.run == 2025-12-04T09:03:18.0911190Z + pre_install_nvidia_driver_amzn2 2025-12-04T09:03:18.0913148Z + sudo yum remove -y nvidia-driver-latest-dkms 2025-12-04T09:03:18.7438900Z No match for argument: nvidia-driver-latest-dkms 2025-12-04T09:03:18.7439326Z No packages marked for removal. 2025-12-04T09:03:18.7492660Z Dependencies resolved. 2025-12-04T09:03:18.7501384Z Nothing to do. 2025-12-04T09:03:18.7502259Z Complete! 2025-12-04T09:03:18.8181097Z + install_nvidia_driver_common 2025-12-04T09:03:18.8187036Z + echo 'Before installing NVIDIA driver' 2025-12-04T09:03:18.8188617Z Before installing NVIDIA driver 2025-12-04T09:03:18.8190422Z + lspci 2025-12-04T09:03:18.9056237Z 00:00.0 Host bridge: Intel Corporation 440FX - 82441FX PMC [Natoma] 2025-12-04T09:03:18.9056746Z 00:01.0 ISA bridge: Intel Corporation 82371SB PIIX3 ISA [Natoma/Triton II] 2025-12-04T09:03:18.9057283Z 00:01.3 Non-VGA unclassified device: Intel Corporation 82371AB/EB/MB PIIX4 ACPI (rev 08) 2025-12-04T09:03:18.9057775Z 00:03.0 VGA compatible controller: Amazon.com, Inc. Device 1111 2025-12-04T09:03:18.9058262Z 00:04.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe EBS Controller 2025-12-04T09:03:18.9058698Z 01:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9059082Z 02:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9059425Z 03:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9059739Z 03:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9060044Z 03:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9060619Z 03:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9060941Z 03:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9061238Z 03:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9061528Z 03:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9061840Z 03:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9062142Z 03:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9062438Z 03:01.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9062743Z 03:01.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9063018Z 03:01.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9063398Z 03:01.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9063734Z 03:01.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9064043Z 03:01.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9064284Z 03:01.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9064525Z 03:02.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9064765Z 03:02.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9065004Z 03:02.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9065239Z 03:02.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9065477Z 03:02.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9065717Z 03:02.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9065955Z 03:02.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9066202Z 03:02.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9066440Z 03:03.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9066696Z 03:03.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9066944Z 03:03.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9067183Z 03:03.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9067421Z 03:03.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9067666Z 03:03.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9067904Z 03:03.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9068144Z 03:03.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9068385Z 24:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9068668Z 25:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9068920Z 26:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9069160Z 26:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9069395Z 26:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9069635Z 26:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9069886Z 26:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9070131Z 26:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9070366Z 26:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9070606Z 26:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9070849Z 26:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9071163Z 27:00.0 Ethernet controller: Amazon.com, Inc. Elastic Network Adapter (ENA) 2025-12-04T09:03:18.9071510Z 30:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9071764Z 31:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9072002Z 32:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9072330Z 33:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-12-04T09:03:18.9072664Z 34:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:18.9072947Z 35:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-12-04T09:03:18.9073196Z + lsmod 2025-12-04T09:03:18.9097819Z Module Size Used by 2025-12-04T09:03:18.9098111Z nvidia_uvm 1925120 0 2025-12-04T09:03:18.9098361Z nvidia 14286848 1 nvidia_uvm 2025-12-04T09:03:18.9098635Z drm 602112 1 nvidia 2025-12-04T09:03:18.9098923Z drm_panel_orientation_quirks 32768 1 drm 2025-12-04T09:03:18.9100192Z backlight 24576 1 drm 2025-12-04T09:03:18.9100491Z i2c_core 110592 2 nvidia,drm 2025-12-04T09:03:18.9100774Z xt_conntrack 16384 1 2025-12-04T09:03:18.9101013Z nft_chain_nat 16384 3 2025-12-04T09:03:18.9101256Z xt_MASQUERADE 20480 1 2025-12-04T09:03:18.9101531Z nf_nat 57344 2 nft_chain_nat,xt_MASQUERADE 2025-12-04T09:03:18.9101851Z nf_conntrack_netlink 57344 0 2025-12-04T09:03:18.9102408Z nf_conntrack 184320 4 xt_conntrack,nf_nat,nf_conntrack_netlink,xt_MASQUERADE 2025-12-04T09:03:18.9102847Z nf_defrag_ipv6 24576 1 nf_conntrack 2025-12-04T09:03:18.9103125Z nf_defrag_ipv4 16384 1 nf_conntrack 2025-12-04T09:03:18.9103441Z xfrm_user 57344 1 2025-12-04T09:03:18.9103637Z xfrm_algo 16384 1 xfrm_user 2025-12-04T09:03:18.9103846Z xt_addrtype 16384 2 2025-12-04T09:03:18.9104030Z nft_compat 20480 4 2025-12-04T09:03:18.9104245Z nf_tables 311296 57 nft_compat,nft_chain_nat 2025-12-04T09:03:18.9104823Z nfnetlink 20480 4 nft_compat,nf_conntrack_netlink,nf_tables 2025-12-04T09:03:18.9105113Z br_netfilter 36864 0 2025-12-04T09:03:18.9105308Z bridge 323584 1 br_netfilter 2025-12-04T09:03:18.9105527Z stp 16384 1 bridge 2025-12-04T09:03:18.9105744Z llc 16384 2 bridge,stp 2025-12-04T09:03:18.9105947Z overlay 167936 0 2025-12-04T09:03:18.9106129Z tls 139264 0 2025-12-04T09:03:18.9106308Z nls_ascii 16384 1 2025-12-04T09:03:18.9106485Z nls_cp437 20480 1 2025-12-04T09:03:18.9106662Z vfat 24576 1 2025-12-04T09:03:18.9106840Z fat 86016 1 vfat 2025-12-04T09:03:18.9107032Z sunrpc 700416 1 2025-12-04T09:03:18.9107205Z ena 184320 0 2025-12-04T09:03:18.9107394Z i8042 45056 0 2025-12-04T09:03:18.9107581Z serio 28672 3 i8042 2025-12-04T09:03:18.9107781Z ghash_clmulni_intel 16384 0 2025-12-04T09:03:18.9107975Z button 24576 0 2025-12-04T09:03:18.9108156Z dm_mod 188416 0 2025-12-04T09:03:18.9108331Z sch_fq_codel 20480 9 2025-12-04T09:03:18.9108510Z fuse 184320 1 2025-12-04T09:03:18.9108690Z loop 36864 0 2025-12-04T09:03:18.9108860Z configfs 57344 1 2025-12-04T09:03:18.9109039Z dmi_sysfs 20480 0 2025-12-04T09:03:18.9109219Z crc32_pclmul 16384 0 2025-12-04T09:03:18.9109398Z crc32c_intel 24576 0 2025-12-04T09:03:18.9109570Z efivarfs 24576 1 2025-12-04T09:03:18.9109753Z + modinfo nvidia 2025-12-04T09:03:18.9115029Z filename: /lib/modules/6.1.150-174.273.amzn2023.x86_64/kernel/drivers/video/nvidia.ko 2025-12-04T09:03:18.9115365Z import_ns: DMA_BUF 2025-12-04T09:03:18.9115553Z alias: char-major-195-* 2025-12-04T09:03:18.9115749Z version: 580.82.07 2025-12-04T09:03:18.9115937Z supported: external 2025-12-04T09:03:18.9116129Z license: Dual MIT/GPL 2025-12-04T09:03:18.9116355Z firmware: nvidia/580.82.07/gsp_tu10x.bin 2025-12-04T09:03:18.9116603Z firmware: nvidia/580.82.07/gsp_ga10x.bin 2025-12-04T09:03:18.9116834Z srcversion: BA7240A71DCF7DC6FE88C1D 2025-12-04T09:03:18.9117084Z alias: of:N*T*Cnvidia,tegra264-displayC* 2025-12-04T09:03:18.9117355Z alias: of:N*T*Cnvidia,tegra264-display 2025-12-04T09:03:18.9117606Z alias: of:N*T*Cnvidia,tegra234-displayC* 2025-12-04T09:03:18.9117853Z alias: of:N*T*Cnvidia,tegra234-display 2025-12-04T09:03:18.9118100Z alias: pci:v000010DEd*sv*sd*bc06sc80i00* 2025-12-04T09:03:18.9118354Z alias: pci:v000010DEd*sv*sd*bc03sc02i00* 2025-12-04T09:03:18.9118589Z alias: pci:v000010DEd*sv*sd*bc03sc00i00* 2025-12-04T09:03:18.9118819Z depends: i2c-core,drm 2025-12-04T09:03:18.9119015Z retpoline: Y 2025-12-04T09:03:18.9119344Z name: nvidia 2025-12-04T09:03:18.9119628Z vermagic: 6.1.150-174.273.amzn2023.x86_64 SMP preempt mod_unload modversions 2025-12-04T09:03:18.9119984Z parm: NvSwitchRegDwords:NvSwitch regkey (charp) 2025-12-04T09:03:18.9120310Z parm: NvSwitchBlacklist:NvSwitchBlacklist=uuid[,uuid...] (charp) 2025-12-04T09:03:18.9120619Z parm: NVreg_ResmanDebugLevel:int 2025-12-04T09:03:18.9120846Z parm: NVreg_RmLogonRC:int 2025-12-04T09:03:18.9121075Z parm: NVreg_ModifyDeviceFiles:int 2025-12-04T09:03:18.9121307Z parm: NVreg_DeviceFileUID:int 2025-12-04T09:03:18.9121524Z parm: NVreg_DeviceFileGID:int 2025-12-04T09:03:18.9121887Z parm: NVreg_DeviceFileMode:int 2025-12-04T09:03:18.9122151Z parm: NVreg_InitializeSystemMemoryAllocations:int 2025-12-04T09:03:18.9122436Z parm: NVreg_UsePageAttributeTable:int 2025-12-04T09:03:18.9122688Z parm: NVreg_EnablePCIeGen3:int 2025-12-04T09:03:18.9122904Z parm: NVreg_EnableMSI:int 2025-12-04T09:03:18.9123129Z parm: NVreg_EnableStreamMemOPs:int 2025-12-04T09:03:18.9123405Z parm: NVreg_RestrictProfilingToAdminUsers:int 2025-12-04T09:03:18.9123693Z parm: NVreg_PreserveVideoMemoryAllocations:int 2025-12-04T09:03:18.9123971Z parm: NVreg_EnableS0ixPowerManagement:int 2025-12-04T09:03:18.9124268Z parm: NVreg_S0ixPowerManagementVideoMemoryThreshold:int 2025-12-04T09:03:18.9124567Z parm: NVreg_DynamicPowerManagement:int 2025-12-04T09:03:18.9124863Z parm: NVreg_DynamicPowerManagementVideoMemoryThreshold:int 2025-12-04T09:03:18.9125164Z parm: NVreg_EnableGpuFirmware:int 2025-12-04T09:03:18.9125418Z parm: NVreg_EnableGpuFirmwareLogs:int 2025-12-04T09:03:18.9125679Z parm: NVreg_OpenRmEnableUnsupportedGpus:int 2025-12-04T09:03:18.9125953Z parm: NVreg_EnableUserNUMAManagement:int 2025-12-04T09:03:18.9126202Z parm: NVreg_MemoryPoolSize:int 2025-12-04T09:03:18.9126428Z parm: NVreg_KMallocHeapMaxSize:int 2025-12-04T09:03:18.9126668Z parm: NVreg_VMallocHeapMaxSize:int 2025-12-04T09:03:18.9126897Z parm: NVreg_IgnoreMMIOCheck:int 2025-12-04T09:03:18.9127128Z parm: NVreg_NvLinkDisable:int 2025-12-04T09:03:18.9127372Z parm: NVreg_EnablePCIERelaxedOrderingMode:int 2025-12-04T09:03:18.9127631Z parm: NVreg_RegisterPCIDriver:int 2025-12-04T09:03:18.9127884Z parm: NVreg_RegisterPlatformDeviceDriver:int 2025-12-04T09:03:18.9128134Z parm: NVreg_EnableResizableBar:int 2025-12-04T09:03:18.9128372Z parm: NVreg_EnableDbgBreakpoint:int 2025-12-04T09:03:18.9128619Z parm: NVreg_EnableNonblockingOpen:int 2025-12-04T09:03:18.9128876Z parm: NVreg_CoherentGPUMemoryMode:charp 2025-12-04T09:03:18.9129122Z parm: NVreg_RegistryDwords:charp 2025-12-04T09:03:18.9129366Z parm: NVreg_RegistryDwordsPerDevice:charp 2025-12-04T09:03:18.9129608Z parm: NVreg_RmMsg:charp 2025-12-04T09:03:18.9129813Z parm: NVreg_GpuBlacklist:charp 2025-12-04T09:03:18.9130047Z parm: NVreg_TemporaryFilePath:charp 2025-12-04T09:03:18.9130280Z parm: NVreg_ExcludedGpus:charp 2025-12-04T09:03:18.9130502Z parm: NVreg_DmaRemapPeerMmio:int 2025-12-04T09:03:18.9130736Z parm: NVreg_RmNvlinkBandwidth:charp 2025-12-04T09:03:18.9131005Z parm: NVreg_RmNvlinkBandwidthLinkCount:int 2025-12-04T09:03:18.9131250Z parm: NVreg_ImexChannelCount:int 2025-12-04T09:03:18.9131481Z parm: NVreg_CreateImexChannel0:int 2025-12-04T09:03:18.9131724Z parm: NVreg_GrdmaPciTopoCheckOverride:int 2025-12-04T09:03:18.9131964Z parm: rm_firmware_active:charp 2025-12-04T09:03:18.9132180Z + HAS_NVIDIA_DRIVER=0 2025-12-04T09:03:18.9132361Z ++ command -v nvidia-smi 2025-12-04T09:03:18.9132559Z + '[' -x /usr/bin/nvidia-smi ']' 2025-12-04T09:03:18.9132743Z + set +e 2025-12-04T09:03:18.9133055Z ++ nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0 2025-12-04T09:03:20.5275777Z + INSTALLED_DRIVER_VERSION=580.82.07 2025-12-04T09:03:20.5276388Z + NVIDIA_SMI_STATUS=0 2025-12-04T09:03:20.5277117Z + '[' 0 -ne 0 ']' 2025-12-04T09:03:20.5277487Z + '[' 580.82.07 '!=' 580.82.07 ']' 2025-12-04T09:03:20.5277953Z + HAS_NVIDIA_DRIVER=1 2025-12-04T09:03:20.5278691Z + echo 'NVIDIA driver (580.82.07) has already been installed. Skipping NVIDIA driver installation' 2025-12-04T09:03:20.5279234Z + set -e 2025-12-04T09:03:20.5279426Z + '[' 1 -eq 0 ']' 2025-12-04T09:03:20.5279828Z NVIDIA driver (580.82.07) has already been installed. Skipping NVIDIA driver installation 2025-12-04T09:03:20.5280267Z + post_install_nvidia_driver_common 2025-12-04T09:03:20.5282679Z + sudo modprobe nvidia 2025-12-04T09:03:20.6399759Z + echo 'After installing NVIDIA driver' 2025-12-04T09:03:20.6400218Z + lspci 2025-12-04T09:03:20.6400452Z After installing NVIDIA driver 2025-12-04T09:03:20.6563826Z 00:00.0 Host bridge: Intel Corporation 440FX - 82441FX PMC [Natoma] 2025-12-04T09:03:20.6564404Z 00:01.0 ISA bridge: Intel Corporation 82371SB PIIX3 ISA [Natoma/Triton II] 2025-12-04T09:03:20.6564954Z 00:01.3 Non-VGA unclassified device: Intel Corporation 82371AB/EB/MB PIIX4 ACPI (rev 08) 2025-12-04T09:03:20.6565451Z 00:03.0 VGA compatible controller: Amazon.com, Inc. Device 1111 2025-12-04T09:03:20.6565900Z 00:04.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe EBS Controller 2025-12-04T09:03:20.6566420Z 01:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6566890Z 02:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6567207Z 03:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6567503Z 03:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6567824Z 03:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6568127Z 03:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6568413Z 03:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6568710Z 03:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6569009Z 03:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6569304Z 03:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6569595Z 03:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6569899Z 03:01.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6570198Z 03:01.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6570485Z 03:01.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6570778Z 03:01.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6571090Z 03:01.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6571382Z 03:01.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6571685Z 03:01.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6571979Z 03:02.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6572275Z 03:02.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6572566Z 03:02.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6572867Z 03:02.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6573164Z 03:02.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6573572Z 03:02.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6573838Z 03:02.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6574081Z 03:02.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6574316Z 03:03.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6574559Z 03:03.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6574799Z 03:03.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6575043Z 03:03.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6575283Z 03:03.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6575526Z 03:03.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6575767Z 03:03.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6576003Z 03:03.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6576475Z 24:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6576730Z 25:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6576970Z 26:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6577235Z 26:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6577483Z 26:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6577719Z 26:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6577959Z 26:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6578197Z 26:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6578436Z 26:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6578807Z 26:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6579048Z 26:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6579366Z 27:00.0 Ethernet controller: Amazon.com, Inc. Elastic Network Adapter (ENA) 2025-12-04T09:03:20.6579686Z 30:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6579928Z 31:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6580171Z 32:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6580479Z 33:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-12-04T09:03:20.6580805Z 34:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-12-04T09:03:20.6581086Z 35:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-12-04T09:03:20.6581346Z + lsmod 2025-12-04T09:03:20.6594197Z Module Size Used by 2025-12-04T09:03:20.6594544Z nvidia_uvm 1925120 0 2025-12-04T09:03:20.6594808Z nvidia 14286848 1 nvidia_uvm 2025-12-04T09:03:20.6595074Z drm 602112 1 nvidia 2025-12-04T09:03:20.6595369Z drm_panel_orientation_quirks 32768 1 drm 2025-12-04T09:03:20.6595688Z backlight 24576 1 drm 2025-12-04T09:03:20.6596072Z i2c_core 110592 2 nvidia,drm 2025-12-04T09:03:20.6596333Z xt_conntrack 16384 1 2025-12-04T09:03:20.6596576Z nft_chain_nat 16384 3 2025-12-04T09:03:20.6596808Z xt_MASQUERADE 20480 1 2025-12-04T09:03:20.6597069Z nf_nat 57344 2 nft_chain_nat,xt_MASQUERADE 2025-12-04T09:03:20.6597378Z nf_conntrack_netlink 57344 0 2025-12-04T09:03:20.6597740Z nf_conntrack 184320 4 xt_conntrack,nf_nat,nf_conntrack_netlink,xt_MASQUERADE 2025-12-04T09:03:20.6598138Z nf_defrag_ipv6 24576 1 nf_conntrack 2025-12-04T09:03:20.6598418Z nf_defrag_ipv4 16384 1 nf_conntrack 2025-12-04T09:03:20.6598680Z xfrm_user 57344 1 2025-12-04T09:03:20.6598935Z xfrm_algo 16384 1 xfrm_user 2025-12-04T09:03:20.6599200Z xt_addrtype 16384 2 2025-12-04T09:03:20.6599434Z nft_compat 20480 4 2025-12-04T09:03:20.6599719Z nf_tables 311296 57 nft_compat,nft_chain_nat 2025-12-04T09:03:20.6600093Z nfnetlink 20480 4 nft_compat,nf_conntrack_netlink,nf_tables 2025-12-04T09:03:20.6600439Z br_netfilter 36864 0 2025-12-04T09:03:20.6600689Z bridge 323584 1 br_netfilter 2025-12-04T09:03:20.6600952Z stp 16384 1 bridge 2025-12-04T09:03:20.6601211Z llc 16384 2 bridge,stp 2025-12-04T09:03:20.6601467Z overlay 167936 0 2025-12-04T09:03:20.6601688Z tls 139264 0 2025-12-04T09:03:20.6601897Z nls_ascii 16384 1 2025-12-04T09:03:20.6602118Z nls_cp437 20480 1 2025-12-04T09:03:20.6602335Z vfat 24576 1 2025-12-04T09:03:20.6602552Z fat 86016 1 vfat 2025-12-04T09:03:20.6602796Z sunrpc 700416 1 2025-12-04T09:03:20.6603019Z ena 184320 0 2025-12-04T09:03:20.6603235Z i8042 45056 0 2025-12-04T09:03:20.6603421Z serio 28672 3 i8042 2025-12-04T09:03:20.6603624Z ghash_clmulni_intel 16384 0 2025-12-04T09:03:20.6603806Z button 24576 0 2025-12-04T09:03:20.6603980Z dm_mod 188416 0 2025-12-04T09:03:20.6604266Z sch_fq_codel 20480 9 2025-12-04T09:03:20.6604445Z fuse 184320 1 2025-12-04T09:03:20.6604915Z loop 36864 0 2025-12-04T09:03:20.6605105Z configfs 57344 1 2025-12-04T09:03:20.6605289Z dmi_sysfs 20480 0 2025-12-04T09:03:20.6605465Z crc32_pclmul 16384 0 2025-12-04T09:03:20.6605646Z crc32c_intel 24576 0 2025-12-04T09:03:20.6605823Z efivarfs 24576 1 2025-12-04T09:03:20.6605996Z + modinfo nvidia 2025-12-04T09:03:20.6611201Z filename: /lib/modules/6.1.150-174.273.amzn2023.x86_64/kernel/drivers/video/nvidia.ko 2025-12-04T09:03:20.6611710Z import_ns: DMA_BUF 2025-12-04T09:03:20.6611895Z alias: char-major-195-* 2025-12-04T09:03:20.6612093Z version: 580.82.07 2025-12-04T09:03:20.6612277Z supported: external 2025-12-04T09:03:20.6612466Z license: Dual MIT/GPL 2025-12-04T09:03:20.6612678Z firmware: nvidia/580.82.07/gsp_tu10x.bin 2025-12-04T09:03:20.6612948Z firmware: nvidia/580.82.07/gsp_ga10x.bin 2025-12-04T09:03:20.6613325Z srcversion: BA7240A71DCF7DC6FE88C1D 2025-12-04T09:03:20.6613652Z alias: of:N*T*Cnvidia,tegra264-displayC* 2025-12-04T09:03:20.6613981Z alias: of:N*T*Cnvidia,tegra264-display 2025-12-04T09:03:20.6614301Z alias: of:N*T*Cnvidia,tegra234-displayC* 2025-12-04T09:03:20.6614606Z alias: of:N*T*Cnvidia,tegra234-display 2025-12-04T09:03:20.6614928Z alias: pci:v000010DEd*sv*sd*bc06sc80i00* 2025-12-04T09:03:20.6615236Z alias: pci:v000010DEd*sv*sd*bc03sc02i00* 2025-12-04T09:03:20.6615659Z alias: pci:v000010DEd*sv*sd*bc03sc00i00* 2025-12-04T09:03:20.6616093Z depends: i2c-core,drm 2025-12-04T09:03:20.6616338Z retpoline: Y 2025-12-04T09:03:20.6616550Z name: nvidia 2025-12-04T09:03:20.6616883Z vermagic: 6.1.150-174.273.amzn2023.x86_64 SMP preempt mod_unload modversions 2025-12-04T09:03:20.6617339Z parm: NvSwitchRegDwords:NvSwitch regkey (charp) 2025-12-04T09:03:20.6617761Z parm: NvSwitchBlacklist:NvSwitchBlacklist=uuid[,uuid...] (charp) 2025-12-04T09:03:20.6618149Z parm: NVreg_ResmanDebugLevel:int 2025-12-04T09:03:20.6618430Z parm: NVreg_RmLogonRC:int 2025-12-04T09:03:20.6618707Z parm: NVreg_ModifyDeviceFiles:int 2025-12-04T09:03:20.6618996Z parm: NVreg_DeviceFileUID:int 2025-12-04T09:03:20.6619268Z parm: NVreg_DeviceFileGID:int 2025-12-04T09:03:20.6619544Z parm: NVreg_DeviceFileMode:int 2025-12-04T09:03:20.6619873Z parm: NVreg_InitializeSystemMemoryAllocations:int 2025-12-04T09:03:20.6620221Z parm: NVreg_UsePageAttributeTable:int 2025-12-04T09:03:20.6620529Z parm: NVreg_EnablePCIeGen3:int 2025-12-04T09:03:20.6620816Z parm: NVreg_EnableMSI:int 2025-12-04T09:03:20.6621091Z parm: NVreg_EnableStreamMemOPs:int 2025-12-04T09:03:20.6621424Z parm: NVreg_RestrictProfilingToAdminUsers:int 2025-12-04T09:03:20.6621790Z parm: NVreg_PreserveVideoMemoryAllocations:int 2025-12-04T09:03:20.6622135Z parm: NVreg_EnableS0ixPowerManagement:int 2025-12-04T09:03:20.6622501Z parm: NVreg_S0ixPowerManagementVideoMemoryThreshold:int 2025-12-04T09:03:20.6622866Z parm: NVreg_DynamicPowerManagement:int 2025-12-04T09:03:20.6623244Z parm: NVreg_DynamicPowerManagementVideoMemoryThreshold:int 2025-12-04T09:03:20.6623593Z parm: NVreg_EnableGpuFirmware:int 2025-12-04T09:03:20.6623838Z parm: NVreg_EnableGpuFirmwareLogs:int 2025-12-04T09:03:20.6624106Z parm: NVreg_OpenRmEnableUnsupportedGpus:int 2025-12-04T09:03:20.6624383Z parm: NVreg_EnableUserNUMAManagement:int 2025-12-04T09:03:20.6624637Z parm: NVreg_MemoryPoolSize:int 2025-12-04T09:03:20.6624870Z parm: NVreg_KMallocHeapMaxSize:int 2025-12-04T09:03:20.6625108Z parm: NVreg_VMallocHeapMaxSize:int 2025-12-04T09:03:20.6625331Z parm: NVreg_IgnoreMMIOCheck:int 2025-12-04T09:03:20.6625691Z parm: NVreg_NvLinkDisable:int 2025-12-04T09:03:20.6625952Z parm: NVreg_EnablePCIERelaxedOrderingMode:int 2025-12-04T09:03:20.6626211Z parm: NVreg_RegisterPCIDriver:int 2025-12-04T09:03:20.6626468Z parm: NVreg_RegisterPlatformDeviceDriver:int 2025-12-04T09:03:20.6626732Z parm: NVreg_EnableResizableBar:int 2025-12-04T09:03:20.6626972Z parm: NVreg_EnableDbgBreakpoint:int 2025-12-04T09:03:20.6627221Z parm: NVreg_EnableNonblockingOpen:int 2025-12-04T09:03:20.6627479Z parm: NVreg_CoherentGPUMemoryMode:charp 2025-12-04T09:03:20.6627730Z parm: NVreg_RegistryDwords:charp 2025-12-04T09:03:20.6628048Z parm: NVreg_RegistryDwordsPerDevice:charp 2025-12-04T09:03:20.6628289Z parm: NVreg_RmMsg:charp 2025-12-04T09:03:20.6628502Z parm: NVreg_GpuBlacklist:charp 2025-12-04T09:03:20.6628739Z parm: NVreg_TemporaryFilePath:charp 2025-12-04T09:03:20.6628978Z parm: NVreg_ExcludedGpus:charp 2025-12-04T09:03:20.6629210Z parm: NVreg_DmaRemapPeerMmio:int 2025-12-04T09:03:20.6629440Z parm: NVreg_RmNvlinkBandwidth:charp 2025-12-04T09:03:20.6629694Z parm: NVreg_RmNvlinkBandwidthLinkCount:int 2025-12-04T09:03:20.6629951Z parm: NVreg_ImexChannelCount:int 2025-12-04T09:03:20.6630179Z parm: NVreg_CreateImexChannel0:int 2025-12-04T09:03:20.6630429Z parm: NVreg_GrdmaPciTopoCheckOverride:int 2025-12-04T09:03:20.6630674Z parm: rm_firmware_active:charp 2025-12-04T09:03:20.6630910Z + set +e 2025-12-04T09:03:20.6631046Z + nvidia-smi 2025-12-04T09:03:22.0787213Z Thu Dec 4 09:03:22 2025 2025-12-04T09:03:22.0787652Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:03:22.0788148Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-12-04T09:03:22.0788627Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:03:22.0789114Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T09:03:22.0789613Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-12-04T09:03:22.0789992Z | | | MIG M. | 2025-12-04T09:03:22.0790302Z |=========================================+========================+======================| 2025-12-04T09:03:22.0856497Z | 0 NVIDIA L4 Off | 00000000:35:00.0 Off | 0 | 2025-12-04T09:03:22.0856953Z | N/A 41C P0 29W / 72W | 0MiB / 23034MiB | 4% Default | 2025-12-04T09:03:22.0857315Z | | | N/A | 2025-12-04T09:03:22.0857761Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:03:22.0858222Z 2025-12-04T09:03:22.0858414Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:03:22.0858826Z | Processes: | 2025-12-04T09:03:22.0859232Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T09:03:22.0859610Z | ID ID Usage | 2025-12-04T09:03:22.0859918Z |=========================================================================================| 2025-12-04T09:03:22.0861355Z | No running processes found | 2025-12-04T09:03:22.0861793Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:03:22.4069202Z + nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 2025-12-04T09:03:23.8331165Z NVIDIA L4 2025-12-04T09:03:24.0135054Z + NVIDIA_SMI_STATUS=0 2025-12-04T09:03:24.0135797Z + '[' 0 -eq 0 ']' 2025-12-04T09:03:24.0136027Z + echo 'INFO: Ignoring allowed status 0' 2025-12-04T09:03:24.0136265Z + set -e 2025-12-04T09:03:24.0136432Z INFO: Ignoring allowed status 0 2025-12-04T09:03:24.0143277Z == Installing nvidia container toolkit for amzn2023 == 2025-12-04T09:03:24.0146855Z + sudo yum install -y yum-utils 2025-12-04T09:03:24.4660932Z Last metadata expiration check: 0:07:51 ago on Thu Dec 4 08:55:33 2025. 2025-12-04T09:03:24.4876721Z Package dnf-utils-4.3.0-13.amzn2023.0.5.noarch is already installed. 2025-12-04T09:03:24.5302206Z Dependencies resolved. 2025-12-04T09:03:24.5526601Z Nothing to do. 2025-12-04T09:03:24.5527262Z Complete! 2025-12-04T09:03:24.6261393Z + [[ amzn2023 == \a\m\z\n\2\0\2\3 ]] 2025-12-04T09:03:24.6261895Z + YUM_REPO_URL=https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-12-04T09:03:24.6262624Z + sudo yum-config-manager --add-repo https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-12-04T09:03:24.9238746Z Adding repo from: https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-12-04T09:03:24.9619934Z + sudo yum install -y nvidia-container-toolkit-1.17.8 libnvidia-container-tools-1.17.8 libnvidia-container1-1.17.8 nvidia-container-toolkit-base-1.17.8 2025-12-04T09:03:25.4611558Z nvidia-container-toolkit 24 kB/s | 833 B 00:00 2025-12-04T09:03:25.5286466Z Dependencies resolved. 2025-12-04T09:03:25.5511838Z ================================================================================ 2025-12-04T09:03:25.5512295Z Package Arch Version Repository Size 2025-12-04T09:03:25.5512698Z ================================================================================ 2025-12-04T09:03:25.5513005Z Downgrading: 2025-12-04T09:03:25.5513347Z libnvidia-container-tools x86_64 1.17.8-1 nvidia-container-toolkit 40 k 2025-12-04T09:03:25.5513903Z libnvidia-container1 x86_64 1.17.8-1 nvidia-container-toolkit 1.0 M 2025-12-04T09:03:25.5514421Z nvidia-container-toolkit x86_64 1.17.8-1 nvidia-container-toolkit 1.2 M 2025-12-04T09:03:25.5514961Z nvidia-container-toolkit-base x86_64 1.17.8-1 nvidia-container-toolkit 5.8 M 2025-12-04T09:03:25.5515460Z 2025-12-04T09:03:25.5515586Z Transaction Summary 2025-12-04T09:03:25.5515900Z ================================================================================ 2025-12-04T09:03:25.5516152Z Downgrade 4 Packages 2025-12-04T09:03:25.5516272Z 2025-12-04T09:03:25.5516368Z Total download size: 8.0 M 2025-12-04T09:03:25.5517181Z Downloading Packages: 2025-12-04T09:03:25.5834014Z (1/4): libnvidia-container-tools-1.17.8-1.x86_6 1.3 MB/s | 40 kB 00:00 2025-12-04T09:03:25.5979921Z (2/4): nvidia-container-toolkit-1.17.8-1.x86_64 28 MB/s | 1.2 MB 00:00 2025-12-04T09:03:25.6045342Z (3/4): libnvidia-container1-1.17.8-1.x86_64.rpm 19 MB/s | 1.0 MB 00:00 2025-12-04T09:03:25.6335241Z (4/4): nvidia-container-toolkit-base-1.17.8-1.x 116 MB/s | 5.8 MB 00:00 2025-12-04T09:03:25.6343958Z -------------------------------------------------------------------------------- 2025-12-04T09:03:25.6346850Z Total 97 MB/s | 8.0 MB 00:00 2025-12-04T09:03:25.6349190Z Running transaction check 2025-12-04T09:03:25.6463971Z Transaction check succeeded. 2025-12-04T09:03:25.6464523Z Running transaction test 2025-12-04T09:03:25.6889093Z Transaction test succeeded. 2025-12-04T09:03:25.6891961Z Running transaction 2025-12-04T09:03:26.1745935Z Preparing : 1/1 2025-12-04T09:03:26.3430309Z Downgrading : nvidia-container-toolkit-base-1.17.8-1.x86_64 1/8 2025-12-04T09:03:26.3693011Z Downgrading : libnvidia-container1-1.17.8-1.x86_64 2/8 2025-12-04T09:03:26.4067497Z Running scriptlet: libnvidia-container1-1.17.8-1.x86_64 2/8 2025-12-04T09:03:26.5035082Z Downgrading : libnvidia-container-tools-1.17.8-1.x86_64 3/8 2025-12-04T09:03:26.5234396Z Downgrading : nvidia-container-toolkit-1.17.8-1.x86_64 4/8 2025-12-04T09:03:26.5342256Z Running scriptlet: nvidia-container-toolkit-1.17.8-1.x86_64 4/8 2025-12-04T09:03:26.5396488Z Running scriptlet: nvidia-container-toolkit-1.18.1-1.x86_64 5/8 2025-12-04T09:03:26.5399309Z Cleanup : nvidia-container-toolkit-1.18.1-1.x86_64 5/8 2025-12-04T09:03:26.5585716Z Running scriptlet: nvidia-container-toolkit-1.18.1-1.x86_64 5/8 2025-12-04T09:03:26.5636981Z Running scriptlet: libnvidia-container-tools-1.18.1-1.x86_64 6/8 2025-12-04T09:03:26.5638252Z Cleanup : libnvidia-container-tools-1.18.1-1.x86_64 6/8 2025-12-04T09:03:26.5879809Z Running scriptlet: libnvidia-container-tools-1.18.1-1.x86_64 6/8 2025-12-04T09:03:26.5936448Z Running scriptlet: libnvidia-container1-1.18.1-1.x86_64 7/8 2025-12-04T09:03:26.5937184Z Cleanup : libnvidia-container1-1.18.1-1.x86_64 7/8 2025-12-04T09:03:26.6109921Z Running scriptlet: libnvidia-container1-1.18.1-1.x86_64 7/8 2025-12-04T09:03:26.6162461Z Running scriptlet: nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8 2025-12-04T09:03:26.6163430Z Cleanup : nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8 2025-12-04T09:03:26.6388259Z Running scriptlet: nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8 2025-12-04T09:03:26.6823389Z Running scriptlet: nvidia-container-toolkit-1.17.8-1.x86_64 8/8 2025-12-04T09:04:14.6375157Z Running scriptlet: nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8 2025-12-04T09:04:14.6378024Z Verifying : libnvidia-container-tools-1.17.8-1.x86_64 1/8 2025-12-04T09:04:14.6378962Z Verifying : libnvidia-container-tools-1.18.1-1.x86_64 2/8 2025-12-04T09:04:14.6379495Z Verifying : libnvidia-container1-1.17.8-1.x86_64 3/8 2025-12-04T09:04:14.6379977Z Verifying : libnvidia-container1-1.18.1-1.x86_64 4/8 2025-12-04T09:04:14.6380447Z Verifying : nvidia-container-toolkit-1.17.8-1.x86_64 5/8 2025-12-04T09:04:14.6380922Z Verifying : nvidia-container-toolkit-1.18.1-1.x86_64 6/8 2025-12-04T09:04:14.6381392Z Verifying : nvidia-container-toolkit-base-1.17.8-1.x86_64 7/8 2025-12-04T09:04:14.7797790Z Verifying : nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8================================================================================ 2025-12-04T09:04:14.7798424Z WARNING: 2025-12-04T09:04:14.7798683Z A newer release of "Amazon Linux" is available. 2025-12-04T09:04:14.7798926Z 2025-12-04T09:04:14.7799024Z Available Versions: 2025-12-04T09:04:14.7799185Z 2025-12-04T09:04:14.7799286Z Version 2023.9.20250929: 2025-12-04T09:04:14.7799586Z Run the following command to upgrade to 2023.9.20250929: 2025-12-04T09:04:14.7799829Z 2025-12-04T09:04:14.7799962Z dnf upgrade --releasever=2023.9.20250929 2025-12-04T09:04:14.7800165Z 2025-12-04T09:04:14.7800271Z Release notes: 2025-12-04T09:04:14.7800862Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20250929.html 2025-12-04T09:04:14.7801236Z 2025-12-04T09:04:14.7801326Z Version 2023.9.20251014: 2025-12-04T09:04:14.7801616Z Run the following command to upgrade to 2023.9.20251014: 2025-12-04T09:04:14.7801853Z 2025-12-04T09:04:14.7801942Z dnf upgrade --releasever=2023.9.20251014 2025-12-04T09:04:14.7802110Z 2025-12-04T09:04:14.7802170Z Release notes: 2025-12-04T09:04:14.7802464Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251014.html 2025-12-04T09:04:14.7802740Z 2025-12-04T09:04:14.7802802Z Version 2023.9.20251020: 2025-12-04T09:04:14.7803319Z Run the following command to upgrade to 2023.9.20251020: 2025-12-04T09:04:14.7803533Z 2025-12-04T09:04:14.7803625Z dnf upgrade --releasever=2023.9.20251020 2025-12-04T09:04:14.7803783Z 2025-12-04T09:04:14.7803847Z Release notes: 2025-12-04T09:04:14.7804143Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251020.html 2025-12-04T09:04:14.7804430Z 2025-12-04T09:04:14.7804717Z Version 2023.9.20251027: 2025-12-04T09:04:14.7804969Z Run the following command to upgrade to 2023.9.20251027: 2025-12-04T09:04:14.7805157Z 2025-12-04T09:04:14.7805242Z dnf upgrade --releasever=2023.9.20251027 2025-12-04T09:04:14.7805407Z 2025-12-04T09:04:14.7805664Z Release notes: 2025-12-04T09:04:14.7805972Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251027.html 2025-12-04T09:04:14.7806249Z 2025-12-04T09:04:14.7806319Z Version 2023.9.20251105: 2025-12-04T09:04:14.7806547Z Run the following command to upgrade to 2023.9.20251105: 2025-12-04T09:04:14.7806739Z 2025-12-04T09:04:14.7806828Z dnf upgrade --releasever=2023.9.20251105 2025-12-04T09:04:14.7806988Z 2025-12-04T09:04:14.7807054Z Release notes: 2025-12-04T09:04:14.7807346Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251105.html 2025-12-04T09:04:14.7807618Z 2025-12-04T09:04:14.7807683Z Version 2023.9.20251110: 2025-12-04T09:04:14.7807912Z Run the following command to upgrade to 2023.9.20251110: 2025-12-04T09:04:14.7808095Z 2025-12-04T09:04:14.7808182Z dnf upgrade --releasever=2023.9.20251110 2025-12-04T09:04:14.7808347Z 2025-12-04T09:04:14.7808410Z Release notes: 2025-12-04T09:04:14.7808700Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251110.html 2025-12-04T09:04:14.7808976Z 2025-12-04T09:04:14.7809040Z Version 2023.9.20251117: 2025-12-04T09:04:14.7809261Z Run the following command to upgrade to 2023.9.20251117: 2025-12-04T09:04:14.7809441Z 2025-12-04T09:04:14.7809523Z dnf upgrade --releasever=2023.9.20251117 2025-12-04T09:04:14.7809681Z 2025-12-04T09:04:14.7809741Z Release notes: 2025-12-04T09:04:14.7810032Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251117.html 2025-12-04T09:04:14.7810298Z 2025-12-04T09:04:14.7810386Z ================================================================================ 2025-12-04T09:04:14.8270706Z 2025-12-04T09:04:14.8270844Z 2025-12-04T09:04:14.8270938Z Downgraded: 2025-12-04T09:04:14.8271303Z libnvidia-container-tools-1.17.8-1.x86_64 2025-12-04T09:04:14.8271847Z libnvidia-container1-1.17.8-1.x86_64 2025-12-04T09:04:14.8272381Z nvidia-container-toolkit-1.17.8-1.x86_64 2025-12-04T09:04:14.8272903Z nvidia-container-toolkit-base-1.17.8-1.x86_64 2025-12-04T09:04:14.8273223Z 2025-12-04T09:04:14.8273301Z Complete! 2025-12-04T09:04:14.8954181Z + sudo systemctl restart docker 2025-12-04T09:04:21.2155246Z Thu Dec 4 09:04:21 2025 2025-12-04T09:04:21.2155909Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:04:21.2156433Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-12-04T09:04:21.2156899Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:04:21.2157372Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T09:04:21.2157862Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-12-04T09:04:21.2158267Z | | | MIG M. | 2025-12-04T09:04:21.2158564Z |=========================================+========================+======================| 2025-12-04T09:04:21.2228885Z | 0 NVIDIA L4 On | 00000000:35:00.0 Off | 0 | 2025-12-04T09:04:21.2229768Z | N/A 41C P0 29W / 72W | 0MiB / 23034MiB | 4% Default | 2025-12-04T09:04:21.2230295Z | | | N/A | 2025-12-04T09:04:21.2230738Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:04:21.2231061Z 2025-12-04T09:04:21.2231252Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:04:21.2231795Z | Processes: | 2025-12-04T09:04:21.2232371Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T09:04:21.2233164Z | ID ID Usage | 2025-12-04T09:04:21.2233627Z |=========================================================================================| 2025-12-04T09:04:21.2234190Z | No running processes found | 2025-12-04T09:04:21.2234735Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:04:21.3819452Z Unable to find image 'public.ecr.aws/docker/library/python:3.13' locally 2025-12-04T09:04:21.5488713Z 3.13: Pulling from docker/library/python 2025-12-04T09:04:21.5813706Z 53c88f1dfeb7: Pulling fs layer 2025-12-04T09:04:21.5814167Z eae668646f44: Pulling fs layer 2025-12-04T09:04:21.5814556Z ff2e6e687b6c: Pulling fs layer 2025-12-04T09:04:21.5814894Z 7c40a3faff76: Pulling fs layer 2025-12-04T09:04:21.5815147Z 967a3b1c8fef: Pulling fs layer 2025-12-04T09:04:21.5815431Z a64e1a44f22a: Pulling fs layer 2025-12-04T09:04:21.5815675Z 52655f8a5bcc: Pulling fs layer 2025-12-04T09:04:21.5815903Z a64e1a44f22a: Waiting 2025-12-04T09:04:21.5816110Z 7c40a3faff76: Waiting 2025-12-04T09:04:21.5816320Z 967a3b1c8fef: Waiting 2025-12-04T09:04:21.6963206Z eae668646f44: Verifying Checksum 2025-12-04T09:04:21.6963728Z eae668646f44: Download complete 2025-12-04T09:04:21.7947391Z 53c88f1dfeb7: Download complete 2025-12-04T09:04:21.8019048Z ff2e6e687b6c: Verifying Checksum 2025-12-04T09:04:21.8019338Z ff2e6e687b6c: Download complete 2025-12-04T09:04:21.8403797Z 967a3b1c8fef: Verifying Checksum 2025-12-04T09:04:21.8404094Z 967a3b1c8fef: Download complete 2025-12-04T09:04:21.8867589Z 52655f8a5bcc: Verifying Checksum 2025-12-04T09:04:21.8867943Z 52655f8a5bcc: Download complete 2025-12-04T09:04:21.9068561Z a64e1a44f22a: Verifying Checksum 2025-12-04T09:04:21.9068911Z a64e1a44f22a: Download complete 2025-12-04T09:04:22.3186588Z 7c40a3faff76: Verifying Checksum 2025-12-04T09:04:22.3187037Z 7c40a3faff76: Download complete 2025-12-04T09:04:23.0626553Z 53c88f1dfeb7: Pull complete 2025-12-04T09:04:23.5968169Z eae668646f44: Pull complete 2025-12-04T09:04:25.3340916Z ff2e6e687b6c: Pull complete 2025-12-04T09:04:30.1171952Z 7c40a3faff76: Pull complete 2025-12-04T09:04:30.3965069Z 967a3b1c8fef: Pull complete 2025-12-04T09:04:31.0770741Z a64e1a44f22a: Pull complete 2025-12-04T09:04:31.2099272Z 52655f8a5bcc: Pull complete 2025-12-04T09:04:31.2689830Z Digest: sha256:3f986299a7b8b44b0d8cf9bda2b22361ce5c3058ef5d7cb17fb7452506680ab0 2025-12-04T09:04:31.3153310Z Status: Downloaded newer image for public.ecr.aws/docker/library/python:3.13 2025-12-04T09:04:38.8195428Z Thu Dec 4 09:04:38 2025 2025-12-04T09:04:38.8195862Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:04:38.8196375Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-12-04T09:04:38.8196851Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:04:38.8197383Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T09:04:38.8197903Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-12-04T09:04:38.8198614Z | | | MIG M. | 2025-12-04T09:04:38.8198938Z |=========================================+========================+======================| 2025-12-04T09:04:38.8307086Z | 0 NVIDIA L4 On | 00000000:35:00.0 Off | 0 | 2025-12-04T09:04:38.8307526Z | N/A 40C P8 12W / 72W | 0MiB / 23034MiB | 0% Default | 2025-12-04T09:04:38.8308004Z | | | N/A | 2025-12-04T09:04:38.8308388Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:04:38.8310591Z 2025-12-04T09:04:38.8310789Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:04:38.8311138Z | Processes: | 2025-12-04T09:04:38.8311597Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T09:04:38.8311922Z | ID ID Usage | 2025-12-04T09:04:38.8312271Z |=========================================================================================| 2025-12-04T09:04:38.8315721Z | No running processes found | 2025-12-04T09:04:38.8316086Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:04:40.1487807Z Command completed after 1 attempt(s). 2025-12-04T09:04:40.1575410Z Prepare all required actions 2025-12-04T09:04:40.1599329Z ##[group]Run ./.github/actions/get-workflow-job-id 2025-12-04T09:04:40.1599574Z with: 2025-12-04T09:04:40.1600083Z github-token: *** 2025-12-04T09:04:40.1600301Z env: 2025-12-04T09:04:40.1600475Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:04:40.1600711Z HAS_NVIDIA_GPU: true 2025-12-04T09:04:40.1600976Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:04:40.1601285Z ##[endgroup] 2025-12-04T09:04:40.1615377Z ##[group]Run set -eux 2025-12-04T09:04:40.1615568Z set -eux 2025-12-04T09:04:40.1615881Z python3 .github/scripts/get_workflow_job_id.py "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-12-04T09:04:40.1628187Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:04:40.1628461Z env: 2025-12-04T09:04:40.1628629Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:04:40.1628823Z HAS_NVIDIA_GPU: true 2025-12-04T09:04:40.1629088Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:04:40.1629456Z GITHUB_TOKEN: *** 2025-12-04T09:04:40.1629629Z ##[endgroup] 2025-12-04T09:04:40.1661662Z + python3 .github/scripts/get_workflow_job_id.py 19922768520 i-077765b42bd7d5290 2025-12-04T09:04:41.4506983Z Setting output job-id=57116084869 2025-12-04T09:04:41.4507689Z Setting output job-name=linux-jammy-cuda12.8-py3.10-gcc11 / test (default, 4, 5, lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests) 2025-12-04T09:04:41.4606206Z ##[group]Run python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-12-04T09:04:41.4606803Z python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-12-04T09:04:41.4607499Z python3 -m tools.stats.monitor --log-interval "$MONITOR_LOG_INTERVAL" --data-collect-interval "$MONITOR_DATA_COLLECT_INTERVAL" > usage_log.txt 2>&1 & 2025-12-04T09:04:41.4608100Z echo "monitor-script-pid=${!}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:04:41.4616543Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:04:41.4616833Z env: 2025-12-04T09:04:41.4616996Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:04:41.4617185Z HAS_NVIDIA_GPU: true 2025-12-04T09:04:41.4617401Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:04:41.4617666Z JOB_ID: 57116084869 2025-12-04T09:04:41.4618123Z JOB_NAME: linux-jammy-cuda12.8-py3.10-gcc11 / test (default, 4, 5, lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests) 2025-12-04T09:04:41.4618620Z WORKFLOW_NAME: trunk 2025-12-04T09:04:41.4618801Z WORKFLOW_RUN_ID: 19922768520 2025-12-04T09:04:41.4618995Z MONITOR_LOG_INTERVAL: 5 2025-12-04T09:04:41.4619190Z MONITOR_DATA_COLLECT_INTERVAL: 1 2025-12-04T09:04:41.4619384Z ##[endgroup] 2025-12-04T09:04:41.7212490Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T09:04:42.0568394Z Collecting psutil==5.9.8 2025-12-04T09:04:42.0740877Z Downloading psutil-5.9.8-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (288 kB) 2025-12-04T09:04:42.1445576Z Collecting dataclasses_json==0.6.7 2025-12-04T09:04:42.1484921Z Downloading dataclasses_json-0.6.7-py3-none-any.whl (28 kB) 2025-12-04T09:04:42.1744783Z Collecting nvidia-ml-py==11.525.84 2025-12-04T09:04:42.1783578Z Downloading nvidia_ml_py-11.525.84-py3-none-any.whl (34 kB) 2025-12-04T09:04:42.2858577Z Collecting marshmallow<4.0.0,>=3.18.0 2025-12-04T09:04:42.2897000Z Downloading marshmallow-3.26.1-py3-none-any.whl (50 kB) 2025-12-04T09:04:42.3119828Z Collecting typing-inspect<1,>=0.4.0 2025-12-04T09:04:42.3156965Z Downloading typing_inspect-0.9.0-py3-none-any.whl (8.8 kB) 2025-12-04T09:04:42.3654432Z Collecting packaging>=17.0 2025-12-04T09:04:42.3693450Z Downloading packaging-25.0-py3-none-any.whl (66 kB) 2025-12-04T09:04:42.3904106Z Collecting mypy-extensions>=0.3.0 2025-12-04T09:04:42.3943881Z Downloading mypy_extensions-1.1.0-py3-none-any.whl (5.0 kB) 2025-12-04T09:04:42.4391287Z Collecting typing-extensions>=3.7.4 2025-12-04T09:04:42.4432787Z Downloading typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2025-12-04T09:04:42.5291048Z Installing collected packages: typing-extensions, packaging, mypy-extensions, typing-inspect, marshmallow, psutil, nvidia-ml-py, dataclasses-json 2025-12-04T09:04:42.7770830Z Successfully installed dataclasses-json-0.6.7 marshmallow-3.26.1 mypy-extensions-1.1.0 nvidia-ml-py-11.525.84 packaging-25.0 psutil-5.9.8 typing-extensions-4.15.0 typing-inspect-0.9.0 2025-12-04T09:04:42.9342779Z Prepare all required actions 2025-12-04T09:04:42.9343189Z Getting action download info 2025-12-04T09:04:43.0980864Z Download action repository 'seemethere/download-artifact-s3@v4' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-12-04T09:04:43.3752680Z Download action repository 'actions/download-artifact@v4' (SHA:d3f86a106a0bac45b974a628896c90dbdf5c8093) 2025-12-04T09:04:43.7649824Z ##[group]Run ./.github/actions/download-build-artifacts 2025-12-04T09:04:43.7650095Z with: 2025-12-04T09:04:43.7650291Z name: linux-jammy-cuda12.8-py3.10-gcc11 2025-12-04T09:04:43.7650550Z s3-bucket: gha-artifacts 2025-12-04T09:04:43.7650740Z env: 2025-12-04T09:04:43.7650887Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:04:43.7651082Z HAS_NVIDIA_GPU: true 2025-12-04T09:04:43.7651329Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:04:43.7651585Z ##[endgroup] 2025-12-04T09:04:43.7690150Z ##[group]Run seemethere/download-artifact-s3@v4 2025-12-04T09:04:43.7690569Z with: 2025-12-04T09:04:43.7690912Z name: linux-jammy-cuda12.8-py3.10-gcc11 2025-12-04T09:04:43.7691327Z s3-bucket: gha-artifacts 2025-12-04T09:04:43.7691672Z region: us-east-1 2025-12-04T09:04:43.7691957Z env: 2025-12-04T09:04:43.7692224Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:04:43.7692555Z HAS_NVIDIA_GPU: true 2025-12-04T09:04:43.7692951Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:04:43.7693501Z ##[endgroup] 2025-12-04T09:04:44.2149345Z (node:58930) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-12-04T09:04:44.2150053Z 2025-12-04T09:04:44.2150324Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-12-04T09:04:44.2151047Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-12-04T09:04:44.2151824Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-12-04T09:04:44.4782652Z Found 1 objects with prefix pytorch/pytorch/19922768520/linux-jammy-cuda12.8-py3.10-gcc11/ 2025-12-04T09:04:44.4783615Z Starting download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-12-04T09:04:52.7294904Z Finished download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-12-04T09:04:52.7301019Z Artifact download has finished successfully 2025-12-04T09:04:52.7575291Z ##[group]Run unzip -o artifacts.zip 2025-12-04T09:04:52.7575628Z unzip -o artifacts.zip 2025-12-04T09:04:52.7583359Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:04:52.7583653Z env: 2025-12-04T09:04:52.7584001Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:04:52.7584222Z HAS_NVIDIA_GPU: true 2025-12-04T09:04:52.7584508Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:04:52.7584844Z ##[endgroup] 2025-12-04T09:04:52.7657673Z Archive: artifacts.zip 2025-12-04T09:04:52.7658809Z creating: dist/ 2025-12-04T09:04:52.7777261Z inflating: dist/.ninja_log 2025-12-04T09:04:54.7586786Z inflating: dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl 2025-12-04T09:04:54.7587559Z creating: build/ 2025-12-04T09:04:54.7587919Z creating: build/custom_test_artifacts/ 2025-12-04T09:04:54.7588433Z creating: build/custom_test_artifacts/custom-op-build/ 2025-12-04T09:04:54.7588897Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/ 2025-12-04T09:04:54.7589427Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:04:54.7595483Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:04:54.7596245Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/ 2025-12-04T09:04:54.7597075Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:04:54.7597854Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:04:54.7598796Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:04:54.7599663Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:04:54.7601284Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:04:54.7602368Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:04:54.7603125Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:04:54.7603780Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:04:54.7605634Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:04:54.7607311Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:04:54.7608251Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:04:54.7610045Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:04:54.7612023Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:04:54.7612602Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-12-04T09:04:54.7613115Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-12-04T09:04:54.7660479Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-12-04T09:04:54.7708958Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-12-04T09:04:54.7709911Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-12-04T09:04:54.7761123Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-12-04T09:04:54.7762014Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-12-04T09:04:54.7763153Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-12-04T09:04:54.7764086Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-12-04T09:04:54.7765358Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-12-04T09:04:54.7766107Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-12-04T09:04:54.7766836Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-12-04T09:04:54.7767695Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-12-04T09:04:54.7768920Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-12-04T09:04:54.7769746Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-12-04T09:04:54.7770499Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-12-04T09:04:54.7771331Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-12-04T09:04:54.7772297Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-12-04T09:04:54.7773524Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-12-04T09:04:54.7775962Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-12-04T09:04:54.7838349Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-12-04T09:04:54.7839268Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-12-04T09:04:54.7902239Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-12-04T09:04:54.7903159Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:04:54.7903979Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeTmp/ 2025-12-04T09:04:54.7904881Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/cmake.check_cache 2025-12-04T09:04:54.7905628Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/ 2025-12-04T09:04:54.7906164Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.ts 2025-12-04T09:04:54.7906855Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.make 2025-12-04T09:04:54.7907561Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/depend.make 2025-12-04T09:04:54.7908193Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/link.txt 2025-12-04T09:04:54.7908746Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/cmake_clean.cmake 2025-12-04T09:04:54.7909316Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/build.make 2025-12-04T09:04:54.7910019Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/DependInfo.cmake 2025-12-04T09:04:54.7910763Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/flags.make 2025-12-04T09:04:54.7911722Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/progress.make 2025-12-04T09:04:54.7929916Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o.d 2025-12-04T09:04:54.8091448Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o 2025-12-04T09:04:54.8092248Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/ 2025-12-04T09:04:54.8093162Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.ts 2025-12-04T09:04:54.8094567Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.make 2025-12-04T09:04:54.8095524Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/depend.make 2025-12-04T09:04:54.8096099Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/link.txt 2025-12-04T09:04:54.8096791Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/cmake_clean.cmake 2025-12-04T09:04:54.8097481Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/build.make 2025-12-04T09:04:54.8098181Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/DependInfo.cmake 2025-12-04T09:04:54.8098764Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/flags.make 2025-12-04T09:04:54.8099455Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/progress.make 2025-12-04T09:04:54.8117384Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o.d 2025-12-04T09:04:54.8183080Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o 2025-12-04T09:04:54.8184209Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:04:54.8185249Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:04:54.8185915Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/progress.marks 2025-12-04T09:04:54.8186630Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile2 2025-12-04T09:04:54.8187821Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile.cmake 2025-12-04T09:04:54.8188520Z inflating: build/custom_test_artifacts/custom-op-build/detect_cuda_version.cc 2025-12-04T09:04:54.8190999Z inflating: build/custom_test_artifacts/custom-op-build/CMakeCache.txt 2025-12-04T09:04:54.8191810Z inflating: build/custom_test_artifacts/custom-op-build/Makefile 2025-12-04T09:04:54.8192632Z inflating: build/custom_test_artifacts/custom-op-build/cmake_install.cmake 2025-12-04T09:04:54.8328922Z inflating: build/custom_test_artifacts/custom-op-build/libcustom_ops.so 2025-12-04T09:04:54.8372875Z inflating: build/custom_test_artifacts/custom-op-build/test_custom_ops 2025-12-04T09:04:54.8373395Z creating: build/custom_test_artifacts/jit-hook-build/ 2025-12-04T09:04:54.8373823Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/ 2025-12-04T09:04:54.8374334Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:04:54.8380761Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:04:54.8381341Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/ 2025-12-04T09:04:54.8381910Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:04:54.8382519Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:04:54.8383143Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:04:54.8384697Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:04:54.8386386Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:04:54.8387084Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:04:54.8387582Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:04:54.8388075Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:04:54.8390433Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:04:54.8391986Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:04:54.8393039Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:04:54.8394798Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:04:54.8396634Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:04:54.8397190Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-12-04T09:04:54.8397688Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-12-04T09:04:54.8445086Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-12-04T09:04:54.8493352Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-12-04T09:04:54.8494295Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-12-04T09:04:54.8545411Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-12-04T09:04:54.8546375Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-12-04T09:04:54.8547531Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-12-04T09:04:54.8548804Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-12-04T09:04:54.8549891Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-12-04T09:04:54.8550958Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-12-04T09:04:54.8551988Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-12-04T09:04:54.8553053Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-12-04T09:04:54.8554060Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-12-04T09:04:54.8555090Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-12-04T09:04:54.8555893Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-12-04T09:04:54.8556639Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-12-04T09:04:54.8557377Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-12-04T09:04:54.8557978Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-12-04T09:04:54.8559593Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-12-04T09:04:54.8621947Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-12-04T09:04:54.8622686Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-12-04T09:04:54.8685192Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-12-04T09:04:54.8686111Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:04:54.8687187Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeTmp/ 2025-12-04T09:04:54.8687885Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/cmake.check_cache 2025-12-04T09:04:54.8688599Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/ 2025-12-04T09:04:54.8689401Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.ts 2025-12-04T09:04:54.8690310Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.make 2025-12-04T09:04:54.8691173Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/depend.make 2025-12-04T09:04:54.8691984Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/link.txt 2025-12-04T09:04:54.8692832Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/cmake_clean.cmake 2025-12-04T09:04:54.8693787Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/build.make 2025-12-04T09:04:54.8694657Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/DependInfo.cmake 2025-12-04T09:04:54.8695360Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/flags.make 2025-12-04T09:04:54.8696259Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/progress.make 2025-12-04T09:04:54.8712613Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o.d 2025-12-04T09:04:54.8763661Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o 2025-12-04T09:04:54.8764502Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:04:54.8765187Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:04:54.8765936Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/progress.marks 2025-12-04T09:04:54.8766883Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile2 2025-12-04T09:04:54.8768503Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile.cmake 2025-12-04T09:04:54.8769240Z inflating: build/custom_test_artifacts/jit-hook-build/detect_cuda_version.cc 2025-12-04T09:04:54.8771533Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeCache.txt 2025-12-04T09:04:54.8772374Z inflating: build/custom_test_artifacts/jit-hook-build/Makefile 2025-12-04T09:04:54.8773166Z inflating: build/custom_test_artifacts/jit-hook-build/cmake_install.cmake 2025-12-04T09:04:54.8803957Z inflating: build/custom_test_artifacts/jit-hook-build/test_jit_hooks 2025-12-04T09:04:54.8804903Z creating: build/custom_test_artifacts/custom-backend-build/ 2025-12-04T09:04:54.8805513Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/ 2025-12-04T09:04:54.8806069Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:04:54.8812066Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:04:54.8812858Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/ 2025-12-04T09:04:54.8813611Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:04:54.8814554Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:04:54.8815189Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:04:54.8816552Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:04:54.8818253Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:04:54.8819103Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:04:54.8819698Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:04:54.8820247Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:04:54.8822307Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:04:54.8823897Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:04:54.8824884Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:04:54.8826650Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:04:54.8828457Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:04:54.8829068Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-12-04T09:04:54.8829609Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-12-04T09:04:54.8876783Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-12-04T09:04:54.8924978Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-12-04T09:04:54.8925997Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-12-04T09:04:54.8976690Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-12-04T09:04:54.8977642Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-12-04T09:04:54.8978631Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-12-04T09:04:54.8979612Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-12-04T09:04:54.8980570Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-12-04T09:04:54.8981490Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-12-04T09:04:54.8982419Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-12-04T09:04:54.8983337Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-12-04T09:04:54.8984503Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-12-04T09:04:54.8985711Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-12-04T09:04:54.8986544Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-12-04T09:04:54.8987247Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-12-04T09:04:54.8988015Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-12-04T09:04:54.8988790Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-12-04T09:04:54.8990697Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-12-04T09:04:54.9053158Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-12-04T09:04:54.9054033Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-12-04T09:04:54.9116712Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-12-04T09:04:54.9117659Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:04:54.9118452Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeTmp/ 2025-12-04T09:04:54.9119187Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/cmake.check_cache 2025-12-04T09:04:54.9119946Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/ 2025-12-04T09:04:54.9120803Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.ts 2025-12-04T09:04:54.9121922Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.make 2025-12-04T09:04:54.9122730Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/depend.make 2025-12-04T09:04:54.9123941Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/link.txt 2025-12-04T09:04:54.9124876Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/cmake_clean.cmake 2025-12-04T09:04:54.9125748Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/build.make 2025-12-04T09:04:54.9126354Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/DependInfo.cmake 2025-12-04T09:04:54.9126954Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/flags.make 2025-12-04T09:04:54.9127557Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/progress.make 2025-12-04T09:04:54.9129416Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o.d 2025-12-04T09:04:54.9226342Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o 2025-12-04T09:04:54.9227288Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/ 2025-12-04T09:04:54.9228197Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.ts 2025-12-04T09:04:54.9229296Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.make 2025-12-04T09:04:54.9230282Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/depend.make 2025-12-04T09:04:54.9231108Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/link.txt 2025-12-04T09:04:54.9232143Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/cmake_clean.cmake 2025-12-04T09:04:54.9233087Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/build.make 2025-12-04T09:04:54.9234031Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/DependInfo.cmake 2025-12-04T09:04:54.9234839Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/flags.make 2025-12-04T09:04:54.9235603Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/progress.make 2025-12-04T09:04:54.9251496Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o.d 2025-12-04T09:04:54.9296267Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o 2025-12-04T09:04:54.9297605Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:04:54.9298536Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:04:54.9299211Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/progress.marks 2025-12-04T09:04:54.9299982Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile2 2025-12-04T09:04:54.9300934Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile.cmake 2025-12-04T09:04:54.9301664Z inflating: build/custom_test_artifacts/custom-backend-build/detect_cuda_version.cc 2025-12-04T09:04:54.9304131Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeCache.txt 2025-12-04T09:04:54.9305134Z inflating: build/custom_test_artifacts/custom-backend-build/Makefile 2025-12-04T09:04:54.9305975Z inflating: build/custom_test_artifacts/custom-backend-build/cmake_install.cmake 2025-12-04T09:04:54.9386757Z inflating: build/custom_test_artifacts/custom-backend-build/libcustom_backend.so 2025-12-04T09:04:54.9418137Z inflating: build/custom_test_artifacts/custom-backend-build/test_custom_backend 2025-12-04T09:04:54.9418565Z creating: build/lib/ 2025-12-04T09:04:54.9485318Z inflating: build/lib/libprotobuf-lite.a 2025-12-04T09:04:54.9842147Z inflating: build/lib/libprotobuf.a 2025-12-04T09:04:55.0242183Z inflating: build/lib/libprotoc.a 2025-12-04T09:04:55.0250212Z inflating: build/lib/libpthreadpool.a 2025-12-04T09:04:55.0257199Z inflating: build/lib/libcpuinfo.a 2025-12-04T09:04:55.0263675Z inflating: build/lib/libcpuinfo_internals.a 2025-12-04T09:04:55.0264566Z inflating: build/lib/libclog.a 2025-12-04T09:04:55.0280014Z inflating: build/lib/libpytorch_qnnpack.a 2025-12-04T09:04:55.0282086Z inflating: build/lib/libnnpack_reference_layers.a 2025-12-04T09:04:55.0296886Z inflating: build/lib/libnnpack.a 2025-12-04T09:04:55.0444567Z inflating: build/lib/libmicrokernels-prod.a 2025-12-04T09:04:55.1128600Z inflating: build/lib/libmicrokernels-all.a 2025-12-04T09:04:55.1184769Z inflating: build/lib/libgtest.a 2025-12-04T09:04:55.1198716Z inflating: build/lib/libgmock.a 2025-12-04T09:04:55.1199414Z inflating: build/lib/libgtest_main.a 2025-12-04T09:04:55.1200284Z inflating: build/lib/libgmock_main.a 2025-12-04T09:04:55.1272813Z inflating: build/lib/libXNNPACK.a 2025-12-04T09:04:55.1333909Z inflating: build/lib/libbenchmark.a 2025-12-04T09:04:55.1334616Z inflating: build/lib/libbenchmark_main.a 2025-12-04T09:04:55.1335592Z inflating: build/lib/libjitprofiling.a 2025-12-04T09:04:55.1342320Z inflating: build/lib/libittnotify.a 2025-12-04T09:04:55.1395034Z inflating: build/lib/libasmjit.a 2025-12-04T09:04:55.2327981Z inflating: build/lib/libfbgemm.a 2025-12-04T09:04:55.2352645Z inflating: build/lib/libtensorpipe_uv.a 2025-12-04T09:04:55.2789192Z inflating: build/lib/libtensorpipe.a 2025-12-04T09:04:55.2994815Z inflating: build/lib/libtensorpipe_cuda.a 2025-12-04T09:04:55.3106942Z inflating: build/lib/libgloo.a 2025-12-04T09:04:55.3146210Z inflating: build/lib/libonnx_proto.a 2025-12-04T09:04:55.3505170Z inflating: build/lib/libgloo_cuda.a 2025-12-04T09:04:55.4084974Z inflating: build/lib/libonnx.a 2025-12-04T09:04:55.4100760Z inflating: build/lib/libfmt.a 2025-12-04T09:04:56.2303809Z inflating: build/lib/libdnnl.a 2025-12-04T09:04:56.2683676Z inflating: build/lib/libkineto.a 2025-12-04T09:04:56.2776684Z inflating: build/lib/libc10.so 2025-12-04T09:04:56.2816736Z inflating: build/lib/libc10_cuda.so 2025-12-04T09:04:56.2818645Z inflating: build/lib/libcaffe2_nvrtc.so 2025-12-04T09:04:56.2820185Z inflating: build/lib/libtorch_global_deps.so 2025-12-04T09:04:58.7646723Z inflating: build/lib/libtorch_cpu.so 2025-12-04T09:04:58.8278988Z inflating: build/lib/libtorch_nvshmem.so 2025-12-04T09:05:01.2231925Z inflating: build/lib/libtorch_cuda.so 2025-12-04T09:05:01.2233088Z inflating: build/lib/libtorch.so 2025-12-04T09:05:01.2274567Z inflating: build/lib/libtorch_cuda_linalg.so 2025-12-04T09:05:01.2332214Z inflating: build/lib/libtorchbind_test.so 2025-12-04T09:05:01.2348017Z inflating: build/lib/libjitbackend_test.so 2025-12-04T09:05:01.2367666Z inflating: build/lib/libbackend_with_compiler.so 2025-12-04T09:05:01.2389312Z inflating: build/lib/libaoti_custom_ops.so 2025-12-04T09:05:01.2391775Z inflating: build/lib/libc10d_cuda_test.so 2025-12-04T09:05:01.2395689Z inflating: build/lib/libshm.so 2025-12-04T09:05:01.4285224Z inflating: build/lib/libtorch_python.so 2025-12-04T09:05:01.4314229Z inflating: build/lib/libnnapi_backend.so 2025-12-04T09:05:01.4314724Z creating: build/bin/ 2025-12-04T09:05:01.4677388Z inflating: build/bin/protoc-3.13.0.0 2025-12-04T09:05:01.5039921Z inflating: build/bin/protoc 2025-12-04T09:05:01.5087272Z inflating: build/bin/c10_AllocatorConfig_test 2025-12-04T09:05:01.5131706Z inflating: build/bin/c10_CompileTimeFunctionPointer_test 2025-12-04T09:05:01.5177302Z inflating: build/bin/c10_DeviceGuard_test 2025-12-04T09:05:01.5224046Z inflating: build/bin/c10_Device_test 2025-12-04T09:05:01.5276980Z inflating: build/bin/c10_DispatchKeySet_test 2025-12-04T09:05:01.5324425Z inflating: build/bin/c10_Scalar_test 2025-12-04T09:05:01.5368149Z inflating: build/bin/c10_StreamGuard_test 2025-12-04T09:05:01.5418178Z inflating: build/bin/c10_SymInt_test 2025-12-04T09:05:01.5466295Z inflating: build/bin/c10_InlineDeviceGuard_test 2025-12-04T09:05:01.5515837Z inflating: build/bin/c10_InlineStreamGuard_test 2025-12-04T09:05:01.5559540Z inflating: build/bin/c10_ConstexprCrc_test 2025-12-04T09:05:01.5609129Z inflating: build/bin/c10_SizesAndStrides_test 2025-12-04T09:05:01.5669838Z inflating: build/bin/c10_cow_test 2025-12-04T09:05:01.5717018Z inflating: build/bin/c10_Bitset_test 2025-12-04T09:05:01.5761550Z inflating: build/bin/c10_ArrayRef_test 2025-12-04T09:05:01.5805770Z inflating: build/bin/c10_DeadlockDetection_test 2025-12-04T09:05:01.5852394Z inflating: build/bin/c10_IntrusiveList_test 2025-12-04T09:05:01.5901967Z inflating: build/bin/c10_LeftRight_test 2025-12-04T09:05:01.5947066Z inflating: build/bin/c10_Half_test 2025-12-04T09:05:01.5990901Z inflating: build/bin/c10_Semaphore_test 2025-12-04T09:05:01.6041443Z inflating: build/bin/c10_Enumerate_test 2025-12-04T09:05:01.6088629Z inflating: build/bin/c10_NetworkFlow_test 2025-12-04T09:05:01.6133564Z inflating: build/bin/c10_Synchronized_test 2025-12-04T09:05:01.6182502Z inflating: build/bin/c10_ThreadLocal_test 2025-12-04T09:05:01.6228639Z inflating: build/bin/c10_accumulate_test 2025-12-04T09:05:01.6274703Z inflating: build/bin/c10_TypeIndex_test 2025-12-04T09:05:01.6319822Z inflating: build/bin/c10_bit_cast_test 2025-12-04T09:05:01.6369047Z inflating: build/bin/c10_bfloat16_test 2025-12-04T09:05:01.6419435Z inflating: build/bin/c10_complex_math_test 2025-12-04T09:05:01.6465923Z inflating: build/bin/c10_exception_test 2025-12-04T09:05:01.6510036Z inflating: build/bin/c10_error_test 2025-12-04T09:05:01.6558750Z inflating: build/bin/c10_complex_test 2025-12-04T09:05:01.6603554Z inflating: build/bin/c10_flags_test 2025-12-04T09:05:01.6648525Z inflating: build/bin/c10_generic_math_test 2025-12-04T09:05:01.6780829Z inflating: build/bin/c10_intrusive_ptr_test 2025-12-04T09:05:01.6826365Z inflating: build/bin/c10_irange_test 2025-12-04T09:05:01.6873690Z inflating: build/bin/c10_lazy_test 2025-12-04T09:05:01.6918284Z inflating: build/bin/c10_nofatal_test 2025-12-04T09:05:01.6968477Z inflating: build/bin/c10_logging_test 2025-12-04T09:05:01.7033625Z inflating: build/bin/c10_optional_test 2025-12-04T09:05:01.7088023Z inflating: build/bin/c10_ordered_preserving_dict_test 2025-12-04T09:05:01.7217577Z inflating: build/bin/c10_small_vector_test 2025-12-04T09:05:01.7264726Z inflating: build/bin/c10_registry_test 2025-12-04T09:05:01.7314850Z inflating: build/bin/c10_string_util_test 2025-12-04T09:05:01.7360567Z inflating: build/bin/c10_ssize_test 2025-12-04T09:05:01.7404282Z inflating: build/bin/c10_string_view_test 2025-12-04T09:05:01.7443319Z inflating: build/bin/c10_intrusive_ptr_benchmark 2025-12-04T09:05:01.7487622Z inflating: build/bin/c10_tempfile_test 2025-12-04T09:05:01.7537425Z inflating: build/bin/c10_typeid_test 2025-12-04T09:05:01.7584139Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_1_var_test 2025-12-04T09:05:01.7631602Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_catches_stream 2025-12-04T09:05:01.7678400Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_blocks_and_threads 2025-12-04T09:05:01.7722468Z inflating: build/bin/c10_cuda_CUDATest 2025-12-04T09:05:01.7769312Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_catches_thread_and_block_and_device 2025-12-04T09:05:01.7816036Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_from_2_processes 2025-12-04T09:05:01.7863088Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_multiple_blocks 2025-12-04T09:05:01.7909927Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_same_block 2025-12-04T09:05:01.8391483Z inflating: build/bin/vec_test_all_types_DEFAULT 2025-12-04T09:05:01.8885118Z inflating: build/bin/vec_test_all_types_AVX512 2025-12-04T09:05:01.9387645Z inflating: build/bin/vec_test_all_types_AVX2 2025-12-04T09:05:01.9431756Z inflating: build/bin/test_vec_half_DEFAULT 2025-12-04T09:05:01.9515504Z inflating: build/bin/test_aoti_abi_check 2025-12-04T09:05:01.9559974Z inflating: build/bin/test_vec_half_AVX512 2025-12-04T09:05:01.9604311Z inflating: build/bin/test_vec_half_AVX2 2025-12-04T09:05:01.9668068Z inflating: build/bin/Dict_test 2025-12-04T09:05:01.9714598Z inflating: build/bin/Dimname_test 2025-12-04T09:05:01.9771241Z inflating: build/bin/MaybeOwned_test 2025-12-04T09:05:01.9821899Z inflating: build/bin/NamedTensor_test 2025-12-04T09:05:01.9873639Z inflating: build/bin/apply_utils_test 2025-12-04T09:05:01.9925772Z inflating: build/bin/atest 2025-12-04T09:05:01.9981600Z inflating: build/bin/basic 2025-12-04T09:05:02.0029858Z inflating: build/bin/broadcast_test 2025-12-04T09:05:02.0074816Z inflating: build/bin/cpu_allocator_test 2025-12-04T09:05:02.0125992Z inflating: build/bin/cpu_generator_test 2025-12-04T09:05:02.0172593Z inflating: build/bin/cpu_profiling_allocator_test 2025-12-04T09:05:02.0252877Z inflating: build/bin/cpu_rng_test 2025-12-04T09:05:02.0298550Z inflating: build/bin/dlconvertor_test 2025-12-04T09:05:02.0349360Z inflating: build/bin/extension_backend_test 2025-12-04T09:05:02.0398025Z inflating: build/bin/half_test 2025-12-04T09:05:02.0481227Z inflating: build/bin/ivalue_test 2025-12-04T09:05:02.0525743Z inflating: build/bin/lazy_tensor_test 2025-12-04T09:05:02.0572474Z inflating: build/bin/math_kernel_test 2025-12-04T09:05:02.0619234Z inflating: build/bin/memory_format_test 2025-12-04T09:05:02.0666583Z inflating: build/bin/memory_overlapping_test 2025-12-04T09:05:02.0713704Z inflating: build/bin/mobile_memory_cleanup 2025-12-04T09:05:02.0763620Z inflating: build/bin/native_test 2025-12-04T09:05:02.0809190Z inflating: build/bin/operator_name_test 2025-12-04T09:05:02.0854312Z inflating: build/bin/operators_test 2025-12-04T09:05:02.0900513Z inflating: build/bin/packedtensoraccessor_test 2025-12-04T09:05:02.0959451Z inflating: build/bin/pow_test 2025-12-04T09:05:02.1009403Z inflating: build/bin/quantized_test 2025-12-04T09:05:02.1053920Z inflating: build/bin/reduce_ops_test 2025-12-04T09:05:02.1099301Z inflating: build/bin/reportMemoryUsage_test 2025-12-04T09:05:02.1148494Z inflating: build/bin/scalar_tensor_test 2025-12-04T09:05:02.1199018Z inflating: build/bin/scalar_test 2025-12-04T09:05:02.1245045Z inflating: build/bin/StorageUtils_test 2025-12-04T09:05:02.1290899Z inflating: build/bin/stride_properties_test 2025-12-04T09:05:02.1359788Z inflating: build/bin/tensor_iterator_test 2025-12-04T09:05:02.1407802Z inflating: build/bin/test_parallel 2025-12-04T09:05:02.1452589Z inflating: build/bin/thread_init_test 2025-12-04T09:05:02.1501238Z inflating: build/bin/type_ptr_test 2025-12-04T09:05:02.1553536Z inflating: build/bin/type_test 2025-12-04T09:05:02.1599857Z inflating: build/bin/undefined_tensor_test 2025-12-04T09:05:02.1643997Z inflating: build/bin/verify_api_visibility 2025-12-04T09:05:02.1705609Z inflating: build/bin/legacy_vmap_test 2025-12-04T09:05:02.1751126Z inflating: build/bin/weakref_test 2025-12-04T09:05:02.1796816Z inflating: build/bin/wrapdim_test 2025-12-04T09:05:02.1842229Z inflating: build/bin/xla_tensor_test 2025-12-04T09:05:02.1894437Z inflating: build/bin/IListRef_test 2025-12-04T09:05:02.1984683Z inflating: build/bin/List_test 2025-12-04T09:05:02.2043079Z inflating: build/bin/KernelFunction_test 2025-12-04T09:05:02.2145114Z inflating: build/bin/kernel_function_legacy_test 2025-12-04T09:05:02.2227165Z inflating: build/bin/kernel_function_test 2025-12-04T09:05:02.2334751Z inflating: build/bin/kernel_lambda_legacy_test 2025-12-04T09:05:02.2421886Z inflating: build/bin/kernel_lambda_test 2025-12-04T09:05:02.2474395Z inflating: build/bin/kernel_stackbased_test 2025-12-04T09:05:02.2556251Z inflating: build/bin/make_boxed_from_unboxed_functor_test 2025-12-04T09:05:02.2601804Z inflating: build/bin/CppSignature_test 2025-12-04T09:05:02.2650638Z inflating: build/bin/backend_fallback_test 2025-12-04T09:05:02.2694301Z inflating: build/bin/op_allowlist_test 2025-12-04T09:05:02.2955325Z inflating: build/bin/op_registration_test 2025-12-04T09:05:02.3013839Z inflating: build/bin/inline_container_test 2025-12-04T09:05:02.3060951Z inflating: build/bin/cuda_allocator_test 2025-12-04T09:05:02.3107923Z inflating: build/bin/cuda_apply_test 2025-12-04T09:05:02.3161067Z inflating: build/bin/cuda_atomic_ops_test 2025-12-04T09:05:02.3210899Z inflating: build/bin/cuda_caching_host_allocator_test 2025-12-04T09:05:02.3272143Z inflating: build/bin/cuda_complex_math_test 2025-12-04T09:05:02.3324350Z inflating: build/bin/cuda_complex_test 2025-12-04T09:05:02.3380958Z inflating: build/bin/cuda_cub_test 2025-12-04T09:05:02.3427904Z inflating: build/bin/cuda_cublas_handle_pool_test 2025-12-04T09:05:02.3472186Z inflating: build/bin/cuda_device_test 2025-12-04T09:05:02.3538437Z inflating: build/bin/cuda_distributions_test 2025-12-04T09:05:02.3584349Z inflating: build/bin/cuda_dlconvertor_test 2025-12-04T09:05:02.3631623Z inflating: build/bin/cuda_event_test 2025-12-04T09:05:02.3675135Z inflating: build/bin/cuda_exchange_device_test 2025-12-04T09:05:02.3725522Z inflating: build/bin/cuda_generator_test 2025-12-04T09:05:02.3770146Z inflating: build/bin/cuda_half_test 2025-12-04T09:05:02.3814661Z inflating: build/bin/cuda_allocatorTraceTracker_test 2025-12-04T09:05:02.3868223Z inflating: build/bin/cuda_stream_test 2025-12-04T09:05:02.3915097Z inflating: build/bin/cuda_reportMemoryUsage_test 2025-12-04T09:05:02.3959072Z inflating: build/bin/cuda_cudnn_test 2025-12-04T09:05:02.4005258Z inflating: build/bin/cuda_integer_divider_test 2025-12-04T09:05:02.4049582Z inflating: build/bin/cuda_optional_test 2025-12-04T09:05:02.4095996Z inflating: build/bin/cuda_packedtensoraccessor_test 2025-12-04T09:05:02.4142568Z inflating: build/bin/cuda_vectorized_test 2025-12-04T09:05:02.5043103Z inflating: build/bin/test_jit 2025-12-04T09:05:02.5335365Z inflating: build/bin/test_lazy 2025-12-04T09:05:02.5381746Z inflating: build/bin/BackoffTest 2025-12-04T09:05:02.5429057Z inflating: build/bin/FileStoreTest 2025-12-04T09:05:02.5479189Z inflating: build/bin/TCPStoreTest 2025-12-04T09:05:02.5526898Z inflating: build/bin/HashStoreTest 2025-12-04T09:05:02.5538936Z inflating: build/bin/ProcessGroupMPITest 2025-12-04T09:05:02.5541636Z inflating: build/bin/example_allreduce 2025-12-04T09:05:02.5590330Z inflating: build/bin/test_dist_autograd 2025-12-04T09:05:02.5650124Z inflating: build/bin/test_cpp_rpc 2025-12-04T09:05:02.5708878Z inflating: build/bin/ProcessGroupGlooTest 2025-12-04T09:05:02.5758969Z inflating: build/bin/ProcessGroupGlooAsyncTest 2025-12-04T09:05:02.5815337Z inflating: build/bin/ProcessGroupNCCLTest 2025-12-04T09:05:02.5869393Z inflating: build/bin/ProcessGroupNCCLErrorsTest 2025-12-04T09:05:02.6826784Z inflating: build/bin/test_api 2025-12-04T09:05:02.6829324Z inflating: build/bin/parallel_benchmark 2025-12-04T09:05:02.6832689Z inflating: build/bin/torch_shm_manager 2025-12-04T09:05:02.6833175Z creating: .additional_ci_files/ 2025-12-04T09:05:02.6885517Z inflating: .additional_ci_files/test-times.json 2025-12-04T09:05:02.7076055Z inflating: .additional_ci_files/test-class-times.json 2025-12-04T09:05:02.7112438Z ##[group]Run rm artifacts.zip 2025-12-04T09:05:02.7112669Z rm artifacts.zip 2025-12-04T09:05:02.7120356Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:05:02.7120636Z env: 2025-12-04T09:05:02.7120939Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:05:02.7121136Z HAS_NVIDIA_GPU: true 2025-12-04T09:05:02.7121365Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:05:02.7121614Z ##[endgroup] 2025-12-04T09:05:02.8190783Z ##[group]Run df -H 2025-12-04T09:05:02.8190973Z df -H 2025-12-04T09:05:02.8197897Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:05:02.8198181Z env: 2025-12-04T09:05:02.8198337Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:05:02.8198542Z HAS_NVIDIA_GPU: true 2025-12-04T09:05:02.8198787Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:05:02.8199059Z ##[endgroup] 2025-12-04T09:05:02.8247118Z Filesystem Size Used Avail Use% Mounted on 2025-12-04T09:05:02.8247685Z devtmpfs 4.2M 0 4.2M 0% /dev 2025-12-04T09:05:02.8248187Z tmpfs 33G 0 33G 0% /dev/shm 2025-12-04T09:05:02.8248490Z tmpfs 13G 775k 13G 1% /run 2025-12-04T09:05:02.8248786Z /dev/nvme0n1p1 161G 54G 108G 34% / 2025-12-04T09:05:02.8249075Z tmpfs 33G 17k 33G 1% /tmp 2025-12-04T09:05:02.8249382Z /dev/nvme0n1p128 11M 1.4M 9.2M 13% /boot/efi 2025-12-04T09:05:02.8249698Z tmpfs 6.5G 0 6.5G 0% /run/user/0 2025-12-04T09:05:02.8277959Z Prepare all required actions 2025-12-04T09:05:02.8278749Z Getting action download info 2025-12-04T09:05:02.9963365Z ##[group]Run ./.github/actions/download-td-artifacts 2025-12-04T09:05:02.9963628Z with: 2025-12-04T09:05:02.9963776Z env: 2025-12-04T09:05:02.9963925Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:05:02.9964122Z HAS_NVIDIA_GPU: true 2025-12-04T09:05:02.9964374Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:05:02.9964624Z ##[endgroup] 2025-12-04T09:05:03.0048458Z ##[group]Run seemethere/download-artifact-s3@v4 2025-12-04T09:05:03.0048726Z with: 2025-12-04T09:05:03.0048873Z name: td_results 2025-12-04T09:05:03.0049040Z s3-bucket: gha-artifacts 2025-12-04T09:05:03.0049229Z region: us-east-1 2025-12-04T09:05:03.0049384Z env: 2025-12-04T09:05:03.0049529Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:05:03.0049716Z HAS_NVIDIA_GPU: true 2025-12-04T09:05:03.0049949Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:05:03.0050242Z ##[endgroup] 2025-12-04T09:05:03.4018279Z (node:58951) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-12-04T09:05:03.4018753Z 2025-12-04T09:05:03.4018947Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-12-04T09:05:03.4019453Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-12-04T09:05:03.4020307Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-12-04T09:05:03.4928444Z Found 1 objects with prefix pytorch/pytorch/19922768520/td_results/ 2025-12-04T09:05:03.4929050Z Starting download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/td_results.json 2025-12-04T09:05:03.5442093Z Finished download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/td_results.json 2025-12-04T09:05:03.5446930Z Artifact download has finished successfully 2025-12-04T09:05:03.5714275Z ##[group]Run mkdir -p .additional_ci_files 2025-12-04T09:05:03.5714634Z mkdir -p .additional_ci_files 2025-12-04T09:05:03.5714949Z mv td_results.json .additional_ci_files/td_results.json || true 2025-12-04T09:05:03.5722588Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:05:03.5722860Z env: 2025-12-04T09:05:03.5723019Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:05:03.5723205Z HAS_NVIDIA_GPU: true 2025-12-04T09:05:03.5723449Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:05:03.5723695Z ##[endgroup] 2025-12-04T09:05:03.5838591Z ##[group]Run .github/scripts/parse_ref.py 2025-12-04T09:05:03.5839063Z .github/scripts/parse_ref.py 2025-12-04T09:05:03.5845825Z shell: /usr/bin/bash -e {0} 2025-12-04T09:05:03.5846029Z env: 2025-12-04T09:05:03.5846188Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:05:03.5846377Z HAS_NVIDIA_GPU: true 2025-12-04T09:05:03.5846598Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:05:03.5846863Z ##[endgroup] 2025-12-04T09:05:03.6044581Z Setting output branch=main 2025-12-04T09:05:03.6146556Z Prepare all required actions 2025-12-04T09:05:03.6146878Z Getting action download info 2025-12-04T09:05:03.7766336Z ##[group]Run ./.github/actions/filter-test-configs 2025-12-04T09:05:03.7766606Z with: 2025-12-04T09:05:03.7766938Z github-token: *** 2025-12-04T09:05:03.7773863Z test-matrix: {"include": [{"config": "default", "shard": 1, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 1, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 2, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 2, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 3, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 3, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 4, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 4, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 5, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 5, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "pr_time_benchmarks", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "pr_time_benchmarks", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "libtorch_agnostic_targetting", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "libtorch_agnostic_targetting", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}]} 2025-12-04T09:05:03.7781289Z job-name: linux-jammy-cuda12.8-py3.10-gcc11 / test (default, 4, 5, lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests) 2025-12-04T09:05:03.7781772Z env: 2025-12-04T09:05:03.7781938Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:05:03.7782143Z HAS_NVIDIA_GPU: true 2025-12-04T09:05:03.7782381Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:05:03.7782644Z ##[endgroup] 2025-12-04T09:05:03.7891921Z ##[group]Run nick-fields/retry@v3.0.0 2025-12-04T09:05:03.7892145Z with: 2025-12-04T09:05:03.7892295Z shell: bash 2025-12-04T09:05:03.7892454Z timeout_minutes: 10 2025-12-04T09:05:03.7892629Z max_attempts: 5 2025-12-04T09:05:03.7892797Z retry_wait_seconds: 30 2025-12-04T09:05:03.7893647Z command: set -eux # PyYAML 6.0 doesn't work with MacOS x86 anymore # This must run on Python-3.7 (AmazonLinux2) so can't use request=3.32.2 python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-12-04T09:05:03.7894258Z polling_interval_seconds: 1 2025-12-04T09:05:03.7894471Z warning_on_retry: true 2025-12-04T09:05:03.7894663Z continue_on_error: false 2025-12-04T09:05:03.7894849Z env: 2025-12-04T09:05:03.7895002Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:05:03.7895189Z HAS_NVIDIA_GPU: true 2025-12-04T09:05:03.7895429Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:05:03.7895812Z GITHUB_TOKEN: *** 2025-12-04T09:05:03.7895988Z ##[endgroup] 2025-12-04T09:05:03.8782391Z + python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-12-04T09:05:04.0886643Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T09:05:04.2503615Z Collecting requests==2.27.1 2025-12-04T09:05:04.2647714Z Downloading requests-2.27.1-py2.py3-none-any.whl (63 kB) 2025-12-04T09:05:04.4607200Z Collecting pyyaml==6.0.2 2025-12-04T09:05:04.4752133Z Downloading PyYAML-6.0.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (737 kB) 2025-12-04T09:05:04.8261248Z Collecting charset-normalizer~=2.0.0 2025-12-04T09:05:04.8300295Z Downloading charset_normalizer-2.0.12-py3-none-any.whl (39 kB) 2025-12-04T09:05:04.8357390Z Requirement already satisfied: idna<4,>=2.5 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (2.10) 2025-12-04T09:05:04.8360985Z Requirement already satisfied: urllib3<1.27,>=1.21.1 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (1.25.10) 2025-12-04T09:05:04.8799763Z Collecting certifi>=2017.4.17 2025-12-04T09:05:04.8839200Z Downloading certifi-2025.11.12-py3-none-any.whl (159 kB) 2025-12-04T09:05:04.9622164Z Installing collected packages: charset-normalizer, certifi, requests, pyyaml 2025-12-04T09:05:05.0754171Z Successfully installed certifi-2025.11.12 charset-normalizer-2.0.12 pyyaml-6.0.2 requests-2.27.1 2025-12-04T09:05:05.8588749Z Command completed after 1 attempt(s). 2025-12-04T09:05:05.8653704Z ##[group]Run set -x 2025-12-04T09:05:05.8653937Z set -x 2025-12-04T09:05:05.8654109Z  2025-12-04T09:05:05.8654390Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-12-04T09:05:05.8654741Z # in runner workspace 2025-12-04T09:05:05.8655244Z python3 "${GITHUB_ACTION_PATH}/../../scripts/parse_ref.py" 2025-12-04T09:05:05.8662864Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:05:05.8663139Z env: 2025-12-04T09:05:05.8663289Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:05:05.8663479Z HAS_NVIDIA_GPU: true 2025-12-04T09:05:05.8663706Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:05:05.8663952Z ##[endgroup] 2025-12-04T09:05:05.8689860Z + python3 /home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/filter-test-configs/../../scripts/parse_ref.py 2025-12-04T09:05:05.8855514Z Setting output branch=main 2025-12-04T09:05:05.8902051Z ##[group]Run echo "Workflow: ${GITHUB_WORKFLOW}" 2025-12-04T09:05:05.8902380Z echo "Workflow: ${GITHUB_WORKFLOW}" 2025-12-04T09:05:05.8902627Z echo "Job name: ${JOB_NAME}" 2025-12-04T09:05:05.8902843Z  2025-12-04T09:05:05.8903124Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-12-04T09:05:05.8903489Z # in runner workspace 2025-12-04T09:05:05.8903801Z python3 "${GITHUB_ACTION_PATH}/../../scripts/filter_test_configs.py" \ 2025-12-04T09:05:05.8904202Z  --workflow "${GITHUB_WORKFLOW}" \ 2025-12-04T09:05:05.8904456Z  --job-name "${JOB_NAME}" \ 2025-12-04T09:05:05.8911814Z  --test-matrix "{"include": [{"config": "default", "shard": 1, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 1, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 2, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 2, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 3, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 3, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 4, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 4, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 5, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 5, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "pr_time_benchmarks", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "pr_time_benchmarks", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "libtorch_agnostic_targetting", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "libtorch_agnostic_targetting", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}]}" \ 2025-12-04T09:05:05.8919121Z  --selected-test-configs "" \ 2025-12-04T09:05:05.8919384Z  --pr-number "${PR_NUMBER}" \ 2025-12-04T09:05:05.8919611Z  --tag "${TAG}" \ 2025-12-04T09:05:05.8919837Z  --event-name "${EVENT_NAME}" \ 2025-12-04T09:05:05.8920081Z  --schedule "${SCHEDULE}" \ 2025-12-04T09:05:05.8920307Z  --branch "${HEAD_BRANCH}" 2025-12-04T09:05:05.8927369Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:05:05.8927653Z env: 2025-12-04T09:05:05.8927818Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:05:05.8928016Z HAS_NVIDIA_GPU: true 2025-12-04T09:05:05.8928253Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:05:05.8928709Z GITHUB_TOKEN: *** 2025-12-04T09:05:05.8929159Z JOB_NAME: linux-jammy-cuda12.8-py3.10-gcc11 / test (default, 4, 5, lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests) 2025-12-04T09:05:05.8929648Z PR_NUMBER: 2025-12-04T09:05:05.8929818Z TAG: 2025-12-04T09:05:05.8929970Z EVENT_NAME: schedule 2025-12-04T09:05:05.8930162Z SCHEDULE: 29 8 * * * 2025-12-04T09:05:05.8930341Z HEAD_BRANCH: main 2025-12-04T09:05:05.8930508Z ##[endgroup] 2025-12-04T09:05:05.8954373Z Workflow: trunk 2025-12-04T09:05:05.8954974Z Job name: linux-jammy-cuda12.8-py3.10-gcc11 / test (default, 4, 5, lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests) 2025-12-04T09:05:06.0562149Z Setting output keep-going=True 2025-12-04T09:05:06.0562489Z Setting output ci-verbose-test-logs=False 2025-12-04T09:05:06.0562815Z Setting output ci-test-showlocals=False 2025-12-04T09:05:06.0563110Z Setting output ci-no-test-timeout=False 2025-12-04T09:05:06.0563394Z Setting output ci-no-td=False 2025-12-04T09:05:06.0563684Z Setting output ci-td-distributed=False 2025-12-04T09:05:06.0563967Z Setting output is-unstable=False 2025-12-04T09:05:06.0564233Z Setting output reenabled-issues= 2025-12-04T09:05:06.0581037Z Setting output test-matrix={"include": [{"config": "default", "shard": 1, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 1, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 1, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 1, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 2, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 2, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 2, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 2, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 3, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 3, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 3, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 3, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 4, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 4, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 4, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 4, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 5, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 5, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 5, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 5, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "pr_time_benchmarks", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "pr_time_benchmarks", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "pr_time_benchmarks", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "pr_time_benchmarks", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "libtorch_agnostic_targetting", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "libtorch_agnostic_targetting", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "libtorch_agnostic_targetting", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "libtorch_agnostic_targetting", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}]} 2025-12-04T09:05:06.0596703Z Setting output is-test-matrix-empty=False 2025-12-04T09:05:06.0664377Z ##[group]Run echo "Filtered matrix:" 2025-12-04T09:05:06.0664686Z echo "Filtered matrix:" 2025-12-04T09:05:06.0680594Z echo "{"include": [{"config": "default", "shard": 1, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 1, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 1, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 1, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 2, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 2, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 2, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 2, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 3, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 3, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 3, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 3, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 4, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 4, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 4, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 4, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 5, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 5, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 5, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 5, "num_shards": 5, "runner": "lf.linux.g6.4xlarge.experimental.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "lf.linux.g4dn.12xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "pr_time_benchmarks", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "pr_time_benchmarks", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "pr_time_benchmarks", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "pr_time_benchmarks", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "libtorch_agnostic_targetting", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "mem_leak_check": "mem_leak_check"}, {"config": "libtorch_agnostic_targetting", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "libtorch_agnostic_targetting", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "libtorch_agnostic_targetting", "shard": 1, "num_shards": 1, "runner": "linux.g4dn.metal.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests"}]}" 2025-12-04T09:05:06.0696364Z  2025-12-04T09:05:06.0696521Z echo 2025-12-04T09:05:06.0696732Z echo "Is the current job unstable? False" 2025-12-04T09:05:06.0697058Z  2025-12-04T09:05:06.0697203Z echo 2025-12-04T09:05:06.0697395Z echo "Is keep-going label set? True" 2025-12-04T09:05:06.0697639Z  2025-12-04T09:05:06.0697786Z echo 2025-12-04T09:05:06.0697962Z echo "Reenabled issues? " 2025-12-04T09:05:06.0705552Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:05:06.0705854Z env: 2025-12-04T09:05:06.0706017Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:05:06.0706216Z HAS_NVIDIA_GPU: true 2025-12-04T09:05:06.0706455Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:05:06.0706721Z ##[endgroup] 2025-12-04T09:05:06.0732029Z Filtered matrix: 2025-12-04T09:05:06.0752496Z {include: [{config: default, shard: 1, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, mem_leak_check: mem_leak_check}, {config: default, shard: 1, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 1, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: default, shard: 1, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 2, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, mem_leak_check: mem_leak_check}, {config: default, shard: 2, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 2, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: default, shard: 2, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 3, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, mem_leak_check: mem_leak_check}, {config: default, shard: 3, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 3, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: default, shard: 3, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 4, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, mem_leak_check: mem_leak_check}, {config: default, shard: 4, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 4, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: default, shard: 4, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 5, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, mem_leak_check: mem_leak_check}, {config: default, shard: 5, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 5, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: default, shard: 5, num_shards: 5, runner: lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests}, {config: distributed, shard: 1, num_shards: 3, runner: lf.linux.g4dn.12xlarge.nvidia.gpu, mem_leak_check: mem_leak_check}, {config: distributed, shard: 1, num_shards: 3, runner: lf.linux.g4dn.12xlarge.nvidia.gpu, mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: distributed, shard: 1, num_shards: 3, runner: lf.linux.g4dn.12xlarge.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: distributed, shard: 1, num_shards: 3, runner: lf.linux.g4dn.12xlarge.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests}, {config: distributed, shard: 2, num_shards: 3, runner: lf.linux.g4dn.12xlarge.nvidia.gpu, mem_leak_check: mem_leak_check}, {config: distributed, shard: 2, num_shards: 3, runner: lf.linux.g4dn.12xlarge.nvidia.gpu, mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: distributed, shard: 2, num_shards: 3, runner: lf.linux.g4dn.12xlarge.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: distributed, shard: 2, num_shards: 3, runner: lf.linux.g4dn.12xlarge.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests}, {config: distributed, shard: 3, num_shards: 3, runner: lf.linux.g4dn.12xlarge.nvidia.gpu, mem_leak_check: mem_leak_check}, {config: distributed, shard: 3, num_shards: 3, runner: lf.linux.g4dn.12xlarge.nvidia.gpu, mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: distributed, shard: 3, num_shards: 3, runner: lf.linux.g4dn.12xlarge.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: distributed, shard: 3, num_shards: 3, runner: lf.linux.g4dn.12xlarge.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests}, {config: pr_time_benchmarks, shard: 1, num_shards: 1, runner: linux.g4dn.metal.nvidia.gpu, mem_leak_check: mem_leak_check}, {config: pr_time_benchmarks, shard: 1, num_shards: 1, runner: linux.g4dn.metal.nvidia.gpu, mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: pr_time_benchmarks, shard: 1, num_shards: 1, runner: linux.g4dn.metal.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: pr_time_benchmarks, shard: 1, num_shards: 1, runner: linux.g4dn.metal.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests}, {config: libtorch_agnostic_targetting, shard: 1, num_shards: 1, runner: linux.g4dn.metal.nvidia.gpu, mem_leak_check: mem_leak_check}, {config: libtorch_agnostic_targetting, shard: 1, num_shards: 1, runner: linux.g4dn.metal.nvidia.gpu, mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: libtorch_agnostic_targetting, shard: 1, num_shards: 1, runner: linux.g4dn.metal.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: libtorch_agnostic_targetting, shard: 1, num_shards: 1, runner: linux.g4dn.metal.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests}]} 2025-12-04T09:05:06.0767901Z 2025-12-04T09:05:06.0768003Z Is the current job unstable? False 2025-12-04T09:05:06.0768160Z 2025-12-04T09:05:06.0768253Z Is keep-going label set? True 2025-12-04T09:05:06.0768392Z 2025-12-04T09:05:06.0768468Z Reenabled issues? 2025-12-04T09:05:06.0794622Z ##[group]Run echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-12-04T09:05:06.0795055Z echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-12-04T09:05:06.0802055Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:05:06.0802343Z env: 2025-12-04T09:05:06.0802502Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:05:06.0802697Z HAS_NVIDIA_GPU: true 2025-12-04T09:05:06.0802928Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:05:06.0803186Z JOB_TIMEOUT: 360 2025-12-04T09:05:06.0803348Z ##[endgroup] 2025-12-04T09:05:06.0848273Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:05:06.0848698Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:05:06.0849033Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:05:06.0856001Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:05:06.0856276Z env: 2025-12-04T09:05:06.0856429Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:05:06.0856749Z HAS_NVIDIA_GPU: true 2025-12-04T09:05:06.0856979Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:05:06.0857246Z ##[endgroup] 2025-12-04T09:05:06.0949664Z ##[group]Run set -x 2025-12-04T09:05:06.0949942Z set -x 2025-12-04T09:05:06.0950110Z  2025-12-04T09:05:06.0950299Z if [[ $TEST_CONFIG == 'multigpu' ]]; then 2025-12-04T09:05:06.0950589Z  TEST_COMMAND=.ci/pytorch/multigpu-test.sh 2025-12-04T09:05:06.0950873Z elif [[ $BUILD_ENVIRONMENT == *onnx* ]]; then 2025-12-04T09:05:06.0951144Z  TEST_COMMAND=.ci/onnx/test.sh 2025-12-04T09:05:06.0951364Z else 2025-12-04T09:05:06.0951551Z  TEST_COMMAND=.ci/pytorch/test.sh 2025-12-04T09:05:06.0951765Z fi 2025-12-04T09:05:06.0951912Z  2025-12-04T09:05:06.0952096Z # Leaving 1GB for the runner and other things 2025-12-04T09:05:06.0952510Z TOTAL_AVAILABLE_MEMORY_IN_GB=$(awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo) 2025-12-04T09:05:06.0953144Z # https://docs.docker.com/engine/containers/resource_constraints/#--memory-swap-details, the 3GB swap 2025-12-04T09:05:06.0953635Z # comes from https://github.com/pytorch/test-infra/pull/6058 2025-12-04T09:05:06.0954017Z TOTAL_MEMORY_WITH_SWAP=$(("${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}" + 3)) 2025-12-04T09:05:06.0954310Z  2025-12-04T09:05:06.0954516Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-12-04T09:05:06.0954763Z  SHM_OPTS= 2025-12-04T09:05:06.0954945Z  JENKINS_USER= 2025-12-04T09:05:06.0955202Z  # ensure that docker container cleanly exits in 12 hours 2025-12-04T09:05:06.0955538Z  # if for some reason cleanup action doesn't stop container 2025-12-04T09:05:06.0955816Z  # when job is cancelled 2025-12-04T09:05:06.0956031Z  DOCKER_SHELL_CMD="sleep 12h" 2025-12-04T09:05:06.0956259Z  USED_IMAGE="${DOCKER_IMAGE_S390X}" 2025-12-04T09:05:06.0956481Z else 2025-12-04T09:05:06.0956658Z  SHM_OPTS="--shm-size=${SHM_SIZE}" 2025-12-04T09:05:06.0956889Z  JENKINS_USER="--user jenkins" 2025-12-04T09:05:06.0957107Z  DOCKER_SHELL_CMD= 2025-12-04T09:05:06.0957306Z  USED_IMAGE="${DOCKER_IMAGE}" 2025-12-04T09:05:06.0957505Z fi 2025-12-04T09:05:06.0957643Z  2025-12-04T09:05:06.0957875Z # detached container should get cleaned up by teardown_ec2_linux 2025-12-04T09:05:06.0958263Z # TODO: Stop building test binaries as part of the build phase 2025-12-04T09:05:06.0958684Z # Used for GPU_FLAG, SHM_OPTS, JENKINS_USER and DOCKER_SHELL_CMD since that doesn't play nice 2025-12-04T09:05:06.0959057Z # shellcheck disable=SC2086,SC2090 2025-12-04T09:05:06.0959294Z container_name=$(docker run \ 2025-12-04T09:05:06.0959518Z  ${GPU_FLAG:-} \ 2025-12-04T09:05:06.0959726Z  ${SCCACHE_SERVER_PORT_DOCKER_FLAG:-} \ 2025-12-04T09:05:06.0959974Z  -e BUILD_ENVIRONMENT \ 2025-12-04T09:05:06.0960185Z  -e PR_NUMBER \ 2025-12-04T09:05:06.0960376Z  -e GITHUB_ACTIONS \ 2025-12-04T09:05:06.0960578Z  -e GITHUB_REPOSITORY \ 2025-12-04T09:05:06.0960787Z  -e GITHUB_WORKFLOW \ 2025-12-04T09:05:06.0960980Z  -e GITHUB_JOB \ 2025-12-04T09:05:06.0961167Z  -e GITHUB_RUN_ID \ 2025-12-04T09:05:06.0961361Z  -e GITHUB_RUN_NUMBER \ 2025-12-04T09:05:06.0961566Z  -e GITHUB_RUN_ATTEMPT \ 2025-12-04T09:05:06.0961763Z  -e JOB_ID \ 2025-12-04T09:05:06.0961943Z  -e JOB_NAME \ 2025-12-04T09:05:06.0962125Z  -e BASE_SHA \ 2025-12-04T09:05:06.0962293Z  -e BRANCH \ 2025-12-04T09:05:06.0962462Z  -e SHA1 \ 2025-12-04T09:05:06.0962637Z  -e AWS_DEFAULT_REGION \ 2025-12-04T09:05:06.0962846Z  -e IN_WHEEL_TEST \ 2025-12-04T09:05:06.0963041Z  -e SHARD_NUMBER \ 2025-12-04T09:05:06.0963371Z  -e TEST_CONFIG \ 2025-12-04T09:05:06.0963555Z  -e NUM_TEST_SHARDS \ 2025-12-04T09:05:06.0963860Z  -e REENABLED_ISSUES \ 2025-12-04T09:05:06.0964081Z  -e CONTINUE_THROUGH_ERROR \ 2025-12-04T09:05:06.0964304Z  -e VERBOSE_TEST_LOGS \ 2025-12-04T09:05:06.0964503Z  -e TEST_SHOWLOCALS \ 2025-12-04T09:05:06.0964700Z  -e NO_TEST_TIMEOUT \ 2025-12-04T09:05:06.0964886Z  -e NO_TD \ 2025-12-04T09:05:06.0965055Z  -e TD_DISTRIBUTED \ 2025-12-04T09:05:06.0965247Z  -e PR_LABELS \ 2025-12-04T09:05:06.0965462Z  -e MAX_JOBS="$(nproc --ignore=2)" \ 2025-12-04T09:05:06.0965697Z  -e SCCACHE_BUCKET \ 2025-12-04T09:05:06.0965896Z  -e SCCACHE_REGION \ 2025-12-04T09:05:06.0966085Z  -e XLA_CUDA \ 2025-12-04T09:05:06.0966275Z  -e XLA_CLANG_CACHE_S3_BUCKET_NAME \ 2025-12-04T09:05:06.0966529Z  -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK \ 2025-12-04T09:05:06.0966793Z  -e PYTORCH_TEST_RERUN_DISABLED_TESTS \ 2025-12-04T09:05:06.0967051Z  -e SKIP_SCCACHE_INITIALIZATION=1 \ 2025-12-04T09:05:06.0967281Z  -e HUGGING_FACE_HUB_TOKEN \ 2025-12-04T09:05:06.0967510Z  -e VLLM_TEST_HUGGING_FACE_TOKEN \ 2025-12-04T09:05:06.0967748Z  -e SCRIBE_GRAPHQL_ACCESS_TOKEN \ 2025-12-04T09:05:06.0967971Z  -e DASHBOARD_TAG \ 2025-12-04T09:05:06.0968182Z  -e ARTIFACTS_FILE_SUFFIX \ 2025-12-04T09:05:06.0968435Z  --memory="${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}g" \ 2025-12-04T09:05:06.0968727Z  --memory-swap="${TOTAL_MEMORY_WITH_SWAP}g" \ 2025-12-04T09:05:06.0969011Z  --env-file="/tmp/github_env_${GITHUB_RUN_ID}" \ 2025-12-04T09:05:06.0969296Z  --security-opt seccomp=unconfined \ 2025-12-04T09:05:06.0969537Z  --cap-add=SYS_PTRACE \ 2025-12-04T09:05:06.0969736Z  --ipc=host \ 2025-12-04T09:05:06.0969922Z  ${SHM_OPTS} \ 2025-12-04T09:05:06.0970101Z  --tty \ 2025-12-04T09:05:06.0970260Z  --detach \ 2025-12-04T09:05:06.0970448Z  --name="${container_name}" \ 2025-12-04T09:05:06.0970668Z  ${JENKINS_USER} \ 2025-12-04T09:05:06.0970910Z  -v "${GITHUB_WORKSPACE}:/var/lib/jenkins/workspace" \ 2025-12-04T09:05:06.0971182Z  -w /var/lib/jenkins/workspace \ 2025-12-04T09:05:06.0971401Z  "${USED_IMAGE}" \ 2025-12-04T09:05:06.0971603Z  ${DOCKER_SHELL_CMD} 2025-12-04T09:05:06.0971781Z ) 2025-12-04T09:05:06.0972014Z echo "DOCKER_CONTAINER_ID=${container_name}" >> "${GITHUB_ENV}" 2025-12-04T09:05:06.0972301Z  2025-12-04T09:05:06.0972481Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-12-04T09:05:06.0972891Z  docker exec -t "${container_name}" sh -c "python3 -m pip install -r .ci/docker/requirements-ci.txt" 2025-12-04T09:05:06.0973341Z fi 2025-12-04T09:05:06.0973493Z  2025-12-04T09:05:06.0973835Z docker exec -t "${container_name}" sh -c "python3 -m pip install $(echo dist/*.whl)[opt-einsum] && ${TEST_COMMAND}" 2025-12-04T09:05:06.0980600Z shell: /usr/bin/bash -e {0} 2025-12-04T09:05:06.0980795Z env: 2025-12-04T09:05:06.0980943Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:05:06.0981128Z HAS_NVIDIA_GPU: true 2025-12-04T09:05:06.0981350Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:05:06.0981653Z BUILD_ENVIRONMENT: linux-jammy-cuda12.8-py3.10-gcc11 2025-12-04T09:05:06.0981903Z PR_NUMBER: 2025-12-04T09:05:06.0982076Z GITHUB_REPOSITORY: pytorch/pytorch 2025-12-04T09:05:06.0982292Z GITHUB_WORKFLOW: trunk 2025-12-04T09:05:06.0982463Z GITHUB_JOB: test 2025-12-04T09:05:06.0982626Z GITHUB_RUN_ID: 19922768520 2025-12-04T09:05:06.0982811Z GITHUB_RUN_NUMBER: 158165 2025-12-04T09:05:06.0982988Z GITHUB_RUN_ATTEMPT: 1 2025-12-04T09:05:06.0983153Z JOB_ID: 57116084869 2025-12-04T09:05:06.0983575Z JOB_NAME: linux-jammy-cuda12.8-py3.10-gcc11 / test (default, 4, 5, lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests) 2025-12-04T09:05:06.0984204Z BRANCH: main 2025-12-04T09:05:06.0984411Z SHA1: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:05:06.0984694Z BASE_SHA: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:05:06.0984931Z TEST_CONFIG: default 2025-12-04T09:05:06.0985090Z SHARD_NUMBER: 4 2025-12-04T09:05:06.0985249Z NUM_TEST_SHARDS: 5 2025-12-04T09:05:06.0985411Z EXTRA_FLAGS: 2025-12-04T09:05:06.0985564Z OP_BENCHMARK_TESTS: 2025-12-04T09:05:06.0985731Z REENABLED_ISSUES: 2025-12-04T09:05:06.0985907Z CONTINUE_THROUGH_ERROR: True 2025-12-04T09:05:06.0986094Z VERBOSE_TEST_LOGS: False 2025-12-04T09:05:06.0986281Z TEST_SHOWLOCALS: False 2025-12-04T09:05:06.0986459Z NO_TEST_TIMEOUT: False 2025-12-04T09:05:06.0986625Z NO_TD: False 2025-12-04T09:05:06.0986781Z TD_DISTRIBUTED: False 2025-12-04T09:05:06.0987000Z SCCACHE_BUCKET: ossci-compiler-cache-circleci-v2 2025-12-04T09:05:06.0987253Z SCCACHE_REGION: us-east-1 2025-12-04T09:05:06.0987447Z SHM_SIZE: 2g 2025-12-04T09:05:06.0987991Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:05:06.0988964Z DOCKER_IMAGE_S390X: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:05:06.0989558Z XLA_CUDA: 2025-12-04T09:05:06.0989810Z XLA_CLANG_CACHE_S3_BUCKET_NAME: ossci-compiler-clang-cache-circleci-xla 2025-12-04T09:05:06.0990129Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK: 0 2025-12-04T09:05:06.0990355Z PYTORCH_TEST_RERUN_DISABLED_TESTS: 1 2025-12-04T09:05:06.0990558Z DASHBOARD_TAG: 2025-12-04T09:05:06.0990868Z VLLM_TEST_HUGGING_FACE_TOKEN: *** 2025-12-04T09:05:06.0991158Z HUGGING_FACE_HUB_TOKEN: *** 2025-12-04T09:05:06.0991454Z SCRIBE_GRAPHQL_ACCESS_TOKEN: *** 2025-12-04T09:05:06.0991836Z ARTIFACTS_FILE_SUFFIX: test-default-4-5-lf.linux.g6.4xlarge.experimental.nvidia.gpu_57116084869 2025-12-04T09:05:06.0992229Z ##[endgroup] 2025-12-04T09:05:06.1015807Z + [[ default == \m\u\l\t\i\g\p\u ]] 2025-12-04T09:05:06.1016163Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *onnx* ]] 2025-12-04T09:05:06.1016511Z + TEST_COMMAND=.ci/pytorch/test.sh 2025-12-04T09:05:06.1018993Z ++ awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo 2025-12-04T09:05:06.1040968Z + TOTAL_AVAILABLE_MEMORY_IN_GB='59.453 ' 2025-12-04T09:05:06.1041438Z + TOTAL_MEMORY_WITH_SWAP=62 2025-12-04T09:05:06.1041919Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *\s\3\9\0\x* ]] 2025-12-04T09:05:06.1042271Z + SHM_OPTS=--shm-size=2g 2025-12-04T09:05:06.1042519Z + JENKINS_USER='--user jenkins' 2025-12-04T09:05:06.1042766Z + DOCKER_SHELL_CMD= 2025-12-04T09:05:06.1043455Z + USED_IMAGE=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:05:06.1050324Z +++ nproc --ignore=2 2025-12-04T09:05:06.1088374Z ++ docker run --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all -e BUILD_ENVIRONMENT -e PR_NUMBER -e GITHUB_ACTIONS -e GITHUB_REPOSITORY -e GITHUB_WORKFLOW -e GITHUB_JOB -e GITHUB_RUN_ID -e GITHUB_RUN_NUMBER -e GITHUB_RUN_ATTEMPT -e JOB_ID -e JOB_NAME -e BASE_SHA -e BRANCH -e SHA1 -e AWS_DEFAULT_REGION -e IN_WHEEL_TEST -e SHARD_NUMBER -e TEST_CONFIG -e NUM_TEST_SHARDS -e REENABLED_ISSUES -e CONTINUE_THROUGH_ERROR -e VERBOSE_TEST_LOGS -e TEST_SHOWLOCALS -e NO_TEST_TIMEOUT -e NO_TD -e TD_DISTRIBUTED -e PR_LABELS -e MAX_JOBS=14 -e SCCACHE_BUCKET -e SCCACHE_REGION -e XLA_CUDA -e XLA_CLANG_CACHE_S3_BUCKET_NAME -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK -e PYTORCH_TEST_RERUN_DISABLED_TESTS -e SKIP_SCCACHE_INITIALIZATION=1 -e HUGGING_FACE_HUB_TOKEN -e VLLM_TEST_HUGGING_FACE_TOKEN -e SCRIBE_GRAPHQL_ACCESS_TOKEN -e DASHBOARD_TAG -e ARTIFACTS_FILE_SUFFIX --memory=59g --memory-swap=62g --env-file=/tmp/github_env_19922768520 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --ipc=host --shm-size=2g --tty --detach --name= --user jenkins -v /home/ec2-user/actions-runner/_work/pytorch/pytorch:/var/lib/jenkins/workspace -w /var/lib/jenkins/workspace 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:05:15.2039169Z + container_name=364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T09:05:15.2040276Z + echo DOCKER_CONTAINER_ID=364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T09:05:15.2040888Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *\s\3\9\0\x* ]] 2025-12-04T09:05:15.2045308Z ++ echo dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl 2025-12-04T09:05:15.2047717Z + docker exec -t 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 sh -c 'python3 -m pip install dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl[opt-einsum] && .ci/pytorch/test.sh' 2025-12-04T09:05:15.6379015Z Processing ./dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl (from torch==2.10.0a0+gitffd9b0f) 2025-12-04T09:05:15.9931804Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.18.0) 2025-12-04T09:05:15.9935682Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (4.12.2) 2025-12-04T09:05:15.9940287Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (1.13.3) 2025-12-04T09:05:15.9945167Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (2.8.8) 2025-12-04T09:05:15.9948823Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.1.6) 2025-12-04T09:05:15.9953577Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (2025.10.0) 2025-12-04T09:05:15.9968096Z Requirement already satisfied: opt-einsum>=3.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.3.0) 2025-12-04T09:05:16.0296726Z Requirement already satisfied: numpy>=1.7 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from opt-einsum>=3.3->torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (1.22.4) 2025-12-04T09:05:16.0314893Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from sympy>=1.13.3->torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (1.3.0) 2025-12-04T09:05:16.0366936Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from jinja2->torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.0.3) 2025-12-04T09:05:16.3874580Z Installing collected packages: torch 2025-12-04T09:05:27.1964553Z Successfully installed torch-2.10.0a0+gitffd9b0f 2025-12-04T09:05:27.2465827Z + export TERM=vt100 2025-12-04T09:05:27.2466101Z + TERM=vt100 2025-12-04T09:05:27.2468321Z ++ dirname .ci/pytorch/test.sh 2025-12-04T09:05:27.2479163Z + source .ci/pytorch/common.sh 2025-12-04T09:05:27.2482407Z +++ dirname .ci/pytorch/common.sh 2025-12-04T09:05:27.2491194Z ++ source .ci/pytorch/common_utils.sh 2025-12-04T09:05:27.2492902Z +++ declare -f -t trap_add 2025-12-04T09:05:27.2498048Z ++ set -ex -o pipefail 2025-12-04T09:05:27.2498306Z ++ [[ linux-jammy-cuda12.8-py3.10-gcc11 == *rocm* ]] 2025-12-04T09:05:27.2498580Z ++ BUILD_TEST_LIBTORCH=0 2025-12-04T09:05:27.2501772Z ++ dirname .ci/pytorch/test.sh 2025-12-04T09:05:27.2509614Z + source .ci/pytorch/common-build.sh 2025-12-04T09:05:27.2511127Z ++ [[ linux-jammy-cuda12.8-py3.10-gcc11 != *win-* ]] 2025-12-04T09:05:27.2517089Z ++++ dirname .ci/pytorch/common-build.sh 2025-12-04T09:05:27.2525065Z +++ cd .ci/pytorch 2025-12-04T09:05:27.2526521Z +++ pwd -P 2025-12-04T09:05:27.2528212Z ++ script_dir=/var/lib/jenkins/workspace/.ci/pytorch 2025-12-04T09:05:27.2528749Z ++ [[ linux-jammy-cuda12.8-py3.10-gcc11 == *-pch* ]] 2025-12-04T09:05:27.2529082Z ++ which sccache 2025-12-04T09:05:27.2541932Z ++ [[ -z ossci-compiler-cache-circleci-v2 ]] 2025-12-04T09:05:27.2542319Z ++ sccache --stop-server 2025-12-04T09:05:27.2568756Z ++ true 2025-12-04T09:05:27.2569078Z ++ rm -f /var/lib/jenkins/sccache_error.log 2025-12-04T09:05:27.2580107Z ++ trap_add sccache_epilogue EXIT 2025-12-04T09:05:27.2580549Z ++ trap_add_cmd=sccache_epilogue 2025-12-04T09:05:27.2580808Z ++ shift 2025-12-04T09:05:27.2581006Z ++ for trap_add_name in "$@" 2025-12-04T09:05:27.2586763Z ++++ trap -p EXIT 2025-12-04T09:05:27.2589830Z +++ eval 'extract_trap_cmd ' 2025-12-04T09:05:27.2590100Z ++++ extract_trap_cmd 2025-12-04T09:05:27.2590361Z ++++ printf '%s\n' '' 2025-12-04T09:05:27.2590780Z +++ printf '%s\n' sccache_epilogue 2025-12-04T09:05:27.2592747Z ++ trap -- ' 2025-12-04T09:05:27.2593105Z sccache_epilogue' EXIT 2025-12-04T09:05:27.2593490Z ++ [[ -n 1 ]] 2025-12-04T09:05:27.2594093Z ++ echo 'Skipping sccache server initialization, setting environment variables' 2025-12-04T09:05:27.2595039Z Skipping sccache server initialization, setting environment variables 2025-12-04T09:05:27.2595691Z ++ export SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:05:27.2595969Z ++ SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:05:27.2596291Z ++ export SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:05:27.2596709Z ++ SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:05:27.2602527Z ++ export RUST_LOG=sccache::server=error 2025-12-04T09:05:27.2602806Z ++ RUST_LOG=sccache::server=error 2025-12-04T09:05:27.2603028Z ++ sccache --zero-stats 2025-12-04T09:05:27.3653496Z Statistics zeroed. 2025-12-04T09:05:27.3658693Z ++ which ccache 2025-12-04T09:05:27.3671929Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 != *rocm* ]] 2025-12-04T09:05:27.3672349Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 != *s390x* ]] 2025-12-04T09:05:27.3672694Z + [[ -d /var/lib/jenkins/workspace ]] 2025-12-04T09:05:27.3674823Z ++ stat -c %u /var/lib/jenkins/workspace 2025-12-04T09:05:27.3690649Z + WORKSPACE_ORIGINAL_OWNER_ID=1000 2025-12-04T09:05:27.3690951Z + trap_add cleanup_workspace EXIT 2025-12-04T09:05:27.3691227Z + trap_add_cmd=cleanup_workspace 2025-12-04T09:05:27.3691468Z + shift 2025-12-04T09:05:27.3691671Z + for trap_add_name in "$@" 2025-12-04T09:05:27.3697409Z +++ trap -p EXIT 2025-12-04T09:05:27.3700276Z ++ eval 'extract_trap_cmd trap -- '\'' 2025-12-04T09:05:27.3711421Z sccache_epilogue'\'' EXIT' 2025-12-04T09:05:27.3711654Z +++ extract_trap_cmd trap -- ' 2025-12-04T09:05:27.3711867Z sccache_epilogue' EXIT 2025-12-04T09:05:27.3712067Z +++ printf '%s\n' ' 2025-12-04T09:05:27.3712250Z sccache_epilogue' 2025-12-04T09:05:27.3712440Z ++ printf '%s\n' cleanup_workspace 2025-12-04T09:05:27.3712661Z + trap -- ' 2025-12-04T09:05:27.3712822Z sccache_epilogue 2025-12-04T09:05:27.3713005Z cleanup_workspace' EXIT 2025-12-04T09:05:27.3713231Z + sudo chown -R jenkins /var/lib/jenkins/workspace 2025-12-04T09:05:28.2721414Z + git config --global --add safe.directory /var/lib/jenkins/workspace 2025-12-04T09:05:28.2742604Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *cuda* ]] 2025-12-04T09:05:28.2746689Z ++ python -c 'import os;import numba.cuda; print(os.path.dirname(numba.cuda.__file__))' 2025-12-04T09:05:28.6324584Z + NUMBA_CUDA_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda 2025-12-04T09:05:28.6325193Z + '[' -n /opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda ']' 2025-12-04T09:05:28.6329821Z +++ realpath .ci/pytorch/test.sh 2025-12-04T09:05:28.6340415Z ++ dirname /var/lib/jenkins/workspace/.ci/pytorch/test.sh 2025-12-04T09:05:28.6349864Z + NUMBA_PATCH=/var/lib/jenkins/workspace/.ci/pytorch/numba-cuda-13.patch 2025-12-04T09:05:28.6350942Z + pushd /opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda 2025-12-04T09:05:28.6351749Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda ~/workspace 2025-12-04T09:05:28.6352189Z + patch -p4 2025-12-04T09:05:28.6365657Z patching file cudadrv/driver.py 2025-12-04T09:05:28.6365989Z Hunk #1 succeeded at 357 (offset -8 lines). 2025-12-04T09:05:28.6424289Z + popd 2025-12-04T09:05:28.6424535Z ~/workspace 2025-12-04T09:05:28.6424754Z + echo 'Environment variables:' 2025-12-04T09:05:28.6425031Z Environment variables: 2025-12-04T09:05:28.6425244Z + env 2025-12-04T09:05:28.6434287Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:05:28.6434878Z CONTINUE_THROUGH_ERROR=True 2025-12-04T09:05:28.6435373Z BUILD_ENVIRONMENT=linux-jammy-cuda12.8-py3.10-gcc11 2025-12-04T09:05:28.6436269Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-12-04T09:05:28.6436689Z HOSTNAME=364ec7eaeaf9 2025-12-04T09:05:28.6437542Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_4c1fca7f-6411-4e46-b998-dde57198cc6c 2025-12-04T09:05:28.6438164Z GITHUB_ACTION=__run_3 2025-12-04T09:05:28.6438367Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-12-04T09:05:28.6438589Z GITHUB_RUN_NUMBER=158165 2025-12-04T09:05:28.6438766Z TEST_CONFIG=default 2025-12-04T09:05:28.6438944Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-12-04T09:05:28.6439190Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-12-04T09:05:28.6439419Z SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:05:28.6439734Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-12-04T09:05:28.6439951Z GITHUB_TRIGGERING_ACTOR=huydhn 2025-12-04T09:05:28.6440155Z GITHUB_REF_TYPE=branch 2025-12-04T09:05:28.6440372Z BASE_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:05:28.6440605Z XLA_CUDA= 2025-12-04T09:05:28.6440771Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-12-04T09:05:28.6441069Z HUGGING_FACE_HUB_TOKEN=*** 2025-12-04T09:05:28.6441348Z *** 2025-12-04T09:05:28.6441508Z GITHUB_REPOSITORY_ID=65600975 2025-12-04T09:05:28.6441713Z GITHUB_ACTIONS=true 2025-12-04T09:05:28.6442016Z NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:05:28.6442272Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:05:28.6442563Z SHA1=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:05:28.6442839Z GITHUB_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:05:28.6443201Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/trunk.yml@refs/heads/main 2025-12-04T09:05:28.6443531Z UCC_HOME=/usr 2025-12-04T09:05:28.6443689Z VERBOSE_TEST_LOGS=False 2025-12-04T09:05:28.6443866Z GITHUB_REF=refs/heads/main 2025-12-04T09:05:28.6444048Z SHARD_NUMBER=4 2025-12-04T09:05:28.6444219Z GITHUB_REF_PROTECTED=true 2025-12-04T09:05:28.6444397Z HOME=/var/lib/jenkins 2025-12-04T09:05:28.6444595Z GITHUB_API_URL=https://api.github.com 2025-12-04T09:05:28.6444842Z PYTORCH_TEST_RERUN_DISABLED_TESTS=1 2025-12-04T09:05:28.6445094Z UCX_COMMIT=7836b165abdbe468a2f607e7254011c07d788152 2025-12-04T09:05:28.6445334Z USE_SYSTEM_NCCL=1 2025-12-04T09:05:28.6445498Z NUM_TEST_SHARDS=5 2025-12-04T09:05:28.6445665Z UCX_HOME=/usr 2025-12-04T09:05:28.6446076Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_4c1fca7f-6411-4e46-b998-dde57198cc6c 2025-12-04T09:05:28.6446796Z JOB_NAME=linux-jammy-cuda12.8-py3.10-gcc11 / test (default, 4, 5, lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests) 2025-12-04T09:05:28.6447484Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_4c1fca7f-6411-4e46-b998-dde57198cc6c 2025-12-04T09:05:28.6448050Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-12-04T09:05:28.6448417Z GITHUB_EVENT_NAME=schedule 2025-12-04T09:05:28.6448602Z DASHBOARD_TAG= 2025-12-04T09:05:28.6448757Z GITHUB_RUN_ID=19922768520 2025-12-04T09:05:28.6448938Z INSTALLED_OPENBLAS= 2025-12-04T09:05:28.6449375Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_4c1fca7f-6411-4e46-b998-dde57198cc6c 2025-12-04T09:05:28.6449842Z GITHUB_ACTOR=huydhn 2025-12-04T09:05:28.6450184Z PR_NUMBER= 2025-12-04T09:05:28.6450334Z DESIRED_CUDA=12.8.1 2025-12-04T09:05:28.6450493Z GITHUB_RUN_ATTEMPT=1 2025-12-04T09:05:28.6450777Z ANACONDA_PYTHON_VERSION=3.10 2025-12-04T09:05:28.6451030Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-12-04T09:05:28.6451273Z TERM=vt100 2025-12-04T09:05:28.6451416Z INSTALLED_VISION=yes 2025-12-04T09:05:28.6451577Z BRANCH=main 2025-12-04T09:05:28.6451734Z SCCACHE_REGION=us-east-1 2025-12-04T09:05:28.6451922Z OPENSSL_ROOT_DIR=/opt/openssl 2025-12-04T09:05:28.6452132Z BUILD_AOT_INDUCTOR_TEST= 2025-12-04T09:05:28.6452315Z CUDA_PATH=/usr/local/cuda 2025-12-04T09:05:28.6452679Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-12-04T09:05:28.6453088Z GITHUB_SERVER_URL=https://github.com 2025-12-04T09:05:28.6453448Z UCC_COMMIT=430e241bf5d38cbc73fc7a6b89155397232e3f96 2025-12-04T09:05:28.6453681Z REENABLED_ISSUES= 2025-12-04T09:05:28.6453835Z DOCS= 2025-12-04T09:05:28.6453973Z SHLVL=1 2025-12-04T09:05:28.6454111Z MAX_JOBS=14 2025-12-04T09:05:28.6454265Z GITHUB_ACTOR_ID=475357 2025-12-04T09:05:28.6454506Z GITHUB_WORKFLOW_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:05:28.6454779Z GITHUB_REF_NAME=main 2025-12-04T09:05:28.6455044Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-12-04T09:05:28.6455345Z GITHUB_JOB=test 2025-12-04T09:05:28.6455510Z NO_TEST_TIMEOUT=False 2025-12-04T09:05:28.6455676Z TD_DISTRIBUTED=False 2025-12-04T09:05:28.6455862Z GITHUB_REPOSITORY=pytorch/pytorch 2025-12-04T09:05:28.6456082Z GITHUB_RETENTION_DAYS=90 2025-12-04T09:05:28.6456267Z OPENSSL_DIR=/opt/openssl 2025-12-04T09:05:28.6456454Z GITHUB_ACTION_REPOSITORY= 2025-12-04T09:05:28.6457005Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:05:28.6457564Z GITHUB_BASE_REF= 2025-12-04T09:05:28.6457725Z INSTALLED_ACL= 2025-12-04T09:05:28.6458071Z ARTIFACTS_FILE_SUFFIX=test-default-4-5-lf.linux.g6.4xlarge.experimental.nvidia.gpu_57116084869 2025-12-04T09:05:28.6458452Z CI=true 2025-12-04T09:05:28.6458607Z GITHUB_REPOSITORY_OWNER=pytorch 2025-12-04T09:05:28.6458852Z RUST_LOG=sccache::server=error 2025-12-04T09:05:28.6459041Z JOB_ID=57116084869 2025-12-04T09:05:28.6459205Z GITHUB_HEAD_REF= 2025-12-04T09:05:28.6459369Z GITHUB_ACTION_REF= 2025-12-04T09:05:28.6459573Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-12-04T09:05:28.6459818Z TEST_SHOWLOCALS=False 2025-12-04T09:05:28.6459994Z GITHUB_WORKFLOW=trunk 2025-12-04T09:05:28.6460172Z DEBIAN_FRONTEND=noninteractive 2025-12-04T09:05:28.6460617Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_4c1fca7f-6411-4e46-b998-dde57198cc6c 2025-12-04T09:05:28.6461063Z NO_TD=False 2025-12-04T09:05:28.6461226Z SKIP_SCCACHE_INITIALIZATION=1 2025-12-04T09:05:28.6461432Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-12-04T09:05:28.6461645Z _=/usr/bin/env 2025-12-04T09:05:28.6461899Z OLDPWD=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda 2025-12-04T09:05:28.6462272Z ++ python -c 'import site; print(site.getsitepackages()[0])' 2025-12-04T09:05:28.6558314Z + TORCH_INSTALL_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch 2025-12-04T09:05:28.6559191Z + TORCH_BIN_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-12-04T09:05:28.6560084Z + TORCH_LIB_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib 2025-12-04T09:05:28.6560665Z + TORCH_TEST_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/test 2025-12-04T09:05:28.6561041Z + BUILD_DIR=build 2025-12-04T09:05:28.6561261Z + BUILD_RENAMED_DIR=build_renamed 2025-12-04T09:05:28.6561525Z + BUILD_BIN_DIR=build/bin 2025-12-04T09:05:28.6561747Z + SHARD_NUMBER=4 2025-12-04T09:05:28.6561951Z + NUM_TEST_SHARDS=5 2025-12-04T09:05:28.6562309Z + export TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:05:28.6562586Z + TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:05:28.6562938Z + export VALGRIND=ON 2025-12-04T09:05:28.6563351Z + VALGRIND=ON 2025-12-04T09:05:28.6563596Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *clang9* ]] 2025-12-04T09:05:28.6564112Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *xpu* ]] 2025-12-04T09:05:28.6564428Z + detect_cuda_arch 2025-12-04T09:05:28.6564675Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *cuda* ]] 2025-12-04T09:05:28.6564982Z + command -v nvidia-smi 2025-12-04T09:05:28.6565213Z /usr/bin/nvidia-smi 2025-12-04T09:05:28.6569299Z ++ nvidia-smi --query-gpu=compute_cap --format=csv 2025-12-04T09:05:28.6570726Z ++ tail -n 1 2025-12-04T09:05:28.6772901Z + TORCH_CUDA_ARCH_LIST=8.9 2025-12-04T09:05:28.6773582Z + export TORCH_CUDA_ARCH_LIST 2025-12-04T09:05:28.6774194Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *s390x* ]] 2025-12-04T09:05:28.6774769Z + [[ 1 == \1 ]] 2025-12-04T09:05:28.6775137Z + ulimit -c 0 2025-12-04T09:05:28.6775611Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 != *bazel* ]] 2025-12-04T09:05:28.6780910Z ++ realpath build/custom_test_artifacts 2025-12-04T09:05:28.6789870Z + CUSTOM_TEST_ARTIFACT_BUILD_DIR=/var/lib/jenkins/workspace/build/custom_test_artifacts 2025-12-04T09:05:28.6790660Z + [[ -n '' ]] 2025-12-04T09:05:28.6790950Z + echo 'Environment variables' 2025-12-04T09:05:28.6791208Z Environment variables 2025-12-04T09:05:28.6791421Z + env 2025-12-04T09:05:28.6798763Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:05:28.6799425Z CONTINUE_THROUGH_ERROR=True 2025-12-04T09:05:28.6799909Z BUILD_ENVIRONMENT=linux-jammy-cuda12.8-py3.10-gcc11 2025-12-04T09:05:28.6800760Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-12-04T09:05:28.6801225Z HOSTNAME=364ec7eaeaf9 2025-12-04T09:05:28.6801766Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_4c1fca7f-6411-4e46-b998-dde57198cc6c 2025-12-04T09:05:28.6802337Z GITHUB_ACTION=__run_3 2025-12-04T09:05:28.6802573Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-12-04T09:05:28.6802979Z GITHUB_RUN_NUMBER=158165 2025-12-04T09:05:28.6803300Z TEST_CONFIG=default 2025-12-04T09:05:28.6803689Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-12-04T09:05:28.6803975Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-12-04T09:05:28.6804258Z SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:05:28.6804887Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-12-04T09:05:28.6805159Z GITHUB_TRIGGERING_ACTOR=huydhn 2025-12-04T09:05:28.6805393Z GITHUB_REF_TYPE=branch 2025-12-04T09:05:28.6805608Z TORCH_CUDA_ARCH_LIST=8.9 2025-12-04T09:05:28.6805877Z BASE_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:05:28.6806163Z XLA_CUDA= 2025-12-04T09:05:28.6806358Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-12-04T09:05:28.6806846Z HUGGING_FACE_HUB_TOKEN=*** 2025-12-04T09:05:28.6807150Z *** 2025-12-04T09:05:28.6807342Z GITHUB_REPOSITORY_ID=65600975 2025-12-04T09:05:28.6807547Z GITHUB_ACTIONS=true 2025-12-04T09:05:28.6807724Z NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:05:28.6807955Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:05:28.6808235Z SHA1=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:05:28.6808632Z GITHUB_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:05:28.6808997Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/trunk.yml@refs/heads/main 2025-12-04T09:05:28.6809429Z UCC_HOME=/usr 2025-12-04T09:05:28.6809601Z TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:05:28.6809795Z VERBOSE_TEST_LOGS=False 2025-12-04T09:05:28.6809973Z GITHUB_REF=refs/heads/main 2025-12-04T09:05:28.6810240Z SHARD_NUMBER=4 2025-12-04T09:05:28.6810490Z GITHUB_REF_PROTECTED=true 2025-12-04T09:05:28.6810683Z HOME=/var/lib/jenkins 2025-12-04T09:05:28.6810894Z GITHUB_API_URL=https://api.github.com 2025-12-04T09:05:28.6811213Z PYTORCH_TEST_RERUN_DISABLED_TESTS=1 2025-12-04T09:05:28.6811478Z UCX_COMMIT=7836b165abdbe468a2f607e7254011c07d788152 2025-12-04T09:05:28.6811720Z USE_SYSTEM_NCCL=1 2025-12-04T09:05:28.6811882Z NUM_TEST_SHARDS=5 2025-12-04T09:05:28.6812027Z UCX_HOME=/usr 2025-12-04T09:05:28.6812436Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_4c1fca7f-6411-4e46-b998-dde57198cc6c 2025-12-04T09:05:28.6813664Z JOB_NAME=linux-jammy-cuda12.8-py3.10-gcc11 / test (default, 4, 5, lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests) 2025-12-04T09:05:28.6814373Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_4c1fca7f-6411-4e46-b998-dde57198cc6c 2025-12-04T09:05:28.6814944Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-12-04T09:05:28.6815313Z GITHUB_EVENT_NAME=schedule 2025-12-04T09:05:28.6815499Z DASHBOARD_TAG= 2025-12-04T09:05:28.6815655Z GITHUB_RUN_ID=19922768520 2025-12-04T09:05:28.6815874Z INSTALLED_OPENBLAS= 2025-12-04T09:05:28.6816315Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_4c1fca7f-6411-4e46-b998-dde57198cc6c 2025-12-04T09:05:28.6816783Z GITHUB_ACTOR=huydhn 2025-12-04T09:05:28.6816934Z PR_NUMBER= 2025-12-04T09:05:28.6817080Z DESIRED_CUDA=12.8.1 2025-12-04T09:05:28.6817244Z GITHUB_RUN_ATTEMPT=1 2025-12-04T09:05:28.6817407Z VALGRIND=ON 2025-12-04T09:05:28.6817565Z ANACONDA_PYTHON_VERSION=3.10 2025-12-04T09:05:28.6817802Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-12-04T09:05:28.6818044Z TERM=vt100 2025-12-04T09:05:28.6818185Z INSTALLED_VISION=yes 2025-12-04T09:05:28.6818346Z BRANCH=main 2025-12-04T09:05:28.6818504Z SCCACHE_REGION=us-east-1 2025-12-04T09:05:28.6818694Z OPENSSL_ROOT_DIR=/opt/openssl 2025-12-04T09:05:28.6818884Z BUILD_AOT_INDUCTOR_TEST= 2025-12-04T09:05:28.6819066Z CUDA_PATH=/usr/local/cuda 2025-12-04T09:05:28.6819430Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-12-04T09:05:28.6819832Z GITHUB_SERVER_URL=https://github.com 2025-12-04T09:05:28.6820083Z UCC_COMMIT=430e241bf5d38cbc73fc7a6b89155397232e3f96 2025-12-04T09:05:28.6820318Z REENABLED_ISSUES= 2025-12-04T09:05:28.6820467Z DOCS= 2025-12-04T09:05:28.6820604Z SHLVL=1 2025-12-04T09:05:28.6820741Z MAX_JOBS=14 2025-12-04T09:05:28.6820884Z GITHUB_ACTOR_ID=475357 2025-12-04T09:05:28.6821136Z GITHUB_WORKFLOW_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:05:28.6821415Z GITHUB_REF_NAME=main 2025-12-04T09:05:28.6821678Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-12-04T09:05:28.6821980Z GITHUB_JOB=test 2025-12-04T09:05:28.6822145Z NO_TEST_TIMEOUT=False 2025-12-04T09:05:28.6822315Z TD_DISTRIBUTED=False 2025-12-04T09:05:28.6822502Z GITHUB_REPOSITORY=pytorch/pytorch 2025-12-04T09:05:28.6822711Z GITHUB_RETENTION_DAYS=90 2025-12-04T09:05:28.6822896Z OPENSSL_DIR=/opt/openssl 2025-12-04T09:05:28.6823077Z GITHUB_ACTION_REPOSITORY= 2025-12-04T09:05:28.6823643Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:05:28.6824201Z GITHUB_BASE_REF= 2025-12-04T09:05:28.6824358Z INSTALLED_ACL= 2025-12-04T09:05:28.6824693Z ARTIFACTS_FILE_SUFFIX=test-default-4-5-lf.linux.g6.4xlarge.experimental.nvidia.gpu_57116084869 2025-12-04T09:05:28.6825068Z CI=true 2025-12-04T09:05:28.6825219Z GITHUB_REPOSITORY_OWNER=pytorch 2025-12-04T09:05:28.6825451Z RUST_LOG=sccache::server=error 2025-12-04T09:05:28.6825639Z JOB_ID=57116084869 2025-12-04T09:05:28.6825790Z GITHUB_HEAD_REF= 2025-12-04T09:05:28.6825942Z GITHUB_ACTION_REF= 2025-12-04T09:05:28.6826140Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-12-04T09:05:28.6826382Z TEST_SHOWLOCALS=False 2025-12-04T09:05:28.6826546Z GITHUB_WORKFLOW=trunk 2025-12-04T09:05:28.6826732Z DEBIAN_FRONTEND=noninteractive 2025-12-04T09:05:28.6827176Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_4c1fca7f-6411-4e46-b998-dde57198cc6c 2025-12-04T09:05:28.6827615Z NO_TD=False 2025-12-04T09:05:28.6827777Z SKIP_SCCACHE_INITIALIZATION=1 2025-12-04T09:05:28.6827991Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-12-04T09:05:28.6828299Z OLDPWD=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda 2025-12-04T09:05:28.6828590Z _=/usr/bin/env 2025-12-04T09:05:28.6828753Z + echo 'Testing pytorch' 2025-12-04T09:05:28.6829008Z Testing pytorch 2025-12-04T09:05:28.6829297Z + export LANG=C.UTF-8 2025-12-04T09:05:28.6829472Z + LANG=C.UTF-8 2025-12-04T09:05:28.6829701Z + PR_NUMBER= 2025-12-04T09:05:28.6829865Z + [[ default == \d\e\f\a\u\l\t ]] 2025-12-04T09:05:28.6830074Z + export CUDA_VISIBLE_DEVICES=0 2025-12-04T09:05:28.6830269Z + CUDA_VISIBLE_DEVICES=0 2025-12-04T09:05:28.6830448Z + export HIP_VISIBLE_DEVICES=0 2025-12-04T09:05:28.6830645Z + HIP_VISIBLE_DEVICES=0 2025-12-04T09:05:28.6830831Z + [[ default == \d\i\s\t\r\i\b\u\t\e\d ]] 2025-12-04T09:05:28.6831038Z + [[ default == \s\l\o\w ]] 2025-12-04T09:05:28.6831285Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *slow-gradcheck* ]] 2025-12-04T09:05:28.6831593Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *cuda* ]] 2025-12-04T09:05:28.6831856Z + export PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-12-04T09:05:28.6832098Z + PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-12-04T09:05:28.6832317Z + [[ default == *crossref* ]] 2025-12-04T09:05:28.6832539Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *rocm* ]] 2025-12-04T09:05:28.6832814Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *xpu* ]] 2025-12-04T09:05:28.6833094Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 != *-bazel-* ]] 2025-12-04T09:05:28.6833349Z + pip_install ninja==1.10.2 2025-12-04T09:05:28.6833718Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-12-04T09:05:28.6834065Z + python3 -m pip install --progress-bar off ninja==1.10.2 2025-12-04T09:05:29.4680113Z Collecting ninja==1.10.2 2025-12-04T09:05:29.4897312Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl.metadata (5.0 kB) 2025-12-04T09:05:29.5292959Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl (108 kB) 2025-12-04T09:05:29.8781033Z Installing collected packages: ninja 2025-12-04T09:05:29.8781549Z Attempting uninstall: ninja 2025-12-04T09:05:29.8787718Z Found existing installation: ninja 1.11.1.4 2025-12-04T09:05:29.8810637Z Uninstalling ninja-1.11.1.4: 2025-12-04T09:05:29.8964581Z Successfully uninstalled ninja-1.11.1.4 2025-12-04T09:05:29.9444948Z Successfully installed ninja-1.10.2 2025-12-04T09:05:29.9884617Z + export PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:05:29.9886111Z + PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:05:29.9886983Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *aarch64* ]] 2025-12-04T09:05:29.9887354Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *asan* ]] 2025-12-04T09:05:29.9887725Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *-debug* ]] 2025-12-04T09:05:29.9888100Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 != *-bazel-* ]] 2025-12-04T09:05:29.9888536Z + echo 'We are not in debug mode: linux-jammy-cuda12.8-py3.10-gcc11. Expect the assertion to pass' 2025-12-04T09:05:29.9889163Z We are not in debug mode: linux-jammy-cuda12.8-py3.10-gcc11. Expect the assertion to pass 2025-12-04T09:05:29.9889524Z + cd test 2025-12-04T09:05:29.9889781Z + python -c 'import torch; torch._C._crash_if_debug_asserts_fail(424242)' 2025-12-04T09:05:31.4003391Z + [[ default == \n\o\g\p\u\_\N\O\_\A\V\X\2 ]] 2025-12-04T09:05:31.4003756Z + [[ default == \n\o\g\p\u\_\A\V\X\5\1\2 ]] 2025-12-04T09:05:31.4004093Z + [[ default == \l\e\g\a\c\y\_\n\v\i\d\i\a\_\d\r\i\v\e\r ]] 2025-12-04T09:05:31.4009031Z + DYNAMO_BENCHMARK_FLAGS=() 2025-12-04T09:05:31.4009557Z + [[ default == *pr_time_benchmarks* ]] 2025-12-04T09:05:31.4009923Z + [[ default == *dynamo_eager* ]] 2025-12-04T09:05:31.4010197Z + [[ default == *aot_eager* ]] 2025-12-04T09:05:31.4010565Z + [[ default == *aot_inductor* ]] 2025-12-04T09:05:31.4010855Z + [[ default == *max_autotune_inductor* ]] 2025-12-04T09:05:31.4011147Z + [[ default == *inductor* ]] 2025-12-04T09:05:31.4011387Z + [[ default == *dynamic* ]] 2025-12-04T09:05:31.4012029Z + [[ default == *cpu* ]] 2025-12-04T09:05:31.4012260Z + [[ default == *xpu* ]] 2025-12-04T09:05:31.4012521Z + DYNAMO_BENCHMARK_FLAGS+=(--device cuda) 2025-12-04T09:05:31.4030623Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *libtorch* ]] 2025-12-04T09:05:31.4031076Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *-bazel-* ]] 2025-12-04T09:05:31.4033446Z + cd test 2025-12-04T09:05:31.4034131Z + python -c 'import torch; print(torch.__config__.show())' 2025-12-04T09:05:32.8335513Z PyTorch built with: 2025-12-04T09:05:32.8335815Z - GCC 11.4 2025-12-04T09:05:32.8336026Z - C++ Version: 201703 2025-12-04T09:05:32.8336552Z - Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-12-04T09:05:32.8337239Z - Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-12-04T09:05:32.8337649Z - OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-12-04T09:05:32.8337954Z - LAPACK is enabled (usually provided by MKL) 2025-12-04T09:05:32.8338279Z - NNPACK is enabled 2025-12-04T09:05:32.8338521Z - CPU capability usage: AVX2 2025-12-04T09:05:32.8338777Z - CUDA Runtime 12.8 2025-12-04T09:05:32.8339151Z - NVCC architecture flags: -gencode;arch=compute_75,code=sm_75;-gencode;arch=compute_89,code=sm_89 2025-12-04T09:05:32.8339552Z - CuDNN 91.0.2 (built against CUDA 12.9) 2025-12-04T09:05:32.8342971Z - Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, COMMIT_SHA=35b7a9a26c5923d98aebaa41a031dae21788a9ee, CUDA_VERSION=12.8, CUDNN_VERSION=9.10.2, CXX_COMPILER=/opt/cache/bin/c++, CXX_FLAGS= -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOROCTRACER -DLIBKINETO_NOXPUPTI=ON -DUSE_FBGEMM -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -DC10_NODEPRECATED -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=range-loop-construct -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-unknown-pragmas -Wno-unused-parameter -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=old-style-cast -faligned-new -Werror -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, FORCE_FALLBACK_CUDA_MPI=1, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, TORCH_VERSION=2.10.0, USE_CUDA=ON, USE_CUDNN=ON, USE_CUSPARSELT=ON, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_GLOO=ON, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=ON, USE_NCCL=ON, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, USE_ROCM_KERNEL_ASSERT=OFF, USE_XCCL=OFF, USE_XPU=OFF, 2025-12-04T09:05:32.8346446Z 2025-12-04T09:05:33.0984295Z + cd test 2025-12-04T09:05:33.0984677Z + python -c 'import torch; print(torch.__config__.parallel_info())' 2025-12-04T09:05:34.2498712Z ATen/Parallel: 2025-12-04T09:05:34.2499043Z at::get_num_threads() : 8 2025-12-04T09:05:34.2499319Z at::get_num_interop_threads() : 8 2025-12-04T09:05:34.2499595Z OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-12-04T09:05:34.2499864Z omp_get_max_threads() : 8 2025-12-04T09:05:34.2500393Z Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-12-04T09:05:34.2500919Z mkl_get_max_threads() : 8 2025-12-04T09:05:34.2501261Z Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-12-04T09:05:34.2501662Z std::thread::hardware_concurrency() : 16 2025-12-04T09:05:34.2501937Z Environment variables: 2025-12-04T09:05:34.2502164Z OMP_NUM_THREADS : [not set] 2025-12-04T09:05:34.2502403Z MKL_NUM_THREADS : [not set] 2025-12-04T09:05:34.2502648Z ATen parallel backend: OpenMP 2025-12-04T09:05:34.2502812Z 2025-12-04T09:05:34.4839461Z + [[ default == *numpy_2* ]] 2025-12-04T09:05:34.4840030Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *aarch64* ]] 2025-12-04T09:05:34.4840596Z + [[ default == *backward* ]] 2025-12-04T09:05:34.4840998Z + [[ default == *libtorch_agnostic_targetting* ]] 2025-12-04T09:05:34.4841394Z + [[ default == *xla* ]] 2025-12-04T09:05:34.4841736Z + [[ default == *vllm* ]] 2025-12-04T09:05:34.4841965Z + [[ default == *executorch* ]] 2025-12-04T09:05:34.4842649Z + [[ default == \j\i\t\_\l\e\g\a\c\y ]] 2025-12-04T09:05:34.4843110Z + [[ default == \q\u\a\n\t\i\z\a\t\i\o\n ]] 2025-12-04T09:05:34.4843447Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *libtorch* ]] 2025-12-04T09:05:34.4843783Z + [[ default == distributed ]] 2025-12-04T09:05:34.4844042Z + [[ default == *operator_benchmark* ]] 2025-12-04T09:05:34.4844336Z + [[ default == *operator_microbenchmark* ]] 2025-12-04T09:05:34.4844666Z + [[ default == *attention_microbenchmark* ]] 2025-12-04T09:05:34.4844965Z + [[ default == *inductor_distributed* ]] 2025-12-04T09:05:34.4845257Z + [[ default == *inductor-halide* ]] 2025-12-04T09:05:34.4845536Z + [[ default == *inductor-pallas* ]] 2025-12-04T09:05:34.4846024Z + [[ default == *inductor-triton-cpu* ]] 2025-12-04T09:05:34.4846378Z + [[ default == *inductor-micro-benchmark* ]] 2025-12-04T09:05:34.4846697Z + [[ default == *aoti_cross_compile_for_windows* ]] 2025-12-04T09:05:34.4846999Z + [[ default == *huggingface* ]] 2025-12-04T09:05:34.4847255Z + [[ default == *timm* ]] 2025-12-04T09:05:34.4847480Z + [[ default == cachebench ]] 2025-12-04T09:05:34.4847721Z + [[ default == verify_cachebench ]] 2025-12-04T09:05:34.4847976Z + [[ default == *torchbench* ]] 2025-12-04T09:05:34.4848235Z + [[ default == *inductor_cpp_wrapper* ]] 2025-12-04T09:05:34.4848501Z + [[ default == *inductor_core* ]] 2025-12-04T09:05:34.4848750Z + [[ default == *inductor* ]] 2025-12-04T09:05:34.4848981Z + [[ default == *einops* ]] 2025-12-04T09:05:34.4849203Z + [[ default == *dynamo_core* ]] 2025-12-04T09:05:34.4849437Z + [[ default == *dynamo_wrapped* ]] 2025-12-04T09:05:34.4849674Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *rocm* ]] 2025-12-04T09:05:34.4849914Z + [[ 4 == 1 ]] 2025-12-04T09:05:34.4850056Z + [[ 4 == 2 ]] 2025-12-04T09:05:34.4850207Z + [[ 4 -gt 2 ]] 2025-12-04T09:05:34.4850369Z + install_torchvision 2025-12-04T09:05:34.4850545Z + local orig_preload 2025-12-04T09:05:34.4850710Z + local commit 2025-12-04T09:05:34.4850872Z ++ get_pinned_commit vision 2025-12-04T09:05:34.4851069Z ++ cat .github/ci_commit_pins/vision.txt 2025-12-04T09:05:34.4861803Z + commit=617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:05:34.4862140Z + orig_preload= 2025-12-04T09:05:34.4862341Z + '[' -n '' ']' 2025-12-04T09:05:34.4862689Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 == *cuda* ]] 2025-12-04T09:05:34.4863149Z + export FORCE_CUDA=1 2025-12-04T09:05:34.4863365Z + FORCE_CUDA=1 2025-12-04T09:05:34.4863658Z + export WITH_CUDA=1 2025-12-04T09:05:34.4863894Z + WITH_CUDA=1 2025-12-04T09:05:34.4864546Z + pip_build_and_install git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e dist/vision 2025-12-04T09:05:34.4865358Z + local build_target=git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:05:34.4865875Z + local wheel_dir=dist/vision 2025-12-04T09:05:34.4866118Z + local found_whl=0 2025-12-04T09:05:34.4866330Z + for file in "${wheel_dir}"/*.whl 2025-12-04T09:05:34.4866592Z + [[ -f dist/vision/*.whl ]] 2025-12-04T09:05:34.4866832Z + '[' 0 == 0 ']' 2025-12-04T09:05:34.4867463Z + python3 -m pip wheel --no-build-isolation --no-deps -w dist/vision git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:05:34.7660914Z Collecting git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:05:34.7664805Z Cloning https://github.com/pytorch/vision.git (to revision 617079d944b0e72632311c30ae2bbdf1168b901e) to /tmp/pip-req-build-g3xdi6sf 2025-12-04T09:05:34.7686507Z Running command git clone --filter=blob:none --quiet https://github.com/pytorch/vision.git /tmp/pip-req-build-g3xdi6sf 2025-12-04T09:05:36.2177354Z Running command git rev-parse -q --verify 'sha^617079d944b0e72632311c30ae2bbdf1168b901e' 2025-12-04T09:05:36.2206558Z Running command git fetch -q https://github.com/pytorch/vision.git 617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:05:36.3217392Z Resolved https://github.com/pytorch/vision.git to commit 617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:05:38.1244312Z Preparing metadata (pyproject.toml) ... [?25l- \ | done 2025-12-04T09:05:38.1277656Z [?25hBuilding wheels for collected packages: torchvision 2025-12-04T09:06:51.1134193Z Building wheel for torchvision (pyproject.toml) ... [?25l- \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ done 2025-12-04T09:06:51.1162566Z [?25h Created wheel for torchvision: filename=torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl size=1786228 sha256=cd9c005e268df9ce6928b66b8245668616c178ca8e58f0afdf18c2edabad7093 2025-12-04T09:06:51.1165029Z Stored in directory: /var/lib/jenkins/.cache/pip/wheels/12/b2/29/1f82685c5b5173629e1f36a9b93989ce92ce563e5fb91d27ac 2025-12-04T09:06:51.1196342Z Successfully built torchvision 2025-12-04T09:06:51.2150311Z + for file in "${wheel_dir}"/*.whl 2025-12-04T09:06:51.2151192Z + pip_install_whl dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:06:51.2151900Z + args=('dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl') 2025-12-04T09:06:51.2152304Z + local args 2025-12-04T09:06:51.2152663Z + [[ dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl == *\ * ]] 2025-12-04T09:06:51.2153089Z + for path in "${args[@]}" 2025-12-04T09:06:51.2153511Z + echo 'Installing dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl' 2025-12-04T09:06:51.2154144Z Installing dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:06:51.2154866Z + python3 -mpip install --no-index --no-deps dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:06:51.5040903Z Processing ./dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:06:51.5122406Z Installing collected packages: torchvision 2025-12-04T09:06:51.9253607Z Successfully installed torchvision-0.25.0a0+617079d 2025-12-04T09:06:51.9533401Z + '[' -n '' ']' 2025-12-04T09:06:51.9533671Z + test_python_shard 4 2025-12-04T09:06:51.9533904Z + [[ -z 5 ]] 2025-12-04T09:06:51.9534643Z + python test/run_test.py --exclude-jit-executor --exclude-distributed-tests --exclude-quantization-tests --shard 4 5 --verbose --upload-artifacts-while-running 2025-12-04T09:06:56.2537169Z Downloading https://ossci-metrics.s3.amazonaws.com/disabled-tests-condensed.json to /var/lib/jenkins/workspace/test/.pytorch-disabled-tests.json 2025-12-04T09:06:56.2994099Z Ignoring disabled issues: [''] 2025-12-04T09:06:56.3072972Z Found test times from artifacts 2025-12-04T09:06:56.3389862Z Found test times from artifacts 2025-12-04T09:06:56.3398918Z Running all tests 2025-12-04T09:06:56.3916527Z Running parallel tests on 3 processes 2025-12-04T09:06:56.3922998Z Name: tests to run (est. time: 102.78min) 2025-12-04T09:06:56.3923318Z Serial tests (0): 2025-12-04T09:06:56.3923558Z Parallel tests (113): 2025-12-04T09:06:56.3923851Z inductor/test_aot_inductor 3/4 2025-12-04T09:06:56.3924175Z dynamo/test_model_output 1/1 2025-12-04T09:06:56.3924475Z inductor/test_extension_backend 1/1 2025-12-04T09:06:56.3924784Z inductor/test_cooperative_reductions 1/1 2025-12-04T09:06:56.3925075Z dynamo/test_fx_graph_runnable 1/1 2025-12-04T09:06:56.3925343Z dynamo/test_backends 1/1 2025-12-04T09:06:56.3925590Z dynamo/test_functions 1/1 2025-12-04T09:06:56.3925844Z inductor/test_mix_order_reduction 1/1 2025-12-04T09:06:56.3926117Z inductor/test_padding 1/1 2025-12-04T09:06:56.3926351Z dynamo/test_dicts 1/1 2025-12-04T09:06:56.3926586Z dynamo/test_aot_compile 1/1 2025-12-04T09:06:56.3926816Z dynamo/test_sets 1/1 2025-12-04T09:06:56.3927049Z dynamo/test_callback 1/1 2025-12-04T09:06:56.3927352Z inductor/test_cudagraph_trees_expandable_segments 1/1 2025-12-04T09:06:56.3927632Z dynamo/test_exceptions 1/1 2025-12-04T09:06:56.3927816Z test_cuda 1/1 2025-12-04T09:06:56.3927983Z test_transformers 1/1 2025-12-04T09:06:56.3928493Z higher_order_ops/test_local_map 1/1 2025-12-04T09:06:56.3928720Z test_dataloader 1/1 2025-12-04T09:06:56.3929068Z test_decomp 6/22 2025-12-04T09:06:56.3929248Z test_decomp 8/22 2025-12-04T09:06:56.3929414Z test_decomp 12/22 2025-12-04T09:06:56.3929586Z test_decomp 18/22 2025-12-04T09:06:56.3929751Z test_decomp 20/22 2025-12-04T09:06:56.3929911Z test_ops 2/9 2025-12-04T09:06:56.3930069Z test_ops 8/9 2025-12-04T09:06:56.3930225Z functorch/test_dims 1/1 2025-12-04T09:06:56.3930423Z test_torchfuzz_repros 1/1 2025-12-04T09:06:56.3930629Z inductor/test_torchinductor 1/2 2025-12-04T09:06:56.3930840Z inductor/test_torchinductor 2/2 2025-12-04T09:06:56.3931056Z inductor/test_kernel_benchmark 1/1 2025-12-04T09:06:56.3931291Z inductor/test_torchinductor_opinfo 1/13 2025-12-04T09:06:56.3931532Z inductor/test_torchinductor_opinfo 5/13 2025-12-04T09:06:56.3931755Z inductor/test_torchinductor_opinfo 7/13 2025-12-04T09:06:56.3931994Z inductor/test_torchinductor_opinfo 13/13 2025-12-04T09:06:56.3932224Z inductor/test_layout_optim 1/1 2025-12-04T09:06:56.3932438Z inductor/test_aot_inductor_arrayref 2/2 2025-12-04T09:06:56.3932686Z inductor/test_torchinductor_strided_blocks 1/1 2025-12-04T09:06:56.3932929Z test_custom_ops 1/1 2025-12-04T09:06:56.3933114Z test_content_store 1/1 2025-12-04T09:06:56.3933437Z inductor/test_flex_decoding 1/3 2025-12-04T09:06:56.3933655Z inductor/test_flex_decoding 3/3 2025-12-04T09:06:56.3933867Z inductor/test_deterministic 3/3 2025-12-04T09:06:56.3934082Z inductor/test_b2b_gemm 1/1 2025-12-04T09:06:56.3934275Z export/test_tree_utils 1/1 2025-12-04T09:06:56.3934479Z inductor/test_triton_wrapper 1/1 2025-12-04T09:06:56.3934706Z inductor/test_static_cuda_launcher 1/1 2025-12-04T09:06:56.3934929Z export/test_dynamic_shapes 1/1 2025-12-04T09:06:56.3935135Z dynamo/test_sdpa 1/1 2025-12-04T09:06:56.3935334Z inductor/test_aot_inductor_package 1/1 2025-12-04T09:06:56.3935566Z inductor/test_compiled_optimizers 3/3 2025-12-04T09:06:56.3935798Z inductor/test_aot_inductor_utils 1/1 2025-12-04T09:06:56.3936026Z inductor/test_control_flow 3/5 2025-12-04T09:06:56.3936224Z test_mkl_verbose 1/1 2025-12-04T09:06:56.3936412Z test_comparison_utils 1/1 2025-12-04T09:06:56.3936619Z functorch/test_ac_logging 1/1 2025-12-04T09:06:56.3936823Z test_mkldnn_verbose 1/1 2025-12-04T09:06:56.3937004Z test_utils_config_module 1/1 2025-12-04T09:06:56.3937199Z test_hop_infra 1/1 2025-12-04T09:06:56.3937383Z test_appending_byte_serializer 1/1 2025-12-04T09:06:56.3937587Z test_license 1/1 2025-12-04T09:06:56.3937757Z test_ao_sparsity 1/1 2025-12-04T09:06:56.3937934Z test_autoload 1/1 2025-12-04T09:06:56.3938112Z nn/attention/test_open_registry 1/1 2025-12-04T09:06:56.3938345Z test_as_strided 1/1 2025-12-04T09:06:56.3938513Z test_foreach 1/1 2025-12-04T09:06:56.3938672Z xpu/test_gemm 1/1 2025-12-04T09:06:56.3938849Z higher_order_ops/test_print 1/1 2025-12-04T09:06:56.3939065Z test_per_overload_api 1/1 2025-12-04T09:06:56.3939279Z torch_np/numpy_tests/core/test_einsum 1/1 2025-12-04T09:06:56.3939506Z test_out_dtype_op 1/1 2025-12-04T09:06:56.3939693Z torch_np/test_ufuncs_basic 1/1 2025-12-04T09:06:56.3939896Z lazy/test_step_closures 1/1 2025-12-04T09:06:56.3940101Z functorch/dim/test_getsetitem 1/1 2025-12-04T09:06:56.3940316Z test_ops_fwd_gradients 1/3 2025-12-04T09:06:56.3940512Z test_ops_fwd_gradients 2/3 2025-12-04T09:06:56.3940688Z test_meta 4/4 2025-12-04T09:06:56.3940849Z test_nestedtensor 3/4 2025-12-04T09:06:56.3941029Z test_nestedtensor 4/4 2025-12-04T09:06:56.3941195Z test_modules 2/2 2025-12-04T09:06:56.3941375Z test_tensorboard 1/1 2025-12-04T09:06:56.3941579Z torch_np/numpy_tests/core/test_indexing 1/1 2025-12-04T09:06:56.3941799Z test_futures 1/1 2025-12-04T09:06:56.3941962Z nn/test_dropout 1/1 2025-12-04T09:06:56.3942146Z functorch/dim/test_split 1/1 2025-12-04T09:06:56.3942470Z torch_np/numpy_tests/lib/test_type_check 1/1 2025-12-04T09:06:56.3942798Z cpp_extensions/test_libtorch_agnostic 1/1 2025-12-04T09:06:56.3943045Z profiler/test_execution_trace 1/1 2025-12-04T09:06:56.3943258Z test_jit 1/1 2025-12-04T09:06:56.3943422Z test_datapipe 1/1 2025-12-04T09:06:56.3943605Z test_numba_integration 1/1 2025-12-04T09:06:56.3943798Z test_functional_optim 1/1 2025-12-04T09:06:56.3943994Z test_maskedtensor 1/1 2025-12-04T09:06:56.3944200Z benchmark_utils/test_benchmark_utils 1/1 2025-12-04T09:06:56.3944454Z torch_np/numpy_tests/core/test_scalarmath 1/1 2025-12-04T09:06:56.3944691Z test_scaled_matmul_cuda 1/1 2025-12-04T09:06:56.3944920Z torch_np/numpy_tests/core/test_shape_base 1/1 2025-12-04T09:06:56.3945148Z test_vulkan 1/1 2025-12-04T09:06:56.3945312Z lazy/test_generator 1/1 2025-12-04T09:06:56.3945526Z torch_np/numpy_tests/linalg/test_linalg 1/1 2025-12-04T09:06:56.3945772Z torch_np/numpy_tests/core/test_dtype 1/1 2025-12-04T09:06:56.3945993Z lazy/test_debug_util 1/1 2025-12-04T09:06:56.3946185Z nn/test_load_state_dict 1/1 2025-12-04T09:06:56.3946377Z test_shape_ops 1/1 2025-12-04T09:06:56.3946545Z nn/test_module_hooks 1/1 2025-12-04T09:06:56.3946752Z torch_np/numpy_tests/lib/test_twodim_base 1/1 2025-12-04T09:06:56.3946998Z profiler/test_memory_profiler 1/1 2025-12-04T09:06:56.3947212Z test_jit_llga_fuser 1/1 2025-12-04T09:06:56.3947400Z test_serialization 1/1 2025-12-04T09:06:56.3947585Z test_sparse_csr 1/3 2025-12-04T09:06:56.3947760Z functorch/test_ops 6/9 2025-12-04T09:06:56.3947932Z functorch/test_ops 8/9 2025-12-04T09:06:56.3948102Z test_mkldnn 1/1 2025-12-04T09:06:56.3948274Z Name: excluded (est. time: 0.0min) 2025-12-04T09:06:56.3948470Z Serial tests (0): 2025-12-04T09:06:56.3948634Z Parallel tests (0): 2025-12-04T09:06:56.3948942Z Running inductor/test_aot_inductor 3/4 ... [2025-12-04 09:06:56.394032][852.353579769] 2025-12-04T09:06:56.3949277Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:06:56.3950197Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_aot_inductor.py', '-m', 'serial', '--shard-id=3', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:06:56.394352] 2025-12-04T09:07:04.8527878Z 2025-12-04T09:07:04.8528781Z inductor/test_aot_inductor 3/4 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_aot_inductor_3.4_4414340a188212de_.log 2025-12-04T09:07:04.8529510Z Running 0 items in this shard: 2025-12-04T09:07:04.8529696Z 2025-12-04T09:07:04.8529993Z Finished inductor/test_aot_inductor 3/4 ... [2025-12-04 09:07:04.852694][860.812240305], took 0.14min 2025-12-04T09:07:04.8535781Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-78423a5d6c0345b8.xml 2025-12-04T09:07:05.2291796Z Uploading artifacts took 0.11 seconds 2025-12-04T09:07:05.2295484Z Running dynamo/test_model_output 1/1 ... [2025-12-04 09:07:05.229315][861.188859892] 2025-12-04T09:07:05.2296035Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:07:05.2299033Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_model_output.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:07:05.229628] 2025-12-04T09:07:08.6140491Z 2025-12-04T09:07:08.6141343Z dynamo/test_model_output 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_model_output_1.1_4617a3ec054d38e3_.log 2025-12-04T09:07:08.6142102Z Running 0 items in this shard: 2025-12-04T09:07:08.6142276Z 2025-12-04T09:07:08.6142552Z Finished dynamo/test_model_output 1/1 ... [2025-12-04 09:07:08.613830][864.57337244], took 0.06min 2025-12-04T09:07:08.6148795Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_model_output/dynamo.test_model_output-5ac7859c2c204934.xml 2025-12-04T09:07:08.6429421Z Running inductor/test_extension_backend 1/1 ... [2025-12-04 09:07:08.642700][864.602247246] 2025-12-04T09:07:08.6429900Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:07:08.6433383Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_extension_backend.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:07:08.643053] 2025-12-04T09:07:14.3744968Z 2025-12-04T09:07:14.3745855Z inductor/test_extension_backend 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_extension_backend_1.1_31c714ecf5a69f35_.log 2025-12-04T09:07:14.3746662Z Running 0 items in this shard: 2025-12-04T09:07:14.3746830Z 2025-12-04T09:07:14.3747145Z Finished inductor/test_extension_backend 1/1 ... [2025-12-04 09:07:14.374249][870.33379439], took 0.10min 2025-12-04T09:07:14.3753407Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_extension_backend/inductor.test_extension_backend-4a93539bbcae0552.xml 2025-12-04T09:07:14.4040668Z Running inductor/test_cooperative_reductions 1/1 ... [2025-12-04 09:07:14.403850][870.363398925] 2025-12-04T09:07:14.4041158Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:07:14.4044051Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_cooperative_reductions.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:07:14.404150] 2025-12-04T09:07:19.8251410Z 2025-12-04T09:07:19.8252842Z inductor/test_cooperative_reductions 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_cooperative_reductions_1.1_2b00296025758482_.log 2025-12-04T09:07:19.8254397Z Running 0 items in this shard: 2025-12-04T09:07:19.8254671Z 2025-12-04T09:07:19.8255214Z Finished inductor/test_cooperative_reductions 1/1 ... [2025-12-04 09:07:19.824893][875.784436632], took 0.09min 2025-12-04T09:07:19.8264349Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cooperative_reductions/inductor.test_cooperative_reductions-fc3209be19432af6.xml 2025-12-04T09:07:19.8502960Z Running dynamo/test_fx_graph_runnable 1/1 ... [2025-12-04 09:07:19.850073][875.809619235] 2025-12-04T09:07:19.8503421Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:07:19.8506628Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_fx_graph_runnable.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:07:19.850372] 2025-12-04T09:07:25.1930766Z 2025-12-04T09:07:25.1931678Z dynamo/test_fx_graph_runnable 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_fx_graph_runnable_1.1_d62e5b0a2f03d2b8_.log 2025-12-04T09:07:25.1932465Z Running 0 items in this shard: 2025-12-04T09:07:25.1932645Z 2025-12-04T09:07:25.1932969Z Finished dynamo/test_fx_graph_runnable 1/1 ... [2025-12-04 09:07:25.192817][881.152364106], took 0.09min 2025-12-04T09:07:25.1943777Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_fx_graph_runnable/dynamo.test_fx_graph_runnable-ddbc1ae1b738e2f4.xml 2025-12-04T09:07:25.2262340Z Running dynamo/test_backends 1/1 ... [2025-12-04 09:07:25.225983][881.185531573] 2025-12-04T09:07:25.2263150Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:07:25.2265447Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_backends.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:07:25.226257] 2025-12-04T09:07:30.6644945Z 2025-12-04T09:07:30.6645772Z dynamo/test_backends 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_backends_1.1_dee2ca4cb0571aff_.log 2025-12-04T09:07:30.6646470Z Running 0 items in this shard: 2025-12-04T09:07:30.6646651Z 2025-12-04T09:07:30.6646917Z Finished dynamo/test_backends 1/1 ... [2025-12-04 09:07:30.664259][886.623805618], took 0.09min 2025-12-04T09:07:30.6658936Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_backends/dynamo.test_backends-8659746a8e039725.xml 2025-12-04T09:07:30.6918180Z Running dynamo/test_functions 1/1 ... [2025-12-04 09:07:30.691611][886.651159924] 2025-12-04T09:07:30.6918566Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:07:30.6921551Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_functions.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:07:30.691877] 2025-12-04T09:07:36.6593874Z 2025-12-04T09:07:36.6594670Z dynamo/test_functions 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_functions_1.1_01cba0571253972c_.log 2025-12-04T09:07:36.6595375Z Running 0 items in this shard: 2025-12-04T09:07:36.6595554Z 2025-12-04T09:07:36.6595818Z Finished dynamo/test_functions 1/1 ... [2025-12-04 09:07:36.659157][892.618702883], took 0.10min 2025-12-04T09:07:36.6609119Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_functions/dynamo.test_functions-891767433e56c016.xml 2025-12-04T09:07:36.6859257Z Running inductor/test_mix_order_reduction 1/1 ... [2025-12-04 09:07:36.685692][892.645240479] 2025-12-04T09:07:36.6859744Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:07:36.6862086Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_mix_order_reduction.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:07:36.685957] 2025-12-04T09:07:42.4757940Z 2025-12-04T09:07:42.4758897Z inductor/test_mix_order_reduction 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_mix_order_reduction_1.1_b3c0a6fce4a8dfd0_.log 2025-12-04T09:07:42.4759718Z Running 0 items in this shard: 2025-12-04T09:07:42.4759898Z 2025-12-04T09:07:42.4760232Z Finished inductor/test_mix_order_reduction 1/1 ... [2025-12-04 09:07:42.475571][898.435116411], took 0.10min 2025-12-04T09:07:42.4779142Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-adb9352cbf30e883.xml 2025-12-04T09:07:42.5009319Z Running inductor/test_padding 1/1 ... [2025-12-04 09:07:42.500699][898.460247342] 2025-12-04T09:07:42.5009768Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:07:42.5012835Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_padding.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:07:42.500996] 2025-12-04T09:07:47.8113097Z 2025-12-04T09:07:47.8113843Z inductor/test_padding 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_padding_1.1_df8aa049a65f2ea4_.log 2025-12-04T09:07:47.8114808Z Running 0 items in this shard: 2025-12-04T09:07:47.8115152Z 2025-12-04T09:07:47.8115382Z Finished inductor/test_padding 1/1 ... [2025-12-04 09:07:47.811089][903.770635807], took 0.09min 2025-12-04T09:07:47.8131759Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_padding/inductor.test_padding-2d400f1d2457cdfd.xml 2025-12-04T09:07:47.8382426Z Running dynamo/test_dicts 1/1 ... [2025-12-04 09:07:47.838037][903.797586252] 2025-12-04T09:07:47.8382846Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:07:47.8385711Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_dicts.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:07:47.838327] 2025-12-04T09:07:50.9653430Z 2025-12-04T09:07:50.9654369Z dynamo/test_dicts 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_dicts_1.1_dabfcd9b1024dd1a_.log 2025-12-04T09:07:50.9655060Z Running 0 items in this shard: 2025-12-04T09:07:50.9655242Z 2025-12-04T09:07:50.9655496Z Finished dynamo/test_dicts 1/1 ... [2025-12-04 09:07:50.965135][906.924680029], took 0.05min 2025-12-04T09:07:50.9673495Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_dicts/dynamo.test_dicts-4379238afe4ffe3d.xml 2025-12-04T09:07:50.9913621Z Running dynamo/test_aot_compile 1/1 ... [2025-12-04 09:07:50.991119][906.95066733] 2025-12-04T09:07:50.9914062Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:07:50.9916889Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_aot_compile.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:07:50.991430] 2025-12-04T09:07:53.9746385Z 2025-12-04T09:07:53.9747246Z dynamo/test_aot_compile 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_aot_compile_1.1_52a2ed661618cde4_.log 2025-12-04T09:07:53.9747961Z Running 0 items in this shard: 2025-12-04T09:07:53.9748132Z 2025-12-04T09:07:53.9748407Z Finished dynamo/test_aot_compile 1/1 ... [2025-12-04 09:07:53.974420][909.933965919], took 0.05min 2025-12-04T09:07:53.9768247Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_aot_compile/dynamo.test_aot_compile-79eacfeb4f1e860c.xml 2025-12-04T09:07:54.0044765Z Running dynamo/test_sets 1/1 ... [2025-12-04 09:07:54.004245][909.963794173] 2025-12-04T09:07:54.0045182Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:07:54.0048462Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_sets.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:07:54.004559] 2025-12-04T09:07:57.1426889Z 2025-12-04T09:07:57.1427672Z dynamo/test_sets 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_sets_1.1_0a6d988f4dd8b837_.log 2025-12-04T09:07:57.1428343Z Running 0 items in this shard: 2025-12-04T09:07:57.1428517Z 2025-12-04T09:07:57.1428765Z Finished dynamo/test_sets 1/1 ... [2025-12-04 09:07:57.142467][913.102012899], took 0.05min 2025-12-04T09:07:57.1449866Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_sets/dynamo.test_sets-135ba238b392c31c.xml 2025-12-04T09:07:57.1723431Z Running dynamo/test_callback 1/1 ... [2025-12-04 09:07:57.172124][913.131672861] 2025-12-04T09:07:57.1723866Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:07:57.1727413Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_callback.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:07:57.172426] 2025-12-04T09:08:02.4409399Z 2025-12-04T09:08:02.4410247Z dynamo/test_callback 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_callback_1.1_3b05c69589f6a657_.log 2025-12-04T09:08:02.4410992Z Running 0 items in this shard: 2025-12-04T09:08:02.4411174Z 2025-12-04T09:08:02.4411434Z Finished dynamo/test_callback 1/1 ... [2025-12-04 09:08:02.440738][918.400283053], took 0.09min 2025-12-04T09:08:02.4433710Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_callback/dynamo.test_callback-d90a836cf7319d29.xml 2025-12-04T09:08:02.4697199Z Running inductor/test_cudagraph_trees_expandable_segments 1/1 ... [2025-12-04 09:08:02.469485][918.429033633] 2025-12-04T09:08:02.4697782Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:08:02.4700877Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_cudagraph_trees_expandable_segments.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:08:02.469791] 2025-12-04T09:08:07.9873793Z 2025-12-04T09:08:07.9874948Z inductor/test_cudagraph_trees_expandable_segments 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_cudagraph_trees_expandable_segments_1.1_261e345fe4595b61_.log 2025-12-04T09:08:07.9875909Z Running 0 items in this shard: 2025-12-04T09:08:07.9876081Z 2025-12-04T09:08:07.9876446Z Finished inductor/test_cudagraph_trees_expandable_segments 1/1 ... [2025-12-04 09:08:07.987137][923.946683361], took 0.09min 2025-12-04T09:08:07.9899469Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cudagraph_trees_expandable_segments/inductor.test_cudagraph_trees_expandable_segments-87db9d88f872c44a.xml 2025-12-04T09:08:08.0189730Z Running dynamo/test_exceptions 1/1 ... [2025-12-04 09:08:08.018734][923.978283029] 2025-12-04T09:08:08.0201015Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:08:08.0202263Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_exceptions.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:08:08.019011] 2025-12-04T09:08:11.0191667Z 2025-12-04T09:08:11.0192678Z dynamo/test_exceptions 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_exceptions_1.1_27ab688a8f5f578d_.log 2025-12-04T09:08:11.0193320Z Running 0 items in this shard: 2025-12-04T09:08:11.0193471Z 2025-12-04T09:08:11.0193714Z Finished dynamo/test_exceptions 1/1 ... [2025-12-04 09:08:11.018934][926.978480471], took 0.05min 2025-12-04T09:08:11.0219052Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_exceptions/dynamo.test_exceptions-68332230799da90f.xml 2025-12-04T09:08:11.0460420Z Running test_cuda 1/1 ... [2025-12-04 09:08:11.045815][927.005362855] 2025-12-04T09:08:11.0460817Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:08:11.0463405Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_cuda.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:08:11.046085] 2025-12-04T09:09:26.4522852Z 2025-12-04T09:09:26.4523614Z PRINTING LOG FILE of test_cuda 1/1 (test/test-reports/test_cuda_1.1_c907a69791bbc315_.log) 2025-12-04T09:09:26.4525324Z Test results will be stored in test-reports/python-pytest/test_cuda/test_cuda-0b3575165da88d35.xml 2025-12-04T09:09:26.4526073Z ============================= test session starts ============================== 2025-12-04T09:09:26.4526742Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:09:26.4527402Z cachedir: .pytest_cache 2025-12-04T09:09:26.4528120Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:09:26.4528938Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:09:26.4529316Z configfile: pytest.ini 2025-12-04T09:09:26.4529800Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:09:26.4530476Z collecting ... collected 252 items / 11200 deselected / -10948 selected 2025-12-04T09:09:26.4530834Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:09:26.4599129Z Running 250 items in this shard: test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator 2025-12-04T09:09:26.4643467Z 2025-12-04T09:09:26.4643667Z test_cuda.py::TestCuda::test_host_memory_stats PASSED [0.1396s] [ 0%] 2025-12-04T09:09:26.4644456Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 0%] 2025-12-04T09:09:26.4645221Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0026s] [ 1%] 2025-12-04T09:09:26.4646966Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=dummy_allocator -DTORCH_API_INCLUDE_EXTENSION_H -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -isystem /usr/local/cuda/include -isystem /opt/conda/envs/py_3.10/include/python3.10 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py310_cu128/dummy_allocator/main.cpp -o main.o 2025-12-04T09:09:26.4649086Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib -lc10 -lc10_cuda -ltorch_cpu -ltorch_cuda -ltorch -ltorch_python -L/usr/local/cuda/lib64 -lcudart -o dummy_allocator.so 2025-12-04T09:09:26.4649758Z PASSED [15.5348s] [ 1%] 2025-12-04T09:09:26.4651377Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=dummy_allocator_v1 -DTORCH_API_INCLUDE_EXTENSION_H -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -isystem /usr/local/cuda/include -isystem /opt/conda/envs/py_3.10/include/python3.10 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py310_cu128/dummy_allocator/main.cpp -o main.o 2025-12-04T09:09:26.4653560Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib -lc10 -lc10_cuda -ltorch_cpu -ltorch_cuda -ltorch -ltorch_python -L/usr/local/cuda/lib64 -lcudart -o dummy_allocator_v1.so 2025-12-04T09:09:26.4654238Z PASSED [15.3072s] [ 2%] 2025-12-04T09:09:26.4654515Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1945s] [ 2%] 2025-12-04T09:09:26.4654924Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1621s] [ 2%] 2025-12-04T09:09:26.4655497Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1574s] [ 2%] 2025-12-04T09:09:26.4655979Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1575s] [ 2%] 2025-12-04T09:09:26.4656354Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1571s] [ 2%] 2025-12-04T09:09:26.4656727Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1567s] [ 2%] 2025-12-04T09:09:26.4657095Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1559s] [ 2%] 2025-12-04T09:09:26.4657467Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1586s] [ 2%] 2025-12-04T09:09:26.4657831Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1586s] [ 2%] 2025-12-04T09:09:26.4658205Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1570s] [ 2%] 2025-12-04T09:09:26.4658585Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1583s] [ 2%] 2025-12-04T09:09:26.4658960Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1573s] [ 2%] 2025-12-04T09:09:26.4659343Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1564s] [ 2%] 2025-12-04T09:09:26.4659747Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1569s] [ 2%] 2025-12-04T09:09:26.4660127Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1577s] [ 2%] 2025-12-04T09:09:26.4660524Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1572s] [ 2%] 2025-12-04T09:09:26.4660898Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1564s] [ 2%] 2025-12-04T09:09:26.4661284Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1562s] [ 2%] 2025-12-04T09:09:26.4661760Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1562s] [ 2%] 2025-12-04T09:09:26.4662132Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1568s] [ 2%] 2025-12-04T09:09:26.4662505Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1563s] [ 2%] 2025-12-04T09:09:26.4662884Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1578s] [ 2%] 2025-12-04T09:09:26.4663253Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1572s] [ 2%] 2025-12-04T09:09:26.4663627Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1566s] [ 2%] 2025-12-04T09:09:26.4663998Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1568s] [ 2%] 2025-12-04T09:09:26.4664379Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1568s] [ 2%] 2025-12-04T09:09:26.4664750Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1568s] [ 2%] 2025-12-04T09:09:26.4665124Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1570s] [ 2%] 2025-12-04T09:09:26.4665540Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1574s] [ 2%] 2025-12-04T09:09:26.4665943Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1572s] [ 2%] 2025-12-04T09:09:26.4666311Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1563s] [ 2%] 2025-12-04T09:09:26.4666692Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1565s] [ 2%] 2025-12-04T09:09:26.4667066Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1577s] [ 2%] 2025-12-04T09:09:26.4667434Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1562s] [ 2%] 2025-12-04T09:09:26.4667806Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1565s] [ 2%] 2025-12-04T09:09:26.4668179Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1583s] [ 2%] 2025-12-04T09:09:26.4668556Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1571s] [ 2%] 2025-12-04T09:09:26.4668921Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1564s] [ 2%] 2025-12-04T09:09:26.4669296Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1575s] [ 2%] 2025-12-04T09:09:26.4669669Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1564s] [ 2%] 2025-12-04T09:09:26.4670123Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1562s] [ 2%] 2025-12-04T09:09:26.4670562Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1573s] [ 2%] 2025-12-04T09:09:26.4670939Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1573s] [ 2%] 2025-12-04T09:09:26.4671311Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1581s] [ 2%] 2025-12-04T09:09:26.4671682Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1580s] [ 2%] 2025-12-04T09:09:26.4672053Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1564s] [ 2%] 2025-12-04T09:09:26.4672424Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1620s] [ 2%] 2025-12-04T09:09:26.4672791Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1582s] [ 2%] 2025-12-04T09:09:26.4673160Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1568s] [ 2%] 2025-12-04T09:09:26.4673896Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4674967Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4676005Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4677032Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4678068Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4679115Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4680147Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4681179Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4682208Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4683315Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4684353Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4685387Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4686421Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4687450Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4688670Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4689711Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4690743Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4691775Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4692816Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4693947Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4694985Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4696021Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4697064Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4698109Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4699138Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4700174Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4701209Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4702256Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4703294Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4704320Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4706051Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4707223Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4708383Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4709429Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4710481Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4711526Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4712575Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4713614Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4714653Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4715682Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4716724Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4717769Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4718814Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4719865Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4720901Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4721934Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4722967Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4724002Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4725041Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:09:26.4725845Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0023s] [ 2%] 2025-12-04T09:09:26.4726310Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0019s] [ 2%] 2025-12-04T09:09:26.4726720Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0019s] [ 2%] 2025-12-04T09:09:26.4727123Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0019s] [ 2%] 2025-12-04T09:09:26.4727514Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0019s] [ 2%] 2025-12-04T09:09:26.4727914Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0019s] [ 2%] 2025-12-04T09:09:26.4728305Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4728691Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4729087Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4729492Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0019s] [ 2%] 2025-12-04T09:09:26.4729893Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0023s] [ 2%] 2025-12-04T09:09:26.4730278Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4730669Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4731072Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4731483Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0021s] [ 2%] 2025-12-04T09:09:26.4731892Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0019s] [ 2%] 2025-12-04T09:09:26.4732289Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4732685Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4733076Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4734163Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4734579Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4734976Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:09:26.4735367Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0019s] [ 2%] 2025-12-04T09:09:26.4735761Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4736163Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4736554Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4736953Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4737353Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4737752Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4738146Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4738542Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4738956Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4739349Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4739751Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4740145Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4740538Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4740927Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4741328Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4741835Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4742289Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4742686Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4743081Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4743474Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4743866Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4744258Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4744665Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4745067Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4745464Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4745864Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0018s] [ 2%] 2025-12-04T09:09:26.4747581Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=dummy_allocator_v2 -DTORCH_API_INCLUDE_EXTENSION_H -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -isystem /usr/local/cuda/include -isystem /opt/conda/envs/py_3.10/include/python3.10 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py310_cu128/dummy_allocator/main.cpp -o main.o 2025-12-04T09:09:26.4749667Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib -lc10 -lc10_cuda -ltorch_cpu -ltorch_cuda -ltorch -ltorch_python -L/usr/local/cuda/lib64 -lcudart -o dummy_allocator_v2.so 2025-12-04T09:09:26.4750360Z PASSED [15.2910s] [ 2%] 2025-12-04T09:09:26.4750659Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0015s] [ 2%] 2025-12-04T09:09:26.4751099Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4751530Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0013s] [ 2%] 2025-12-04T09:09:26.4751942Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4752359Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4752779Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4753195Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4753608Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0013s] [ 2%] 2025-12-04T09:09:26.4754024Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4754450Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4754867Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0011s] [ 2%] 2025-12-04T09:09:26.4755290Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4755713Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4756134Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4756545Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4756963Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4757382Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4757806Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4758400Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0013s] [ 2%] 2025-12-04T09:09:26.4758831Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4759253Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4759669Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0011s] [ 2%] 2025-12-04T09:09:26.4760084Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4760512Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4760927Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4761336Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4761753Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0011s] [ 2%] 2025-12-04T09:09:26.4762177Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4762595Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0013s] [ 2%] 2025-12-04T09:09:26.4763006Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:09:26.4763422Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4763836Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4764246Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0013s] [ 2%] 2025-12-04T09:09:26.4764666Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4765084Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0011s] [ 2%] 2025-12-04T09:09:26.4765503Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4765922Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4766344Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4766765Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4767177Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0011s] [ 2%] 2025-12-04T09:09:26.4767596Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:09:26.4768014Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4768436Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0011s] [ 2%] 2025-12-04T09:09:26.4768854Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4769271Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4769697Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4770119Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0012s] [ 2%] 2025-12-04T09:09:26.4770530Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0011s] [ 2%] 2025-12-04T09:09:26.4772274Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=dummy_allocator_v3 -DTORCH_API_INCLUDE_EXTENSION_H -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -isystem /usr/local/cuda/include -isystem /opt/conda/envs/py_3.10/include/python3.10 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py310_cu128/dummy_allocator/main.cpp -o main.o 2025-12-04T09:09:26.4774525Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib -lc10 -lc10_cuda -ltorch_cpu -ltorch_cuda -ltorch -ltorch_python -L/usr/local/cuda/lib64 -lcudart -o dummy_allocator_v3.so 2025-12-04T09:09:26.4775272Z PASSED [15.2203s] [ 2%] 2025-12-04T09:09:26.4775594Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0168s] [ 2%] 2025-12-04T09:09:26.4776077Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0166s] [ 2%] 2025-12-04T09:09:26.4776557Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0165s] [ 2%] 2025-12-04T09:09:26.4777046Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0168s] [ 2%] 2025-12-04T09:09:26.4777519Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0164s] [ 2%] 2025-12-04T09:09:26.4777991Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0165s] [ 2%] 2025-12-04T09:09:26.4778467Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0164s] [ 2%] 2025-12-04T09:09:26.4778949Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0167s] [ 2%] 2025-12-04T09:09:26.4779417Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0164s] [ 2%] 2025-12-04T09:09:26.4779886Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0168s] [ 2%] 2025-12-04T09:09:26.4780357Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0164s] [ 2%] 2025-12-04T09:09:26.4780833Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0162s] [ 2%] 2025-12-04T09:09:26.4781296Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0163s] [ 2%] 2025-12-04T09:09:26.4781773Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0165s] [ 2%] 2025-12-04T09:09:26.4782256Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0164s] [ 2%] 2025-12-04T09:09:26.4782732Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0164s] [ 2%] 2025-12-04T09:09:26.4783215Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0166s] [ 2%] 2025-12-04T09:09:26.4783687Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0168s] [ 2%] 2025-12-04T09:09:26.4784159Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0164s] [ 2%] 2025-12-04T09:09:26.4784628Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0162s] [ 2%] 2025-12-04T09:09:26.4785092Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0163s] [ 2%] 2025-12-04T09:09:26.4785564Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0167s] [ 2%] 2025-12-04T09:09:26.4786038Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0166s] [ 2%] 2025-12-04T09:09:26.4786512Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0163s] [ 2%] 2025-12-04T09:09:26.4786988Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0163s] [ 2%] 2025-12-04T09:09:26.4787459Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0164s] [ 2%] 2025-12-04T09:09:26.4787929Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0163s] [ 2%] 2025-12-04T09:09:26.4788391Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0164s] [ 2%] 2025-12-04T09:09:26.4788866Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0163s] [ 2%] 2025-12-04T09:09:26.4789338Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0164s] [ 2%] 2025-12-04T09:09:26.4789811Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0166s] [ 2%] 2025-12-04T09:09:26.4790275Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0166s] [ 2%] 2025-12-04T09:09:26.4790889Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0167s] [ 2%] 2025-12-04T09:09:26.4791360Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0164s] [ 2%] 2025-12-04T09:09:26.4791829Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0161s] [ 2%] 2025-12-04T09:09:26.4792292Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0168s] [ 2%] 2025-12-04T09:09:26.4792758Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0163s] [ 2%] 2025-12-04T09:09:26.4793240Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0166s] [ 2%] 2025-12-04T09:09:26.4793701Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0165s] [ 2%] 2025-12-04T09:09:26.4794166Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0166s] [ 2%] 2025-12-04T09:09:26.4794638Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0164s] [ 2%] 2025-12-04T09:09:26.4795106Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0164s] [ 2%] 2025-12-04T09:09:26.4795572Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0163s] [ 2%] 2025-12-04T09:09:26.4796041Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0162s] [ 2%] 2025-12-04T09:09:26.4796510Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0165s] [ 2%] 2025-12-04T09:09:26.4796979Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0168s] [ 2%] 2025-12-04T09:09:26.4797443Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0165s] [ 2%] 2025-12-04T09:09:26.4797910Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0165s] [ 2%] 2025-12-04T09:09:26.4798189Z 2025-12-04T09:09:26.4798293Z =================================== FAILURES =================================== 2025-12-04T09:09:26.4798626Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4798921Z Traceback (most recent call last): 2025-12-04T09:09:26.4799285Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4799663Z check_stats(expected) 2025-12-04T09:09:26.4799958Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4800290Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4800763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4801295Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4801576Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4801735Z 2025-12-04T09:09:26.4801809Z Expected 196608 but got 0. 2025-12-04T09:09:26.4802003Z Absolute difference: 196608 2025-12-04T09:09:26.4802203Z Relative difference: 1.0 2025-12-04T09:09:26.4802330Z 2025-12-04T09:09:26.4802461Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4802792Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4802989Z 2025-12-04T09:09:26.4803157Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4803533Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4803865Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4804189Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4804475Z Traceback (most recent call last): 2025-12-04T09:09:26.4805293Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4805928Z check_stats(expected) 2025-12-04T09:09:26.4806407Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4806887Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4807376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4807906Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4808186Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4808344Z 2025-12-04T09:09:26.4808414Z Expected 196608 but got 0. 2025-12-04T09:09:26.4808603Z Absolute difference: 196608 2025-12-04T09:09:26.4808802Z Relative difference: 1.0 2025-12-04T09:09:26.4808923Z 2025-12-04T09:09:26.4809051Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4809379Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4809573Z 2025-12-04T09:09:26.4809737Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4810112Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4810438Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4810746Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4811054Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4811367Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4811663Z Traceback (most recent call last): 2025-12-04T09:09:26.4812019Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4812374Z check_stats(expected) 2025-12-04T09:09:26.4812678Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4813020Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4813561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4814090Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4814376Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4814526Z 2025-12-04T09:09:26.4814601Z Expected 196608 but got 0. 2025-12-04T09:09:26.4814783Z Absolute difference: 196608 2025-12-04T09:09:26.4814976Z Relative difference: 1.0 2025-12-04T09:09:26.4815092Z 2025-12-04T09:09:26.4815226Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4815550Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4815754Z 2025-12-04T09:09:26.4815914Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4816284Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4816610Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4816926Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4817236Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4817544Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4817846Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4818160Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4818453Z Traceback (most recent call last): 2025-12-04T09:09:26.4818811Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4819168Z check_stats(expected) 2025-12-04T09:09:26.4819457Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4819789Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4820245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4820854Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4821135Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4821351Z 2025-12-04T09:09:26.4821429Z Expected 196608 but got 0. 2025-12-04T09:09:26.4821610Z Absolute difference: 196608 2025-12-04T09:09:26.4821800Z Relative difference: 1.0 2025-12-04T09:09:26.4821914Z 2025-12-04T09:09:26.4822049Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4822366Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4822568Z 2025-12-04T09:09:26.4822727Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4823090Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4823406Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4823709Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4824030Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4824339Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4824643Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4824948Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4825250Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4825556Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4825838Z Traceback (most recent call last): 2025-12-04T09:09:26.4826186Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4826543Z check_stats(expected) 2025-12-04T09:09:26.4826831Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4827156Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4827611Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4828146Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4828436Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4828595Z 2025-12-04T09:09:26.4828663Z Expected 196608 but got 0. 2025-12-04T09:09:26.4828852Z Absolute difference: 196608 2025-12-04T09:09:26.4829045Z Relative difference: 1.0 2025-12-04T09:09:26.4829164Z 2025-12-04T09:09:26.4829289Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4829614Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4829810Z 2025-12-04T09:09:26.4829974Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4830337Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4830654Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4830968Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4831278Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4831580Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4831890Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4832193Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4832494Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4832794Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4833101Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4833404Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4833695Z Traceback (most recent call last): 2025-12-04T09:09:26.4834056Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4834494Z check_stats(expected) 2025-12-04T09:09:26.4834848Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4835179Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4835636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4836152Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4836439Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4836590Z 2025-12-04T09:09:26.4836669Z Expected 196608 but got 0. 2025-12-04T09:09:26.4836851Z Absolute difference: 196608 2025-12-04T09:09:26.4837042Z Relative difference: 1.0 2025-12-04T09:09:26.4837163Z 2025-12-04T09:09:26.4837289Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4837617Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4837813Z 2025-12-04T09:09:26.4837975Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4838346Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4838662Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4838962Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4839271Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4839572Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4839881Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4840173Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4840476Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4840776Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4841075Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4841377Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4841696Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4842011Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4842291Z Traceback (most recent call last): 2025-12-04T09:09:26.4842647Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4843009Z check_stats(expected) 2025-12-04T09:09:26.4843297Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4843625Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4844084Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4844599Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4844879Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4845034Z 2025-12-04T09:09:26.4845106Z Expected 196608 but got 0. 2025-12-04T09:09:26.4845299Z Absolute difference: 196608 2025-12-04T09:09:26.4845481Z Relative difference: 1.0 2025-12-04T09:09:26.4845600Z 2025-12-04T09:09:26.4845725Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4846047Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4846247Z 2025-12-04T09:09:26.4846413Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4846775Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4847086Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4847393Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4847695Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4847998Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4848396Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4848779Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4849088Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4849392Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4849701Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4849994Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4850303Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4850605Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4850931Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4851254Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4851549Z Traceback (most recent call last): 2025-12-04T09:09:26.4851905Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4852258Z check_stats(expected) 2025-12-04T09:09:26.4852555Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4852887Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4853434Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4853953Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4854238Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4854388Z 2025-12-04T09:09:26.4854464Z Expected 196608 but got 0. 2025-12-04T09:09:26.4854648Z Absolute difference: 196608 2025-12-04T09:09:26.4854838Z Relative difference: 1.0 2025-12-04T09:09:26.4854952Z 2025-12-04T09:09:26.4855088Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4855416Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4855619Z 2025-12-04T09:09:26.4855782Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4856155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4856483Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4856793Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4857105Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4857408Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4857717Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4858017Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4858322Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4858643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4858948Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4859258Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4859564Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4859863Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4860162Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4860459Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4860765Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4861067Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4861357Z Traceback (most recent call last): 2025-12-04T09:09:26.4861720Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4862157Z check_stats(expected) 2025-12-04T09:09:26.4862531Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4862864Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4863323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4863843Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4864130Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4864282Z 2025-12-04T09:09:26.4864359Z Expected 196608 but got 0. 2025-12-04T09:09:26.4864546Z Absolute difference: 196608 2025-12-04T09:09:26.4864742Z Relative difference: 1.0 2025-12-04T09:09:26.4864863Z 2025-12-04T09:09:26.4864995Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4865325Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4865525Z 2025-12-04T09:09:26.4865688Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4866065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4866384Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4866696Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4867010Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4867316Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4867621Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4867913Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4868217Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4868514Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4868829Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4869143Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4869451Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4869752Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4870053Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4870349Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4870650Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4870942Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4871242Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4871551Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4871838Z Traceback (most recent call last): 2025-12-04T09:09:26.4872183Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4872542Z check_stats(expected) 2025-12-04T09:09:26.4872831Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4873153Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4873607Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4874124Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4874402Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4874550Z 2025-12-04T09:09:26.4874617Z Expected 196608 but got 0. 2025-12-04T09:09:26.4874806Z Absolute difference: 196608 2025-12-04T09:09:26.4875007Z Relative difference: 1.0 2025-12-04T09:09:26.4875121Z 2025-12-04T09:09:26.4875246Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4875569Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4875844Z 2025-12-04T09:09:26.4876080Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4876457Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4876766Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4877071Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4877382Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4877675Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4877982Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4878284Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4878590Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4878893Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4879209Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4879517Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4879824Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4880136Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4880444Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4880746Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4881046Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4881355Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4881662Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4881956Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4882265Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4882573Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4882860Z Traceback (most recent call last): 2025-12-04T09:09:26.4883203Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4883556Z check_stats(expected) 2025-12-04T09:09:26.4883851Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4884170Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4884621Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4885139Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4885414Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4885567Z 2025-12-04T09:09:26.4885635Z Expected 196608 but got 0. 2025-12-04T09:09:26.4885823Z Absolute difference: 196608 2025-12-04T09:09:26.4886013Z Relative difference: 1.0 2025-12-04T09:09:26.4886130Z 2025-12-04T09:09:26.4886270Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4886596Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4886794Z 2025-12-04T09:09:26.4886957Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4887327Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4887637Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4887946Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4888259Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4888564Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4888869Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4889262Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4889627Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4889930Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4890235Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4890538Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4890841Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4891150Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4891454Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4891748Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4892052Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4892352Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4892661Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4892962Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4893327Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4893633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4893933Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4894247Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4894538Z Traceback (most recent call last): 2025-12-04T09:09:26.4894893Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4895253Z check_stats(expected) 2025-12-04T09:09:26.4895545Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4895874Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4896340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4896874Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4897157Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4897306Z 2025-12-04T09:09:26.4897379Z Expected 196608 but got 0. 2025-12-04T09:09:26.4897558Z Absolute difference: 196608 2025-12-04T09:09:26.4897747Z Relative difference: 1.0 2025-12-04T09:09:26.4897859Z 2025-12-04T09:09:26.4897990Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4898307Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4898508Z 2025-12-04T09:09:26.4898665Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4899039Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4899349Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4899650Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4899961Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4900260Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4900559Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4900861Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4901161Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4901461Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4901761Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4902067Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4902374Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4902747Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4903116Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4903417Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4903737Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4904032Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4904332Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4904867Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4905189Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4905496Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4905806Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4906112Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4906420Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4906730Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4907023Z Traceback (most recent call last): 2025-12-04T09:09:26.4907380Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4907748Z check_stats(expected) 2025-12-04T09:09:26.4908044Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4908373Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4908842Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4909373Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4909672Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4909825Z 2025-12-04T09:09:26.4909901Z Expected 196608 but got 0. 2025-12-04T09:09:26.4910096Z Absolute difference: 196608 2025-12-04T09:09:26.4910291Z Relative difference: 1.0 2025-12-04T09:09:26.4910411Z 2025-12-04T09:09:26.4910544Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4910879Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4911082Z 2025-12-04T09:09:26.4911245Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4911625Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4911945Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4912281Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4912594Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4912903Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4913213Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4913522Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4913835Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4914132Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4914438Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4914747Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4915047Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4930809Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4931189Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4931524Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4931857Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4932395Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4932844Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4933249Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4933584Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4933896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4934214Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4934513Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4934825Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4935131Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4935436Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4935751Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4936067Z Traceback (most recent call last): 2025-12-04T09:09:26.4936449Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4936818Z check_stats(expected) 2025-12-04T09:09:26.4937132Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4937473Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4937952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4938489Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4938788Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4938944Z 2025-12-04T09:09:26.4939022Z Expected 196608 but got 0. 2025-12-04T09:09:26.4939225Z Absolute difference: 196608 2025-12-04T09:09:26.4939428Z Relative difference: 1.0 2025-12-04T09:09:26.4939547Z 2025-12-04T09:09:26.4939698Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4940044Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4940260Z 2025-12-04T09:09:26.4940429Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4940820Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4941150Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4941457Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4941776Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4942089Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4942389Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4942684Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4942992Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4943299Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4943618Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4943923Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4944235Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4944535Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4944844Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4945143Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4945459Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4945760Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4946072Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4946461Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4946830Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4947129Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4947434Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4947739Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4948034Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4948325Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4948629Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4948917Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4949216Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4949519Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4949808Z Traceback (most recent call last): 2025-12-04T09:09:26.4950164Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4950530Z check_stats(expected) 2025-12-04T09:09:26.4950817Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4951137Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4951603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4952126Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4952401Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4952554Z 2025-12-04T09:09:26.4952621Z Expected 196608 but got 0. 2025-12-04T09:09:26.4952807Z Absolute difference: 196608 2025-12-04T09:09:26.4952993Z Relative difference: 1.0 2025-12-04T09:09:26.4953110Z 2025-12-04T09:09:26.4953243Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4953572Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4953782Z 2025-12-04T09:09:26.4953947Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4954309Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4954618Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4954917Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4955216Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4955503Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4955797Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4956089Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4956395Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4956691Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4956990Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4957273Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4957568Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4957855Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4958151Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4958436Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4958737Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4959029Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4959325Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4959698Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4960087Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4960389Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4960698Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4961006Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4961313Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4961607Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4961906Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4962211Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4962508Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4962797Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4963114Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4963429Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4963713Z Traceback (most recent call last): 2025-12-04T09:09:26.4964070Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4964427Z check_stats(expected) 2025-12-04T09:09:26.4964711Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4965032Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4965489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4966003Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4966276Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4966427Z 2025-12-04T09:09:26.4966497Z Expected 196608 but got 0. 2025-12-04T09:09:26.4966676Z Absolute difference: 196608 2025-12-04T09:09:26.4966861Z Relative difference: 1.0 2025-12-04T09:09:26.4966980Z 2025-12-04T09:09:26.4967117Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4967436Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4967628Z 2025-12-04T09:09:26.4967789Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4968155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4968466Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4968763Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4969062Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4969351Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4969657Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4969954Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4970250Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4970538Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4970835Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4971125Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4971425Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4971715Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4972012Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4972298Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4972596Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4972991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4973425Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4973720Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4974019Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4974310Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4974604Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4974895Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4975202Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4975498Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4975799Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4976112Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4976421Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4976717Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4977019Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4977319Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4977624Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4977928Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4978225Z Traceback (most recent call last): 2025-12-04T09:09:26.4978589Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4978949Z check_stats(expected) 2025-12-04T09:09:26.4979240Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4979567Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4980029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4980561Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4980847Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4981000Z 2025-12-04T09:09:26.4981075Z Expected 196608 but got 0. 2025-12-04T09:09:26.4981262Z Absolute difference: 196608 2025-12-04T09:09:26.4981471Z Relative difference: 1.0 2025-12-04T09:09:26.4981590Z 2025-12-04T09:09:26.4981728Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4982053Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4982256Z 2025-12-04T09:09:26.4982418Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4982797Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4983136Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4983452Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4983767Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4984070Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4984374Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4984677Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4984988Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4985285Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4985590Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4985891Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4986200Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4986589Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4986967Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4987273Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4987581Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4987875Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4988180Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4988482Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4988789Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4989089Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4989402Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4989706Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4990026Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4990332Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4990639Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4990933Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4991239Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4991536Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4991840Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4992134Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4992440Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4992739Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4993043Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4993358Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.4993653Z Traceback (most recent call last): 2025-12-04T09:09:26.4994012Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.4994371Z check_stats(expected) 2025-12-04T09:09:26.4994667Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.4995001Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.4995462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.4995991Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.4996280Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.4996437Z 2025-12-04T09:09:26.4996511Z Expected 196608 but got 0. 2025-12-04T09:09:26.4996700Z Absolute difference: 196608 2025-12-04T09:09:26.4996892Z Relative difference: 1.0 2025-12-04T09:09:26.4997007Z 2025-12-04T09:09:26.4997146Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.4997475Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.4997682Z 2025-12-04T09:09:26.4997846Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.4998224Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4998547Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4998862Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4999171Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.4999471Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.4999774Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5000155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5000535Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5000844Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5001145Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5001442Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5001760Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5002052Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5002358Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5002660Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5002963Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5003258Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5003568Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5003868Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5004170Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5004467Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5005078Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5005400Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5005716Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5006022Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5006333Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5006630Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5006940Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5007248Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5007551Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5007850Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5008155Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5008452Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5008758Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5009056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5009366Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5009671Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5009966Z Traceback (most recent call last): 2025-12-04T09:09:26.5010332Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5010704Z check_stats(expected) 2025-12-04T09:09:26.5011003Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5011340Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5011823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5012355Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5012640Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5012800Z 2025-12-04T09:09:26.5012869Z Expected 196608 but got 0. 2025-12-04T09:09:26.5013059Z Absolute difference: 196608 2025-12-04T09:09:26.5013309Z Relative difference: 1.0 2025-12-04T09:09:26.5013433Z 2025-12-04T09:09:26.5013564Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5014027Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5014230Z 2025-12-04T09:09:26.5014517Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5014911Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5015235Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5015544Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5015851Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5016156Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5016471Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5016772Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5017078Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5017378Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5017693Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5017981Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5018277Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5018573Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5018891Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5019197Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5019503Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5019802Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5020106Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5020407Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5020717Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5021017Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5021324Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5021626Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5021940Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5022240Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5022543Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5022838Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5023136Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5023427Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5023731Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5024035Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5024335Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5024633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5024935Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5025228Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5025531Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5025826Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5026147Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5026452Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5026743Z Traceback (most recent call last): 2025-12-04T09:09:26.5027176Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5027530Z check_stats(expected) 2025-12-04T09:09:26.5027889Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5028225Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5028687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5029202Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5029488Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5029640Z 2025-12-04T09:09:26.5029714Z Expected 196608 but got 0. 2025-12-04T09:09:26.5029904Z Absolute difference: 196608 2025-12-04T09:09:26.5030096Z Relative difference: 1.0 2025-12-04T09:09:26.5030212Z 2025-12-04T09:09:26.5030361Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5030690Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5030894Z 2025-12-04T09:09:26.5031061Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5031435Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5031754Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5032054Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5032367Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5032672Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5032982Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5033277Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5033581Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5033882Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5034188Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5034490Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5034795Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5035090Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5035389Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5035686Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5035998Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5036291Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5036592Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5036886Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5037193Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5037491Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5037793Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5038089Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5038393Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5038693Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5038994Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5039283Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5039589Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5039885Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5040190Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5040572Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5040960Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5041261Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5041561Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5041859Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5042166Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5042461Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5042761Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5043056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5043362Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5043663Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5043957Z Traceback (most recent call last): 2025-12-04T09:09:26.5044325Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5044685Z check_stats(expected) 2025-12-04T09:09:26.5044970Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5045297Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5045757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5046268Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5046549Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5046705Z 2025-12-04T09:09:26.5046774Z Expected 196608 but got 0. 2025-12-04T09:09:26.5046964Z Absolute difference: 196608 2025-12-04T09:09:26.5047151Z Relative difference: 1.0 2025-12-04T09:09:26.5047270Z 2025-12-04T09:09:26.5047404Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5047733Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5047930Z 2025-12-04T09:09:26.5048094Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5048462Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5048779Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5049094Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5049400Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5049698Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5050008Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5050307Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5050617Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5050919Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5051233Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5051525Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5051830Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5052126Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5052443Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5052730Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5053033Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5053407Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5053704Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5054085Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5054248Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5054377Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5054480Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5054607Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5054706Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5054827Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5054938Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5055063Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5055161Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5055285Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5055386Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5055516Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5055615Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5055734Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5055834Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5055957Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5056055Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5056179Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5056276Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5056401Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5056502Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5056626Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5056728Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5056859Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5056937Z Traceback (most recent call last): 2025-12-04T09:09:26.5057153Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5057221Z check_stats(expected) 2025-12-04T09:09:26.5057398Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5057470Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5057797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5057923Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5058009Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5058013Z 2025-12-04T09:09:26.5058084Z Expected 196608 but got 0. 2025-12-04T09:09:26.5058160Z Absolute difference: 196608 2025-12-04T09:09:26.5058226Z Relative difference: 1.0 2025-12-04T09:09:26.5058229Z 2025-12-04T09:09:26.5058359Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5058486Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5058492Z 2025-12-04T09:09:26.5058655Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5058786Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5058889Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5059014Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5059113Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5059350Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5059514Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5059637Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5059737Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5059863Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5059961Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5060085Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5060198Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5060323Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5060426Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5060547Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5060648Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5060774Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5060873Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5060997Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5061095Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5061215Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5061318Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5061444Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5061543Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5061674Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5061775Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5061903Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5062002Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5062122Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5062222Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5062340Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5062436Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5062570Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5062671Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5062795Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5062894Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5063018Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5063121Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5063242Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5063340Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5063463Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5063560Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5063685Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5063784Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5063918Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5063999Z Traceback (most recent call last): 2025-12-04T09:09:26.5064207Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5064363Z check_stats(expected) 2025-12-04T09:09:26.5064608Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5064684Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5065004Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5065140Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5065221Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5065225Z 2025-12-04T09:09:26.5065297Z Expected 196608 but got 0. 2025-12-04T09:09:26.5065366Z Absolute difference: 196608 2025-12-04T09:09:26.5065433Z Relative difference: 1.0 2025-12-04T09:09:26.5065436Z 2025-12-04T09:09:26.5065569Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5065701Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5065708Z 2025-12-04T09:09:26.5065877Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5066009Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5066118Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5066248Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5066354Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5066477Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5066583Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5066703Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5066804Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5066925Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5067027Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5067154Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5067251Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5067371Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5067478Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5067599Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5067701Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5067821Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5067923Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5068049Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5068148Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5068271Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5068378Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5068497Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5068599Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5068718Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5068825Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5068955Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5069054Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5069176Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5069279Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5069492Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5069664Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5069789Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5069890Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5070016Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5070114Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5070233Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5070335Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5070455Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5070557Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5070677Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5070778Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5070908Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5071008Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5071130Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5071230Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5071362Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5071445Z Traceback (most recent call last): 2025-12-04T09:09:26.5071649Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5071717Z check_stats(expected) 2025-12-04T09:09:26.5071896Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5071973Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5072301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5072437Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5072517Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5072520Z 2025-12-04T09:09:26.5072590Z Expected 196608 but got 0. 2025-12-04T09:09:26.5072657Z Absolute difference: 196608 2025-12-04T09:09:26.5072725Z Relative difference: 1.0 2025-12-04T09:09:26.5072728Z 2025-12-04T09:09:26.5072856Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5072976Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5072979Z 2025-12-04T09:09:26.5073138Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5073269Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5073371Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5073505Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5073604Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5073730Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5073834Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5073954Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5074052Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5074178Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5074277Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5074403Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5074501Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5074698Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5074861Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5074986Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5075086Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5075208Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5075309Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5075434Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5075531Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5075652Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5075753Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5075874Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5075975Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5076101Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5076199Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5076327Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5076435Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5076559Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5076661Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5076781Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5076877Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5077000Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5077100Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5077225Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5077322Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5077442Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5077548Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5077671Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5077770Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5077895Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5077993Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5078116Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5078216Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5078340Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5078447Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5078569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5078668Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5078803Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5078875Z Traceback (most recent call last): 2025-12-04T09:09:26.5079082Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5079149Z check_stats(expected) 2025-12-04T09:09:26.5079321Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5079396Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5079707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5079991Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5080076Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5080080Z 2025-12-04T09:09:26.5080145Z Expected 196608 but got 0. 2025-12-04T09:09:26.5080220Z Absolute difference: 196608 2025-12-04T09:09:26.5080286Z Relative difference: 1.0 2025-12-04T09:09:26.5080290Z 2025-12-04T09:09:26.5080413Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5080534Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5080538Z 2025-12-04T09:09:26.5080693Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5080818Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5080924Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5081046Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5081156Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5081280Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5081379Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5081506Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5081603Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5081723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5081829Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5081949Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5082056Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5082177Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5082285Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5082421Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5082520Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5082643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5082760Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5082883Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5082989Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5083109Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5083210Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5083333Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5083431Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5083556Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5083662Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5083782Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5083885Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5084011Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5084109Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5084235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5084334Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5084455Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5084560Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5084756Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5084937Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5085065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5085164Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5085288Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5085386Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5085505Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5085609Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5085730Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5085835Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5085967Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5086071Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5086202Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5086300Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5086420Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5086520Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5086650Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5086731Z Traceback (most recent call last): 2025-12-04T09:09:26.5086939Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5087008Z check_stats(expected) 2025-12-04T09:09:26.5087183Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5087254Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5087575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5087715Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5087796Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5087800Z 2025-12-04T09:09:26.5087873Z Expected 196608 but got 0. 2025-12-04T09:09:26.5087943Z Absolute difference: 196608 2025-12-04T09:09:26.5088008Z Relative difference: 1.0 2025-12-04T09:09:26.5088012Z 2025-12-04T09:09:26.5088143Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5088262Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5088265Z 2025-12-04T09:09:26.5088428Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5088554Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5088656Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5088786Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5088886Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5089007Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5089111Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5089230Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5089333Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5089453Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5089550Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5089677Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5089775Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5089982Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5090147Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5090272Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5090373Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5090495Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5090593Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5090719Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5090816Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5090937Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5091048Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5091171Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5091277Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5091401Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5091501Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5091624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5091721Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5091840Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5091943Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5092065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5092168Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5092287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5092391Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5092518Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5092616Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5092735Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5092836Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5092958Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5093057Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5093237Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5093339Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5093462Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5093561Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5093686Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5093791Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5093911Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5094009Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5094137Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5094235Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5094360Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5094458Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5094589Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5094665Z Traceback (most recent call last): 2025-12-04T09:09:26.5094876Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5095022Z check_stats(expected) 2025-12-04T09:09:26.5095286Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5095372Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5095695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5095817Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5095899Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5095903Z 2025-12-04T09:09:26.5095978Z Expected 196608 but got 0. 2025-12-04T09:09:26.5096047Z Absolute difference: 196608 2025-12-04T09:09:26.5096113Z Relative difference: 1.0 2025-12-04T09:09:26.5096125Z 2025-12-04T09:09:26.5096252Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5096373Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5096381Z 2025-12-04T09:09:26.5096544Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5096671Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5096776Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5096908Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5097009Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5097139Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5097242Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5097366Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5097471Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5097594Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5097696Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5097823Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5097922Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5098050Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5098149Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5098270Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5098373Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5098497Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5098596Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5098722Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5098822Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5098950Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5099063Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5099196Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5099300Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5099423Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5099523Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5099649Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5099747Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5099874Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5099974Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5100095Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5100348Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5100477Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5100577Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5100703Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5100802Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5100928Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5101038Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5101163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5101266Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5101386Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5101490Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5101620Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5101723Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5101847Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5101952Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5102074Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5102178Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5102299Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5102399Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5102527Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5102628Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5102764Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5102862Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5102993Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5103075Z Traceback (most recent call last): 2025-12-04T09:09:26.5103281Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5103352Z check_stats(expected) 2025-12-04T09:09:26.5103527Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5103598Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5103919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5104055Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5104142Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5104146Z 2025-12-04T09:09:26.5104219Z Expected 196608 but got 0. 2025-12-04T09:09:26.5104293Z Absolute difference: 196608 2025-12-04T09:09:26.5104360Z Relative difference: 1.0 2025-12-04T09:09:26.5104364Z 2025-12-04T09:09:26.5104678Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5104839Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5104844Z 2025-12-04T09:09:26.5105010Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5105139Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5105243Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5105370Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5105471Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5105598Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5105940Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5106080Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5106191Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5106323Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5106429Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5106557Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5106659Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5106781Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5106887Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5107010Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5107118Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5107244Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5107347Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5107475Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5107574Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5107699Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5107805Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5107927Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5108032Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5108155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5108255Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5108389Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5108488Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5108612Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5108725Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5108848Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5108960Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5109082Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5109185Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5109314Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5109412Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5109538Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5109648Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5109771Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5109878Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5110011Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5110113Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5110239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5110338Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5110458Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5110565Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5110688Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5110869Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5111055Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5111157Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5111284Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5111383Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5111504Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5111609Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5111732Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5111836Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5111974Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5112051Z Traceback (most recent call last): 2025-12-04T09:09:26.5112278Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5112347Z check_stats(expected) 2025-12-04T09:09:26.5112524Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5112602Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5112923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5113053Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5113134Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5113138Z 2025-12-04T09:09:26.5113204Z Expected 196608 but got 0. 2025-12-04T09:09:26.5113282Z Absolute difference: 196608 2025-12-04T09:09:26.5113348Z Relative difference: 1.0 2025-12-04T09:09:26.5113352Z 2025-12-04T09:09:26.5113479Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5113610Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5113614Z 2025-12-04T09:09:26.5113784Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5113918Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5114022Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5114147Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5114250Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5114372Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5114471Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5114597Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5114695Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5114824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5114925Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5115048Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5115153Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5115276Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5115374Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5115500Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5115601Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5115730Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5115829Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5115951Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5116150Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5116594Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5116701Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5116829Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5116928Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5117062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5117169Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5117293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5117399Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5117521Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5117622Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5117752Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5117852Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5117981Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5118081Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5118205Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5118308Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5118430Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5118537Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5118666Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5118767Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5118898Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5119001Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5119124Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5119233Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5119355Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5119455Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5119583Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5119682Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5119810Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5119910Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5120038Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5120145Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5120270Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5120371Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5120500Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5120600Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5120727Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5120829Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5120965Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5121045Z Traceback (most recent call last): 2025-12-04T09:09:26.5121253Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5121402Z check_stats(expected) 2025-12-04T09:09:26.5121642Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5121717Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5122039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5122163Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5122242Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5122245Z 2025-12-04T09:09:26.5122319Z Expected 196608 but got 0. 2025-12-04T09:09:26.5122388Z Absolute difference: 196608 2025-12-04T09:09:26.5122454Z Relative difference: 1.0 2025-12-04T09:09:26.5122457Z 2025-12-04T09:09:26.5122588Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5122707Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5122710Z 2025-12-04T09:09:26.5122875Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5123017Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5123121Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5123250Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5123349Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5123474Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5123579Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5123699Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5123802Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5123925Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5124024Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5124153Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5124254Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5124375Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5124480Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5124600Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5124703Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5124824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5124924Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5125051Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5125150Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5125275Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5125384Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5125507Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5125611Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5125735Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5125834Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5125966Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5126067Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5126192Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5126296Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5126418Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5126600Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5126785Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5126886Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5127015Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5127115Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5127242Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5127363Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5127484Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5127588Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5127709Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5127813Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5127942Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5128042Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5128164Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5128269Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5128392Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5128498Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5128623Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5128721Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5128851Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5128950Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5129076Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5129185Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5129317Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5129421Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5129542Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5129642Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5129769Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5129868Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5130000Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5130080Z Traceback (most recent call last): 2025-12-04T09:09:26.5130287Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5130366Z check_stats(expected) 2025-12-04T09:09:26.5130541Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5130613Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5130929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5131051Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5131129Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5131139Z 2025-12-04T09:09:26.5131206Z Expected 196608 but got 0. 2025-12-04T09:09:26.5131275Z Absolute difference: 196608 2025-12-04T09:09:26.5131346Z Relative difference: 1.0 2025-12-04T09:09:26.5131349Z 2025-12-04T09:09:26.5131473Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5131596Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5131682Z 2025-12-04T09:09:26.5131848Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5132077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5132196Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5132324Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5132425Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5132555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5132653Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5132774Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5132879Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5133002Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5133109Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5133305Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5133407Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5133530Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5133628Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5133753Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5133854Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5133973Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5134076Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5134199Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5134296Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5134433Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5134536Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5134658Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5134764Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5134885Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5134984Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5135115Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5135222Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5135352Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5135451Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5135573Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5135681Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5135805Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5135906Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5136032Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5136131Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5136269Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5136371Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5136492Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5136596Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5136720Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5136920Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5137111Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5137211Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5137336Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5137434Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5137555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5137656Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5137777Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5137876Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5138002Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5138101Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5138233Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5138337Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5138458Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5138562Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5138694Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5138794Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5138920Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5139019Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5139146Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5139248Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5139386Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5139468Z Traceback (most recent call last): 2025-12-04T09:09:26.5139676Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5139746Z check_stats(expected) 2025-12-04T09:09:26.5139920Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5139991Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5140311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5140431Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5140515Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5140519Z 2025-12-04T09:09:26.5146685Z Expected 196608 but got 0. 2025-12-04T09:09:26.5146791Z Absolute difference: 196608 2025-12-04T09:09:26.5146868Z Relative difference: 1.0 2025-12-04T09:09:26.5146881Z 2025-12-04T09:09:26.5147042Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5147182Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5147186Z 2025-12-04T09:09:26.5147369Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5147507Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5147623Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5147757Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5147862Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5147990Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5148098Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5148220Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5148450Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5148646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5148746Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5148871Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5148967Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5149085Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5149189Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5149309Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5149412Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5149539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5149651Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5149782Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5149883Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5150003Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5150115Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5150241Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5150347Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5150469Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5150569Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5150695Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5150794Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5150919Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5151022Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5151141Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5151242Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5151360Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5151457Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5151580Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5151675Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5151794Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5151899Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5152018Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5152127Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5152249Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5152347Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5152473Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5152569Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5152689Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5152790Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5152909Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5153011Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5153132Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5153324Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5153520Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5153623Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5153743Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5153849Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5153966Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5154067Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5154186Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5154285Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5154410Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5154519Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5154644Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5154748Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5154865Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5154961Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5155090Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5155169Z Traceback (most recent call last): 2025-12-04T09:09:26.5155404Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5155477Z check_stats(expected) 2025-12-04T09:09:26.5155657Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5155738Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5156073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5156211Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5156299Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5156303Z 2025-12-04T09:09:26.5156371Z Expected 196608 but got 0. 2025-12-04T09:09:26.5156445Z Absolute difference: 196608 2025-12-04T09:09:26.5156514Z Relative difference: 1.0 2025-12-04T09:09:26.5156517Z 2025-12-04T09:09:26.5156644Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5156768Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5156772Z 2025-12-04T09:09:26.5156941Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5157078Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5157189Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5157317Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5157421Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5157546Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5157650Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5157772Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5157873Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5157999Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5158095Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5158215Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5158318Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5158435Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5158638Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5158831Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5158935Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5159056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5159150Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5159266Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5159364Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5159482Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5159578Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5159698Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5159794Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5159917Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5160015Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5160134Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5160231Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5160349Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5160445Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5160565Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5160660Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5160785Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5160880Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5161000Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5161103Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5161219Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5161313Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5161433Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5161538Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5161662Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5161762Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5161878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5161975Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5162092Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5162190Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5162311Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5162406Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5162525Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5162623Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5162742Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5162840Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5162958Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5163063Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5163187Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5163392Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5163575Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5163684Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5163806Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5163906Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5164025Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5164121Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5164242Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5164336Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5164453Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5164550Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5164681Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5164759Z Traceback (most recent call last): 2025-12-04T09:09:26.5164972Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5165040Z check_stats(expected) 2025-12-04T09:09:26.5165214Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5165285Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5165606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5165737Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5165815Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5165819Z 2025-12-04T09:09:26.5165888Z Expected 196608 but got 0. 2025-12-04T09:09:26.5165955Z Absolute difference: 196608 2025-12-04T09:09:26.5166025Z Relative difference: 1.0 2025-12-04T09:09:26.5166028Z 2025-12-04T09:09:26.5166155Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5166276Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5166280Z 2025-12-04T09:09:26.5166438Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5166561Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5166662Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5166784Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5166879Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5166998Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5167098Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5167217Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5167319Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5167442Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5167537Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5167658Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5167753Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5167869Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5167965Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5168081Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5168176Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5168293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5168476Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5168660Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5168758Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5168875Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5168973Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5169090Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5169186Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5169304Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5169398Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5169527Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5169624Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5169747Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5169847Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5169966Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5170063Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5170181Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5170276Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5170393Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5170486Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5170603Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5170698Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5170815Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5170917Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5171035Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5171129Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5171249Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5171343Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5171460Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5171557Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5171674Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5171772Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5171889Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5171987Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5172116Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5172217Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5172334Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5172430Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5172547Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5172642Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5172759Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5172853Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5172971Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5173065Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5173438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5173546Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5173670Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5173771Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5173892Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5173997Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5174121Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5174217Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5174345Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5174419Z Traceback (most recent call last): 2025-12-04T09:09:26.5174639Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5174707Z check_stats(expected) 2025-12-04T09:09:26.5174884Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5174956Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5175272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5175394Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5175471Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5175475Z 2025-12-04T09:09:26.5175542Z Expected 196608 but got 0. 2025-12-04T09:09:26.5175609Z Absolute difference: 196608 2025-12-04T09:09:26.5175672Z Relative difference: 1.0 2025-12-04T09:09:26.5175677Z 2025-12-04T09:09:26.5175799Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5175919Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5175925Z 2025-12-04T09:09:26.5176089Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5176212Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5176312Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5176435Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5176533Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5176654Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5176751Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5176869Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5176967Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5177085Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5177183Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5177305Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5177402Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5177526Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5177624Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5177743Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5177846Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5177966Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5178066Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5178188Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5178372Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5178572Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5178670Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5178787Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5178884Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5179003Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5179107Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5179232Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5179331Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5179453Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5179554Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5179682Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5179794Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5179918Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5180018Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5180141Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5180237Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5180362Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5180458Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5180579Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5180678Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5180802Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5180901Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5181026Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5181122Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5181243Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5181341Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5181458Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5181557Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5181675Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5181772Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5181896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5181996Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5182121Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5182218Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5182337Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5182437Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5182555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5182653Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5182775Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5182871Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5182993Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5183089Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5183286Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5183457Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5183583Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5183679Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5183803Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5183901Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5184027Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5184122Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5184250Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5184326Z Traceback (most recent call last): 2025-12-04T09:09:26.5184531Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5184601Z check_stats(expected) 2025-12-04T09:09:26.5184780Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5184852Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5185177Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5185298Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5185378Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5185382Z 2025-12-04T09:09:26.5185450Z Expected 196608 but got 0. 2025-12-04T09:09:26.5185516Z Absolute difference: 196608 2025-12-04T09:09:26.5185581Z Relative difference: 1.0 2025-12-04T09:09:26.5185585Z 2025-12-04T09:09:26.5185710Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5185829Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5185835Z 2025-12-04T09:09:26.5186000Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5186125Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5186230Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5186356Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5186453Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5186574Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5186677Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5186798Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5186898Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5187018Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5187119Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5187246Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5187344Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5187463Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5187565Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5187684Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5187785Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5187905Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5188002Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5188127Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5188223Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5188427Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5188592Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5188716Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5188819Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5188939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5189035Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5189160Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5189258Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5189377Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5189479Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5189601Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5189702Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5189822Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5189930Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5190056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5190151Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5190270Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5190373Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5190492Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5190592Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5190710Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5190809Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5190933Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5191030Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5191150Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5191250Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5191369Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5191468Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5191586Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5191682Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5191806Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5191907Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5192028Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5192129Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5192248Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5192347Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5192466Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5192564Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5192688Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5192784Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5192907Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5193008Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5193210Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5193373Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5193494Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5193590Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5193713Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5193808Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5193925Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5194036Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5194156Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5194256Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5194383Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5194458Z Traceback (most recent call last): 2025-12-04T09:09:26.5194667Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5194733Z check_stats(expected) 2025-12-04T09:09:26.5194899Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5194975Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5195284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5195406Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5195483Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5195487Z 2025-12-04T09:09:26.5195550Z Expected 196608 but got 0. 2025-12-04T09:09:26.5195620Z Absolute difference: 196608 2025-12-04T09:09:26.5195685Z Relative difference: 1.0 2025-12-04T09:09:26.5195691Z 2025-12-04T09:09:26.5195813Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5195937Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5195940Z 2025-12-04T09:09:26.5196097Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5196227Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5196332Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5196455Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5196558Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5196681Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5196778Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5196904Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5197004Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5197131Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5197229Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5197348Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5197452Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5197574Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5197671Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5197802Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5197899Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5198020Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5198116Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5198324Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5198506Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5198627Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5198724Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5198848Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5198946Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5199075Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5199178Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5199299Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5199398Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5199520Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5199624Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5199748Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5199844Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5199975Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5200079Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5200198Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5200299Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5200418Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5200515Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5200637Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5200739Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5200870Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5200971Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5201091Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5201194Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5201313Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5201411Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5201532Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5201631Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5201753Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5201853Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5201978Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5202080Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5202200Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5202299Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5202422Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5202519Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5202641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5202739Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5202858Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5202965Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5203164Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5203330Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5203457Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5203554Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5203678Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5203775Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5203896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5203996Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5204114Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5204210Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5204336Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5204435Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5204787Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5204899Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5205031Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5205110Z Traceback (most recent call last): 2025-12-04T09:09:26.5205320Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5205388Z check_stats(expected) 2025-12-04T09:09:26.5205566Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5205637Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5205960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5206091Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5206174Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5206181Z 2025-12-04T09:09:26.5206250Z Expected 196608 but got 0. 2025-12-04T09:09:26.5206322Z Absolute difference: 196608 2025-12-04T09:09:26.5206388Z Relative difference: 1.0 2025-12-04T09:09:26.5206391Z 2025-12-04T09:09:26.5206522Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5206644Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5206647Z 2025-12-04T09:09:26.5206807Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5206937Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5207039Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5207164Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5207266Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5207389Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5207498Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5207624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5207727Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5207857Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5207957Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5208092Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5208192Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5208312Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5208413Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5208668Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5208864Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5208997Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5209094Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5209217Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5209314Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5209433Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5209533Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5209659Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5209766Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5209898Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5210003Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5210128Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5210230Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5210351Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5210448Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5210566Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5210663Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5210783Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5210879Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5211001Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5211105Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5211225Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5211326Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5211444Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5211539Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5211661Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5211757Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5211875Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5211977Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5212096Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5212206Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5212330Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5212428Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5212550Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5212645Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5212764Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5212863Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5212981Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5213079Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5213251Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5213350Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5213555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5213712Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5213834Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5213933Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5214064Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5214166Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5214285Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5214380Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5214503Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5214597Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5214716Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5214818Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5214943Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5215043Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5215162Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5215258Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5215381Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5215478Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5215596Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5215698Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5215825Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5215908Z Traceback (most recent call last): 2025-12-04T09:09:26.5216118Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5216188Z check_stats(expected) 2025-12-04T09:09:26.5216381Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5216456Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5216781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5216910Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5216991Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5216995Z 2025-12-04T09:09:26.5217066Z Expected 196608 but got 0. 2025-12-04T09:09:26.5217135Z Absolute difference: 196608 2025-12-04T09:09:26.5217199Z Relative difference: 1.0 2025-12-04T09:09:26.5217203Z 2025-12-04T09:09:26.5217334Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5217462Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5217466Z 2025-12-04T09:09:26.5217635Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5217764Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5217869Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5217994Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5218095Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5218214Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5218316Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5218436Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5218537Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5218755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5218937Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5219066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5219162Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5219282Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5219385Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5219504Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5219605Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5219724Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5219822Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5219945Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5220045Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5220167Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5220268Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5220388Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5220489Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5220609Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5220705Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5220840Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5220939Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5221058Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5221160Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5221285Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5221386Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5221505Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5221601Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5221724Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5221821Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5221939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5222040Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5222157Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5222258Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5222379Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5222475Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5222599Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5222696Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5222813Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5222914Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5223034Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5223134Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5223255Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5223350Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5223472Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5223718Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5223843Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5223952Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5224074Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5224173Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5224293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5224389Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5224511Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5224607Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5224725Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5224828Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5224949Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5225049Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5225170Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5225265Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5225389Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5225486Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5225605Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5225706Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5225824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5225919Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5226047Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5226142Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5226264Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5226360Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5226481Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5226583Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5226711Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5226784Z Traceback (most recent call last): 2025-12-04T09:09:26.5226995Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5227062Z check_stats(expected) 2025-12-04T09:09:26.5227235Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5227311Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5227631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5227755Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5227837Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5227840Z 2025-12-04T09:09:26.5227907Z Expected 196608 but got 0. 2025-12-04T09:09:26.5227977Z Absolute difference: 196608 2025-12-04T09:09:26.5228043Z Relative difference: 1.0 2025-12-04T09:09:26.5228046Z 2025-12-04T09:09:26.5228175Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5228294Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5228297Z 2025-12-04T09:09:26.5228453Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5228581Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5228831Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5228962Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5229061Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5229181Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5229281Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5229399Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5229496Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5229617Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5229713Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5229834Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5229933Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5230054Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5230154Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5230273Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5230369Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5230491Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5230588Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5230711Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5230807Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5230926Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5231024Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5231149Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5231249Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5231372Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5231468Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5231593Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5231690Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5231808Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5231909Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5232028Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5232124Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5232250Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5232348Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5232471Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5232566Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5232682Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5232782Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5232900Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5232994Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5233118Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5233217Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5233349Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5233524Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5233710Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5233819Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5233939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5234043Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5234171Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5234268Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5234390Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5234484Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5234602Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5234705Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5234826Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5234923Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5235040Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5235135Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5235257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5235352Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5235470Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5235571Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5235689Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5235788Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5235911Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5236010Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5236131Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5236226Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5236343Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5236442Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5236560Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5236660Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5236779Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5236874Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5237001Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5237098Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5237217Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5237317Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5237434Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5237533Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5237661Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5237736Z Traceback (most recent call last): 2025-12-04T09:09:26.5237943Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5238010Z check_stats(expected) 2025-12-04T09:09:26.5238181Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5238358Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5238737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5238861Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5238938Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5238942Z 2025-12-04T09:09:26.5239016Z Expected 196608 but got 0. 2025-12-04T09:09:26.5239088Z Absolute difference: 196608 2025-12-04T09:09:26.5239152Z Relative difference: 1.0 2025-12-04T09:09:26.5239156Z 2025-12-04T09:09:26.5239281Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5239401Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5239405Z 2025-12-04T09:09:26.5239565Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5239690Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5239793Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5239916Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5240018Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5240136Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5240231Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5240353Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5240448Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5240570Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5240666Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5240784Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5240883Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5241005Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5241108Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5241239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5241335Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5241455Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5241553Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5241670Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5241770Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5241888Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5241984Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5242109Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5242209Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5242329Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5242427Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5242544Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5242643Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5242759Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5242853Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5242974Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5243069Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5243192Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5243365Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5243552Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5243657Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5243778Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5243875Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5243999Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5244094Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5244216Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5244313Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5244431Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5244534Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5244654Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5244750Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5244870Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5244966Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5245087Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5245192Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5245312Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5245412Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5245532Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5245626Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5245763Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5245861Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5245984Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5246079Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5246198Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5246297Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5246415Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5246511Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5246633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5246729Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5246850Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5246948Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5247070Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5247171Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5247290Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5247387Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5247509Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5247605Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5247727Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5247826Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5247944Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5248131Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5248313Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5248410Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5248531Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5248626Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5248745Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5248845Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5248972Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5249047Z Traceback (most recent call last): 2025-12-04T09:09:26.5249249Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5249314Z check_stats(expected) 2025-12-04T09:09:26.5249488Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5249561Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5249876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5250006Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5250087Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5250091Z 2025-12-04T09:09:26.5250158Z Expected 196608 but got 0. 2025-12-04T09:09:26.5250225Z Absolute difference: 196608 2025-12-04T09:09:26.5250288Z Relative difference: 1.0 2025-12-04T09:09:26.5250292Z 2025-12-04T09:09:26.5250417Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5250543Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5250547Z 2025-12-04T09:09:26.5250705Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5250828Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5250930Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5251051Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5251148Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5251269Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5251373Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5251492Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5251592Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5251709Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5251804Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5251926Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5252027Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5252144Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5252243Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5252372Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5252472Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5252590Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5252686Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5252814Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5252918Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5253044Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5253307Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5253516Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5253626Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5253747Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5253854Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5253983Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5254079Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5254199Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5254301Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5254419Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5254522Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5254646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5254743Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5254865Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5254961Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5255081Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5255181Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5255301Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5255402Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5255521Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5255618Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5255743Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5255842Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5255963Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5256064Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5256187Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5256285Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5256405Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5256504Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5256626Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5256722Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5256839Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5256942Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5257062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5257163Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5257280Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5257375Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5257496Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5257592Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5257710Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5257809Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5257928Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5258121Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5258306Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5258404Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5258527Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5258625Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5258745Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5258844Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5258961Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5259058Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5259176Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5259281Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5259404Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5259503Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5259622Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5259721Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5259839Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5259936Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5260055Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5260149Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5260270Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5260366Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5260498Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5260576Z Traceback (most recent call last): 2025-12-04T09:09:26.5260784Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5260855Z check_stats(expected) 2025-12-04T09:09:26.5261026Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5261098Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5261429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5261561Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5261642Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5261645Z 2025-12-04T09:09:26.5261713Z Expected 196608 but got 0. 2025-12-04T09:09:26.5261780Z Absolute difference: 196608 2025-12-04T09:09:26.5261847Z Relative difference: 1.0 2025-12-04T09:09:26.5261853Z 2025-12-04T09:09:26.5261984Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5262107Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5262110Z 2025-12-04T09:09:26.5262273Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5262400Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5262501Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5262626Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5262725Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5262847Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5262944Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5263062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5263238Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5263423Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5263522Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5263641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5263738Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5263861Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5263957Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5264078Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5264177Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5264307Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5264409Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5264535Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5264634Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5264757Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5264854Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5264972Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5265071Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5265190Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5265286Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5265411Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5265508Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5265635Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5265733Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5265852Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5265951Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5266071Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5266168Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5266291Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5266387Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5266508Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5266603Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5266722Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5266823Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5266955Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5267054Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5267176Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5267272Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5267394Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5267492Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5267611Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5267711Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5267830Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5268001Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5268184Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5268282Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5268404Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5268500Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5268618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5268717Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5268844Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5268941Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5269061Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5269157Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5269282Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5269381Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5269501Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5269600Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5269719Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5269815Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5269950Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5270048Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5270170Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5270266Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5270383Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5270485Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5270606Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5270702Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5270823Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5270919Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5271041Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5271136Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5271255Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5271356Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5271475Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5271574Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5271710Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5271810Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5271931Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5272027Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5272155Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5272231Z Traceback (most recent call last): 2025-12-04T09:09:26.5272433Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5272499Z check_stats(expected) 2025-12-04T09:09:26.5272671Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5272742Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5273145Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5273349Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5273428Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5273433Z 2025-12-04T09:09:26.5273500Z Expected 196608 but got 0. 2025-12-04T09:09:26.5273568Z Absolute difference: 196608 2025-12-04T09:09:26.5273631Z Relative difference: 1.0 2025-12-04T09:09:26.5273635Z 2025-12-04T09:09:26.5273759Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5273876Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5273880Z 2025-12-04T09:09:26.5274039Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5274164Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5274266Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5274394Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5274494Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5274615Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5274715Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5274835Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5274940Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5275068Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5275167Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5275288Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5275383Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5275501Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5275602Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5275723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5275822Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5275942Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5276038Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5276159Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5276255Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5276373Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5276472Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5276592Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5276698Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5276819Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5276916Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5277038Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5277135Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5277253Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5277353Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5277470Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5277569Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5277687Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5277783Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5277981Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5278154Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5278281Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5278382Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5278510Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5278613Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5278741Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5278837Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5278959Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5279054Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5279175Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5279277Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5279397Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5279496Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5279614Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5279711Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5279832Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5279928Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5280047Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5280152Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5280271Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5280370Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5280494Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5280601Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5280724Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5280820Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5280938Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5281037Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5281155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5281254Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5281373Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5281473Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5281597Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5281695Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5281813Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5281912Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5282029Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5282130Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5282248Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5282344Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5282466Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5282562Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5282767Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5282932Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5283053Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5283153Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5283271Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5283367Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5283492Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5283588Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5283705Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5283805Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5283923Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5284029Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5284157Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5284231Z Traceback (most recent call last): 2025-12-04T09:09:26.5284442Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5284507Z check_stats(expected) 2025-12-04T09:09:26.5284673Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5284751Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5285064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5285186Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5285265Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5285271Z 2025-12-04T09:09:26.5285334Z Expected 196608 but got 0. 2025-12-04T09:09:26.5285407Z Absolute difference: 196608 2025-12-04T09:09:26.5285476Z Relative difference: 1.0 2025-12-04T09:09:26.5285480Z 2025-12-04T09:09:26.5285613Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5285735Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5285739Z 2025-12-04T09:09:26.5285898Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5286025Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5286128Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5286249Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5286352Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5286469Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5286572Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5286698Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5286797Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5286921Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5287020Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5287137Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5287237Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5287358Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5287455Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5287578Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5287676Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5287874Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5288038Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5288161Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5288262Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5288381Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5288479Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5288609Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5288707Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5288829Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5288926Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5289047Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5289150Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5289270Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5289367Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5289488Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5289595Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5289721Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5289820Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5289941Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5290040Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5290160Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5290261Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5290387Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5290487Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5290608Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5290705Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5290824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5290924Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5291044Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5291143Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5291268Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5291366Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5291491Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5291589Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5291709Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5291809Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5291927Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5292025Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5292148Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5292243Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5292366Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5292463Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5292675Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5292843Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5292968Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5293063Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5293244Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5293345Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5293469Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5293563Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5293682Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5293781Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5293899Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5293998Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5294121Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5294217Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5294338Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5294433Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5294553Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5294653Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5294770Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5294866Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5294989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5295089Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5295213Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5295309Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5295428Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5295530Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5295650Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5295744Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5295868Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5295965Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5296091Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5296188Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5296322Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5296404Z Traceback (most recent call last): 2025-12-04T09:09:26.5296614Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5296680Z check_stats(expected) 2025-12-04T09:09:26.5296854Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5296926Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5297245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5297366Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5297444Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5297448Z 2025-12-04T09:09:26.5297518Z Expected 196608 but got 0. 2025-12-04T09:09:26.5297589Z Absolute difference: 196608 2025-12-04T09:09:26.5297733Z Relative difference: 1.0 2025-12-04T09:09:26.5297737Z 2025-12-04T09:09:26.5297946Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5298069Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5298072Z 2025-12-04T09:09:26.5298236Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5298359Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5298459Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5298585Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5298682Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5298818Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5298919Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5299040Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5299144Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5299268Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5299365Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5299492Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5299586Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5299705Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5299805Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5299924Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5300025Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5300143Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5300242Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5300370Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5300466Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5300586Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5300687Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5300805Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5300907Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5301027Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5301124Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5301246Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5301343Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5301466Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5301568Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5301689Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5301793Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5301916Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5302015Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5302141Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5302236Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5302357Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5302460Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5302657Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5302819Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5302941Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5303039Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5303164Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5303260Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5303377Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5303475Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5303592Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5303695Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5303813Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5303911Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5304035Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5304143Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5304266Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5304365Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5304485Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5304865Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5305065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5305177Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5305302Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5305405Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5305530Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5305632Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5305753Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5305854Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5305974Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5306071Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5306196Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5306293Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5306412Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5306510Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5306633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5306740Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5306858Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5306956Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5307077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5307173Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5307291Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5307392Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5307509Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5307610Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5307730Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5308069Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5308206Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5308306Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5308428Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5308531Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5308658Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5308763Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5308882Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5308980Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5309101Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5309202Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5309336Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5309414Z Traceback (most recent call last): 2025-12-04T09:09:26.5309629Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5309700Z check_stats(expected) 2025-12-04T09:09:26.5309870Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5309942Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5310265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5310392Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5310471Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5310480Z 2025-12-04T09:09:26.5310546Z Expected 196608 but got 0. 2025-12-04T09:09:26.5310618Z Absolute difference: 196608 2025-12-04T09:09:26.5310685Z Relative difference: 1.0 2025-12-04T09:09:26.5310689Z 2025-12-04T09:09:26.5310818Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5310950Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5310954Z 2025-12-04T09:09:26.5311117Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5311240Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5311346Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5311465Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5311562Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5311686Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5311784Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5311908Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5312013Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5312137Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5312238Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5312357Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5312454Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5312574Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5312680Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5312802Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5312901Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5313018Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5313281Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5313405Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5313513Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5313641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5313737Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5313856Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5313955Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5314074Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5314173Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5314292Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5314389Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5314512Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5314607Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5314727Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5314827Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5314947Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5315045Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5315162Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5315260Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5315381Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5315479Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5315606Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5315717Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5315843Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5315939Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5316060Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5316153Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5316273Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5316371Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5316488Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5316587Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5316708Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5316807Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5316930Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5317025Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5317148Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5317244Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5317364Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5317465Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5317585Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5317695Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5317818Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5317988Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5318183Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5318285Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5318406Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5318505Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5318624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5318719Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5318841Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5318945Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5319067Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5319166Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5319290Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5319390Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5319510Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5319607Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5319730Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5319828Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5319949Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5320045Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5320164Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5320266Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5320391Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5320496Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5324634Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5324799Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5324956Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5325070Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5325204Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5325312Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5325440Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5325546Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5325678Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5325783Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5325909Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5326011Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5326149Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5326227Z Traceback (most recent call last): 2025-12-04T09:09:26.5326454Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5326531Z check_stats(expected) 2025-12-04T09:09:26.5326721Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5326798Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5327142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5327394Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5327552Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5327559Z 2025-12-04T09:09:26.5327633Z Expected 196608 but got 0. 2025-12-04T09:09:26.5327703Z Absolute difference: 196608 2025-12-04T09:09:26.5327773Z Relative difference: 1.0 2025-12-04T09:09:26.5327777Z 2025-12-04T09:09:26.5327908Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5328032Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5328035Z 2025-12-04T09:09:26.5328202Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5328337Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5328449Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5328575Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5328687Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5328821Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5328919Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5329050Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5329154Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5329276Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5329374Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5329493Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5329590Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5329717Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5329815Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5329937Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5330042Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5330175Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5330275Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5330393Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5330490Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5330612Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5330709Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5330828Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5330931Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5331055Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5331158Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5331277Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5331375Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5331502Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5331600Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5331720Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5331821Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5331941Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5332040Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5332163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5332332Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5332542Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5332646Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5332768Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5332871Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5332991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5333094Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5333302Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5333406Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5333534Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5333638Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5333762Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5333864Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5333984Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5334083Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5334202Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5334296Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5334420Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5334517Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5334640Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5334742Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5334863Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5334968Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5335098Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5335201Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5335324Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5335420Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5335539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5335645Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5335765Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5335866Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5335991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5336094Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5336222Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5336320Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5336445Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5336545Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5336667Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5336766Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5336885Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5336982Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5337103Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5337279Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5337469Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5337576Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5337695Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5337794Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5337914Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5338011Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5338131Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5338227Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5338345Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5338444Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5338571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5338670Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5338789Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5338895Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5339017Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5339115Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5339233Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5339331Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5339460Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:09:26.5339539Z Traceback (most recent call last): 2025-12-04T09:09:26.5339757Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:09:26.5339827Z check_stats(expected) 2025-12-04T09:09:26.5340016Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:09:26.5340102Z self.assertEqual(v, stats[k]) 2025-12-04T09:09:26.5340451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5340585Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5340666Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5340670Z 2025-12-04T09:09:26.5340738Z Expected 196608 but got 0. 2025-12-04T09:09:26.5340807Z Absolute difference: 196608 2025-12-04T09:09:26.5340875Z Relative difference: 1.0 2025-12-04T09:09:26.5340879Z 2025-12-04T09:09:26.5341013Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5341136Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5341143Z 2025-12-04T09:09:26.5341307Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5341436Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5341542Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5341663Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5341767Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5341886Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5341995Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5342117Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5342216Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5342338Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5342510Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5342703Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5342810Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5342929Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5343029Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5343148Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5343247Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5343368Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5343464Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5343582Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5343681Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5343816Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5343919Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5344041Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5344138Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5344260Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5344356Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5344474Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5344573Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5344692Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5344790Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5344911Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5345011Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5345134Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5345230Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5345350Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5345449Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5345572Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5345671Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5345789Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5345887Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5346008Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5346109Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5346232Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5346331Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5346451Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5346549Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5346669Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5346765Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5346888Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5346984Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5347102Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5347289Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5347477Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5347577Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5347696Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5347791Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5347913Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5348009Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5348128Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5348228Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5348347Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5348445Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5348575Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5348684Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5348810Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5348907Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5349026Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5349125Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5349243Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5349341Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5349460Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5349556Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5349679Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5349778Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5349898Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5349998Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5350115Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5350213Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5350330Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5350426Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5350547Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5350641Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5350759Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5350876Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5350999Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5351099Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5351217Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5351314Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5351434Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5351530Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5351648Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5351748Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5351866Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5352041Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5352241Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5352339Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5352470Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:09:26.5352570Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:09:26.5352717Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5352793Z Traceback (most recent call last): 2025-12-04T09:09:26.5353031Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5353129Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5353445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5353565Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5353650Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5353656Z 2025-12-04T09:09:26.5353721Z Expected 0 but got 123. 2025-12-04T09:09:26.5353788Z Absolute difference: 123 2025-12-04T09:09:26.5353854Z Relative difference: inf 2025-12-04T09:09:26.5353858Z 2025-12-04T09:09:26.5353982Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5354139Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5354143Z 2025-12-04T09:09:26.5354296Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5354444Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5354521Z Traceback (most recent call last): 2025-12-04T09:09:26.5354751Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5354845Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5355157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5355276Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5355355Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5355359Z 2025-12-04T09:09:26.5355421Z Expected 0 but got 123. 2025-12-04T09:09:26.5355485Z Absolute difference: 123 2025-12-04T09:09:26.5355551Z Relative difference: inf 2025-12-04T09:09:26.5355554Z 2025-12-04T09:09:26.5355685Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5355845Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5355849Z 2025-12-04T09:09:26.5356003Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5356154Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5356235Z Traceback (most recent call last): 2025-12-04T09:09:26.5356470Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5356567Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5356883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5357006Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5357086Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5357089Z 2025-12-04T09:09:26.5357153Z Expected 0 but got 123. 2025-12-04T09:09:26.5357220Z Absolute difference: 123 2025-12-04T09:09:26.5357288Z Relative difference: inf 2025-12-04T09:09:26.5357291Z 2025-12-04T09:09:26.5357414Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5357564Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5357646Z 2025-12-04T09:09:26.5357801Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5358013Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5358092Z Traceback (most recent call last): 2025-12-04T09:09:26.5358322Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5358415Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5358730Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5358849Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5358928Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5358931Z 2025-12-04T09:09:26.5359002Z Expected 0 but got 123. 2025-12-04T09:09:26.5359067Z Absolute difference: 123 2025-12-04T09:09:26.5359134Z Relative difference: inf 2025-12-04T09:09:26.5359141Z 2025-12-04T09:09:26.5359268Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5359418Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5359424Z 2025-12-04T09:09:26.5359586Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5359736Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5359814Z Traceback (most recent call last): 2025-12-04T09:09:26.5360041Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5360132Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5360440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5360553Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5360629Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5360635Z 2025-12-04T09:09:26.5360697Z Expected 0 but got 123. 2025-12-04T09:09:26.5360763Z Absolute difference: 123 2025-12-04T09:09:26.5360830Z Relative difference: inf 2025-12-04T09:09:26.5360833Z 2025-12-04T09:09:26.5360956Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5361108Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5361112Z 2025-12-04T09:09:26.5361265Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5361411Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5361485Z Traceback (most recent call last): 2025-12-04T09:09:26.5361710Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5361801Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5362106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5362228Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5362303Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5362308Z 2025-12-04T09:09:26.5362370Z Expected 0 but got 123. 2025-12-04T09:09:26.5362433Z Absolute difference: 123 2025-12-04T09:09:26.5362501Z Relative difference: inf 2025-12-04T09:09:26.5362504Z 2025-12-04T09:09:26.5362625Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5362777Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5362781Z 2025-12-04T09:09:26.5362935Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5363080Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5363156Z Traceback (most recent call last): 2025-12-04T09:09:26.5363383Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5363621Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5363931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5364046Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5364121Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5364124Z 2025-12-04T09:09:26.5364189Z Expected 0 but got 123. 2025-12-04T09:09:26.5364253Z Absolute difference: 123 2025-12-04T09:09:26.5364320Z Relative difference: inf 2025-12-04T09:09:26.5364324Z 2025-12-04T09:09:26.5364444Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5364592Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5364595Z 2025-12-04T09:09:26.5364753Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5364908Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5364985Z Traceback (most recent call last): 2025-12-04T09:09:26.5365218Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5365313Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5365623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5365741Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5365821Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5365825Z 2025-12-04T09:09:26.5365894Z Expected 0 but got 123. 2025-12-04T09:09:26.5365962Z Absolute difference: 123 2025-12-04T09:09:26.5366027Z Relative difference: inf 2025-12-04T09:09:26.5366034Z 2025-12-04T09:09:26.5366157Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5366313Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5366316Z 2025-12-04T09:09:26.5366478Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5366628Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5366701Z Traceback (most recent call last): 2025-12-04T09:09:26.5366939Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5367033Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5367359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5367478Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5367556Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5367559Z 2025-12-04T09:09:26.5367627Z Expected 0 but got 123. 2025-12-04T09:09:26.5367699Z Absolute difference: 123 2025-12-04T09:09:26.5367764Z Relative difference: inf 2025-12-04T09:09:26.5367769Z 2025-12-04T09:09:26.5367898Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5368049Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5368052Z 2025-12-04T09:09:26.5368215Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5368365Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5368436Z Traceback (most recent call last): 2025-12-04T09:09:26.5368691Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5368785Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5369093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5369290Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5369429Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5369433Z 2025-12-04T09:09:26.5369500Z Expected 0 but got 123. 2025-12-04T09:09:26.5369565Z Absolute difference: 123 2025-12-04T09:09:26.5369629Z Relative difference: inf 2025-12-04T09:09:26.5369632Z 2025-12-04T09:09:26.5369757Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5369910Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5369913Z 2025-12-04T09:09:26.5370069Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5370224Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5370305Z Traceback (most recent call last): 2025-12-04T09:09:26.5370538Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5370632Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5370939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5371057Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5371135Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5371138Z 2025-12-04T09:09:26.5371204Z Expected 0 but got 123. 2025-12-04T09:09:26.5371269Z Absolute difference: 123 2025-12-04T09:09:26.5371333Z Relative difference: inf 2025-12-04T09:09:26.5371336Z 2025-12-04T09:09:26.5371458Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5371608Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5371611Z 2025-12-04T09:09:26.5371762Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5371908Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5371983Z Traceback (most recent call last): 2025-12-04T09:09:26.5372215Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5372306Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5372607Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5372726Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5372803Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5372806Z 2025-12-04T09:09:26.5372874Z Expected 0 but got 123. 2025-12-04T09:09:26.5372938Z Absolute difference: 123 2025-12-04T09:09:26.5373003Z Relative difference: inf 2025-12-04T09:09:26.5373007Z 2025-12-04T09:09:26.5373132Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5373371Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5373378Z 2025-12-04T09:09:26.5373531Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5373681Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5373752Z Traceback (most recent call last): 2025-12-04T09:09:26.5373984Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5374075Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5374379Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5374498Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5374575Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5374578Z 2025-12-04T09:09:26.5374644Z Expected 0 but got 123. 2025-12-04T09:09:26.5374708Z Absolute difference: 123 2025-12-04T09:09:26.5374865Z Relative difference: inf 2025-12-04T09:09:26.5374868Z 2025-12-04T09:09:26.5374993Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5375210Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5375216Z 2025-12-04T09:09:26.5375369Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5375522Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5375594Z Traceback (most recent call last): 2025-12-04T09:09:26.5375824Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5375917Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5376232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5376351Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5376431Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5376434Z 2025-12-04T09:09:26.5376498Z Expected 0 but got 123. 2025-12-04T09:09:26.5376569Z Absolute difference: 123 2025-12-04T09:09:26.5376634Z Relative difference: inf 2025-12-04T09:09:26.5376637Z 2025-12-04T09:09:26.5376763Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5376912Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5376915Z 2025-12-04T09:09:26.5377067Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5377226Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5377301Z Traceback (most recent call last): 2025-12-04T09:09:26.5377535Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5377632Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5377950Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5378079Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5378160Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5378163Z 2025-12-04T09:09:26.5378228Z Expected 0 but got 123. 2025-12-04T09:09:26.5378298Z Absolute difference: 123 2025-12-04T09:09:26.5378362Z Relative difference: inf 2025-12-04T09:09:26.5378365Z 2025-12-04T09:09:26.5378494Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5378654Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5378657Z 2025-12-04T09:09:26.5378811Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5378967Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5379041Z Traceback (most recent call last): 2025-12-04T09:09:26.5379278Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5379387Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5379703Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5379824Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5379900Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5379904Z 2025-12-04T09:09:26.5379967Z Expected 0 but got 123. 2025-12-04T09:09:26.5380036Z Absolute difference: 123 2025-12-04T09:09:26.5380101Z Relative difference: inf 2025-12-04T09:09:26.5380105Z 2025-12-04T09:09:26.5380228Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5380388Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5380390Z 2025-12-04T09:09:26.5380541Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5380833Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5380907Z Traceback (most recent call last): 2025-12-04T09:09:26.5381136Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5381234Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5381551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5381675Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5381751Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5381754Z 2025-12-04T09:09:26.5381817Z Expected 0 but got 123. 2025-12-04T09:09:26.5381885Z Absolute difference: 123 2025-12-04T09:09:26.5381949Z Relative difference: inf 2025-12-04T09:09:26.5381953Z 2025-12-04T09:09:26.5382075Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5382233Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5382237Z 2025-12-04T09:09:26.5382387Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5382535Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5382607Z Traceback (most recent call last): 2025-12-04T09:09:26.5382833Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5382929Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5383231Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5383349Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5383427Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5383433Z 2025-12-04T09:09:26.5383495Z Expected 0 but got 123. 2025-12-04T09:09:26.5383567Z Absolute difference: 123 2025-12-04T09:09:26.5383633Z Relative difference: inf 2025-12-04T09:09:26.5383636Z 2025-12-04T09:09:26.5383758Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5383913Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5383916Z 2025-12-04T09:09:26.5384065Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5384216Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5384287Z Traceback (most recent call last): 2025-12-04T09:09:26.5384512Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5384610Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5384912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5385031Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5385113Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5385116Z 2025-12-04T09:09:26.5385179Z Expected 0 but got 123. 2025-12-04T09:09:26.5385247Z Absolute difference: 123 2025-12-04T09:09:26.5385310Z Relative difference: inf 2025-12-04T09:09:26.5385313Z 2025-12-04T09:09:26.5385444Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5385600Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5385603Z 2025-12-04T09:09:26.5385752Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5385903Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5385974Z Traceback (most recent call last): 2025-12-04T09:09:26.5386200Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5386372Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5386761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5386884Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5386964Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5386967Z 2025-12-04T09:09:26.5387031Z Expected 0 but got 123. 2025-12-04T09:09:26.5387097Z Absolute difference: 123 2025-12-04T09:09:26.5387160Z Relative difference: inf 2025-12-04T09:09:26.5387164Z 2025-12-04T09:09:26.5387285Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5387434Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5387437Z 2025-12-04T09:09:26.5387589Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5387737Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5387814Z Traceback (most recent call last): 2025-12-04T09:09:26.5388043Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5388137Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5388442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5388557Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5388639Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5388642Z 2025-12-04T09:09:26.5388707Z Expected 0 but got 123. 2025-12-04T09:09:26.5388772Z Absolute difference: 123 2025-12-04T09:09:26.5388838Z Relative difference: inf 2025-12-04T09:09:26.5388842Z 2025-12-04T09:09:26.5388963Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5389113Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5389120Z 2025-12-04T09:09:26.5389278Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5389426Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5389500Z Traceback (most recent call last): 2025-12-04T09:09:26.5389727Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5389817Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5390122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5390237Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5390312Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5390318Z 2025-12-04T09:09:26.5390382Z Expected 0 but got 123. 2025-12-04T09:09:26.5390446Z Absolute difference: 123 2025-12-04T09:09:26.5390530Z Relative difference: inf 2025-12-04T09:09:26.5390533Z 2025-12-04T09:09:26.5390659Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5390808Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5390811Z 2025-12-04T09:09:26.5390963Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5391111Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5391185Z Traceback (most recent call last): 2025-12-04T09:09:26.5391412Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5391502Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5391811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5391926Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5392078Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5392081Z 2025-12-04T09:09:26.5392357Z Expected 0 but got 123. 2025-12-04T09:09:26.5392429Z Absolute difference: 123 2025-12-04T09:09:26.5392497Z Relative difference: inf 2025-12-04T09:09:26.5392500Z 2025-12-04T09:09:26.5392622Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5392773Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5392775Z 2025-12-04T09:09:26.5392928Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5393074Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5393146Z Traceback (most recent call last): 2025-12-04T09:09:26.5393375Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5393464Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5393776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5393890Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5393966Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5393969Z 2025-12-04T09:09:26.5394035Z Expected 0 but got 123. 2025-12-04T09:09:26.5394100Z Absolute difference: 123 2025-12-04T09:09:26.5394170Z Relative difference: inf 2025-12-04T09:09:26.5394176Z 2025-12-04T09:09:26.5394297Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5394458Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5394461Z 2025-12-04T09:09:26.5394618Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5394766Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5394840Z Traceback (most recent call last): 2025-12-04T09:09:26.5395072Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5395161Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5395473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5395588Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5395667Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5395670Z 2025-12-04T09:09:26.5395735Z Expected 0 but got 123. 2025-12-04T09:09:26.5395798Z Absolute difference: 123 2025-12-04T09:09:26.5395861Z Relative difference: inf 2025-12-04T09:09:26.5395865Z 2025-12-04T09:09:26.5395991Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5396142Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5396145Z 2025-12-04T09:09:26.5396301Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5396449Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5396522Z Traceback (most recent call last): 2025-12-04T09:09:26.5396754Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5396844Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5397149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5397263Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5397338Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5397341Z 2025-12-04T09:09:26.5397407Z Expected 0 but got 123. 2025-12-04T09:09:26.5397472Z Absolute difference: 123 2025-12-04T09:09:26.5397539Z Relative difference: inf 2025-12-04T09:09:26.5397674Z 2025-12-04T09:09:26.5397810Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5398025Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5398029Z 2025-12-04T09:09:26.5398183Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5398330Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5398401Z Traceback (most recent call last): 2025-12-04T09:09:26.5398632Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5398723Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5399026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5399148Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5399225Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5399231Z 2025-12-04T09:09:26.5399297Z Expected 0 but got 123. 2025-12-04T09:09:26.5399363Z Absolute difference: 123 2025-12-04T09:09:26.5399431Z Relative difference: inf 2025-12-04T09:09:26.5399434Z 2025-12-04T09:09:26.5399574Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5399725Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5399729Z 2025-12-04T09:09:26.5399882Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5400033Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5400105Z Traceback (most recent call last): 2025-12-04T09:09:26.5400335Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5400423Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5400726Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5400849Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5400924Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5400928Z 2025-12-04T09:09:26.5400994Z Expected 0 but got 123. 2025-12-04T09:09:26.5401058Z Absolute difference: 123 2025-12-04T09:09:26.5401124Z Relative difference: inf 2025-12-04T09:09:26.5401127Z 2025-12-04T09:09:26.5401250Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5401401Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5401404Z 2025-12-04T09:09:26.5401556Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5401707Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5401778Z Traceback (most recent call last): 2025-12-04T09:09:26.5402020Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5402117Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5402423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5402542Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5402618Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5402622Z 2025-12-04T09:09:26.5402688Z Expected 0 but got 123. 2025-12-04T09:09:26.5402754Z Absolute difference: 123 2025-12-04T09:09:26.5402817Z Relative difference: inf 2025-12-04T09:09:26.5402820Z 2025-12-04T09:09:26.5402944Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5403091Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5403094Z 2025-12-04T09:09:26.5403242Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5403468Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5403604Z Traceback (most recent call last): 2025-12-04T09:09:26.5403835Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5403925Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5404229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5404356Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5404438Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5404441Z 2025-12-04T09:09:26.5404716Z Expected 0 but got 123. 2025-12-04T09:09:26.5404841Z Absolute difference: 123 2025-12-04T09:09:26.5404952Z Relative difference: inf 2025-12-04T09:09:26.5404958Z 2025-12-04T09:09:26.5405166Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5405330Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5405334Z 2025-12-04T09:09:26.5405489Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5405643Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5405716Z Traceback (most recent call last): 2025-12-04T09:09:26.5405948Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5406040Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5406346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5406467Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5406542Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5406546Z 2025-12-04T09:09:26.5406609Z Expected 0 but got 123. 2025-12-04T09:09:26.5406681Z Absolute difference: 123 2025-12-04T09:09:26.5406745Z Relative difference: inf 2025-12-04T09:09:26.5406748Z 2025-12-04T09:09:26.5406875Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5407025Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5407029Z 2025-12-04T09:09:26.5407179Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5407332Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5407405Z Traceback (most recent call last): 2025-12-04T09:09:26.5407631Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5407724Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5408029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5408150Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5408225Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5408231Z 2025-12-04T09:09:26.5408294Z Expected 0 but got 123. 2025-12-04T09:09:26.5408361Z Absolute difference: 123 2025-12-04T09:09:26.5408426Z Relative difference: inf 2025-12-04T09:09:26.5408429Z 2025-12-04T09:09:26.5408549Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5408702Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5408706Z 2025-12-04T09:09:26.5408855Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5409002Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5409083Z Traceback (most recent call last): 2025-12-04T09:09:26.5409313Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5409561Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5409978Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5410108Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5410189Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5410193Z 2025-12-04T09:09:26.5410256Z Expected 0 but got 123. 2025-12-04T09:09:26.5410325Z Absolute difference: 123 2025-12-04T09:09:26.5410390Z Relative difference: inf 2025-12-04T09:09:26.5410393Z 2025-12-04T09:09:26.5410520Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5410677Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5410681Z 2025-12-04T09:09:26.5410847Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5411000Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5411077Z Traceback (most recent call last): 2025-12-04T09:09:26.5411314Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5411410Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5411719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5411839Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5411915Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5411918Z 2025-12-04T09:09:26.5411981Z Expected 0 but got 123. 2025-12-04T09:09:26.5412048Z Absolute difference: 123 2025-12-04T09:09:26.5412111Z Relative difference: inf 2025-12-04T09:09:26.5412115Z 2025-12-04T09:09:26.5412239Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5412392Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5412398Z 2025-12-04T09:09:26.5412550Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5412703Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5412776Z Traceback (most recent call last): 2025-12-04T09:09:26.5413003Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5413098Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5413474Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5413592Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5413670Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5413674Z 2025-12-04T09:09:26.5413738Z Expected 0 but got 123. 2025-12-04T09:09:26.5413804Z Absolute difference: 123 2025-12-04T09:09:26.5413868Z Relative difference: inf 2025-12-04T09:09:26.5413874Z 2025-12-04T09:09:26.5413995Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5414153Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5414156Z 2025-12-04T09:09:26.5414308Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5414458Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5414531Z Traceback (most recent call last): 2025-12-04T09:09:26.5414757Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5414851Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5415155Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5415271Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5415429Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5415432Z 2025-12-04T09:09:26.5415496Z Expected 0 but got 123. 2025-12-04T09:09:26.5415653Z Absolute difference: 123 2025-12-04T09:09:26.5415724Z Relative difference: inf 2025-12-04T09:09:26.5415727Z 2025-12-04T09:09:26.5415851Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5416003Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5416006Z 2025-12-04T09:09:26.5416156Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5416306Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5416381Z Traceback (most recent call last): 2025-12-04T09:09:26.5416609Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5416701Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5417005Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5417129Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5417209Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5417212Z 2025-12-04T09:09:26.5417274Z Expected 0 but got 123. 2025-12-04T09:09:26.5417339Z Absolute difference: 123 2025-12-04T09:09:26.5417405Z Relative difference: inf 2025-12-04T09:09:26.5417408Z 2025-12-04T09:09:26.5417530Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5417684Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5417688Z 2025-12-04T09:09:26.5417840Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5417985Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5418060Z Traceback (most recent call last): 2025-12-04T09:09:26.5418294Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5418391Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5418701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5418819Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5418897Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5418901Z 2025-12-04T09:09:26.5418964Z Expected 0 but got 123. 2025-12-04T09:09:26.5419029Z Absolute difference: 123 2025-12-04T09:09:26.5419097Z Relative difference: inf 2025-12-04T09:09:26.5419100Z 2025-12-04T09:09:26.5419221Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5419373Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5419376Z 2025-12-04T09:09:26.5419526Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5419678Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5419752Z Traceback (most recent call last): 2025-12-04T09:09:26.5419978Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5420073Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5420379Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5420494Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5420577Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5420581Z 2025-12-04T09:09:26.5420644Z Expected 0 but got 123. 2025-12-04T09:09:26.5420710Z Absolute difference: 123 2025-12-04T09:09:26.5420776Z Relative difference: inf 2025-12-04T09:09:26.5420779Z 2025-12-04T09:09:26.5420909Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5421139Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5421205Z 2025-12-04T09:09:26.5421359Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5421504Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5421578Z Traceback (most recent call last): 2025-12-04T09:09:26.5421805Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5421899Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5422205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5422323Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5422404Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5422407Z 2025-12-04T09:09:26.5422473Z Expected 0 but got 123. 2025-12-04T09:09:26.5422547Z Absolute difference: 123 2025-12-04T09:09:26.5422616Z Relative difference: inf 2025-12-04T09:09:26.5422623Z 2025-12-04T09:09:26.5422744Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5422892Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5422896Z 2025-12-04T09:09:26.5423046Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5423192Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5423265Z Traceback (most recent call last): 2025-12-04T09:09:26.5423492Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5423585Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5423890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5424010Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5424090Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5424093Z 2025-12-04T09:09:26.5424156Z Expected 0 but got 123. 2025-12-04T09:09:26.5424220Z Absolute difference: 123 2025-12-04T09:09:26.5424291Z Relative difference: inf 2025-12-04T09:09:26.5424294Z 2025-12-04T09:09:26.5424414Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5424562Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5424565Z 2025-12-04T09:09:26.5424716Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5424861Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5424936Z Traceback (most recent call last): 2025-12-04T09:09:26.5425162Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5425253Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5425561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5425676Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5425753Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5425757Z 2025-12-04T09:09:26.5425820Z Expected 0 but got 123. 2025-12-04T09:09:26.5425883Z Absolute difference: 123 2025-12-04T09:09:26.5425949Z Relative difference: inf 2025-12-04T09:09:26.5425952Z 2025-12-04T09:09:26.5426071Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5426218Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5426222Z 2025-12-04T09:09:26.5426371Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5426517Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5426675Z Traceback (most recent call last): 2025-12-04T09:09:26.5426969Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5427062Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5427366Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5427481Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5427557Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5427561Z 2025-12-04T09:09:26.5427624Z Expected 0 but got 123. 2025-12-04T09:09:26.5427689Z Absolute difference: 123 2025-12-04T09:09:26.5427755Z Relative difference: inf 2025-12-04T09:09:26.5427758Z 2025-12-04T09:09:26.5427878Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5428034Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5428043Z 2025-12-04T09:09:26.5428199Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5428347Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5428420Z Traceback (most recent call last): 2025-12-04T09:09:26.5428644Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5428737Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5429043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5429159Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5429236Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5429240Z 2025-12-04T09:09:26.5429302Z Expected 0 but got 123. 2025-12-04T09:09:26.5429365Z Absolute difference: 123 2025-12-04T09:09:26.5429433Z Relative difference: inf 2025-12-04T09:09:26.5429436Z 2025-12-04T09:09:26.5429560Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5429708Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5429711Z 2025-12-04T09:09:26.5429862Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5430009Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5430083Z Traceback (most recent call last): 2025-12-04T09:09:26.5430312Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5430402Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5430706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5430821Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5430899Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5430905Z 2025-12-04T09:09:26.5430970Z Expected 0 but got 123. 2025-12-04T09:09:26.5431037Z Absolute difference: 123 2025-12-04T09:09:26.5431103Z Relative difference: inf 2025-12-04T09:09:26.5431106Z 2025-12-04T09:09:26.5431225Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5431373Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5431376Z 2025-12-04T09:09:26.5431529Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5431675Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5431750Z Traceback (most recent call last): 2025-12-04T09:09:26.5431978Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5432069Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5432514Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5432630Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5432706Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5432709Z 2025-12-04T09:09:26.5432775Z Expected 0 but got 123. 2025-12-04T09:09:26.5432839Z Absolute difference: 123 2025-12-04T09:09:26.5432905Z Relative difference: inf 2025-12-04T09:09:26.5432908Z 2025-12-04T09:09:26.5433028Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5433177Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5433181Z 2025-12-04T09:09:26.5433335Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5433481Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5433552Z Traceback (most recent call last): 2025-12-04T09:09:26.5433788Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5433879Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5434183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5434297Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5434372Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5434375Z 2025-12-04T09:09:26.5434440Z Expected 0 but got 123. 2025-12-04T09:09:26.5434503Z Absolute difference: 123 2025-12-04T09:09:26.5434568Z Relative difference: inf 2025-12-04T09:09:26.5434573Z 2025-12-04T09:09:26.5434695Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5434842Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5434845Z 2025-12-04T09:09:26.5435007Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5435162Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:09:26.5435235Z Traceback (most recent call last): 2025-12-04T09:09:26.5435462Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5435553Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:09:26.5435861Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:09:26.5435976Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:09:26.5436050Z AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5436054Z 2025-12-04T09:09:26.5436121Z Expected 0 but got 123. 2025-12-04T09:09:26.5436184Z Absolute difference: 123 2025-12-04T09:09:26.5436247Z Relative difference: inf 2025-12-04T09:09:26.5436251Z 2025-12-04T09:09:26.5436376Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5436527Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5436530Z 2025-12-04T09:09:26.5436682Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5437003Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda/test_cuda-0b3575165da88d35.xml - 2025-12-04T09:09:26.5437106Z =========================== short test summary info ============================ 2025-12-04T09:09:26.5437353Z FAILED [0.1945s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5437356Z 2025-12-04T09:09:26.5437423Z Expected 196608 but got 0. 2025-12-04T09:09:26.5437494Z Absolute difference: 196608 2025-12-04T09:09:26.5437559Z Relative difference: 1.0 2025-12-04T09:09:26.5437562Z 2025-12-04T09:09:26.5437689Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5437904Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5437907Z 2025-12-04T09:09:26.5438145Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5438377Z FAILED [0.1621s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5438382Z 2025-12-04T09:09:26.5438453Z Expected 196608 but got 0. 2025-12-04T09:09:26.5438521Z Absolute difference: 196608 2025-12-04T09:09:26.5438589Z Relative difference: 1.0 2025-12-04T09:09:26.5438592Z 2025-12-04T09:09:26.5438714Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5438832Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5438835Z 2025-12-04T09:09:26.5438991Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5439210Z FAILED [0.1574s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5439216Z 2025-12-04T09:09:26.5439284Z Expected 196608 but got 0. 2025-12-04T09:09:26.5439352Z Absolute difference: 196608 2025-12-04T09:09:26.5439418Z Relative difference: 1.0 2025-12-04T09:09:26.5439421Z 2025-12-04T09:09:26.5439544Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5439656Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5439660Z 2025-12-04T09:09:26.5439812Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5440030Z FAILED [0.1575s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5440033Z 2025-12-04T09:09:26.5440096Z Expected 196608 but got 0. 2025-12-04T09:09:26.5440165Z Absolute difference: 196608 2025-12-04T09:09:26.5440229Z Relative difference: 1.0 2025-12-04T09:09:26.5440233Z 2025-12-04T09:09:26.5440352Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5440469Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5440473Z 2025-12-04T09:09:26.5440630Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5440855Z FAILED [0.1571s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5440858Z 2025-12-04T09:09:26.5440921Z Expected 196608 but got 0. 2025-12-04T09:09:26.5440986Z Absolute difference: 196608 2025-12-04T09:09:26.5441052Z Relative difference: 1.0 2025-12-04T09:09:26.5441055Z 2025-12-04T09:09:26.5441174Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5441287Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5441290Z 2025-12-04T09:09:26.5441437Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5441653Z FAILED [0.1567s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5441659Z 2025-12-04T09:09:26.5441728Z Expected 196608 but got 0. 2025-12-04T09:09:26.5441793Z Absolute difference: 196608 2025-12-04T09:09:26.5441860Z Relative difference: 1.0 2025-12-04T09:09:26.5441866Z 2025-12-04T09:09:26.5441986Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5442100Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5442103Z 2025-12-04T09:09:26.5442254Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5442468Z FAILED [0.1559s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5442471Z 2025-12-04T09:09:26.5442533Z Expected 196608 but got 0. 2025-12-04T09:09:26.5442604Z Absolute difference: 196608 2025-12-04T09:09:26.5442668Z Relative difference: 1.0 2025-12-04T09:09:26.5442671Z 2025-12-04T09:09:26.5442792Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5442991Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5442994Z 2025-12-04T09:09:26.5443209Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5443430Z FAILED [0.1586s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5443434Z 2025-12-04T09:09:26.5443498Z Expected 196608 but got 0. 2025-12-04T09:09:26.5443566Z Absolute difference: 196608 2025-12-04T09:09:26.5443630Z Relative difference: 1.0 2025-12-04T09:09:26.5443633Z 2025-12-04T09:09:26.5443753Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5443866Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5443870Z 2025-12-04T09:09:26.5444018Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5444233Z FAILED [0.1586s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5444242Z 2025-12-04T09:09:26.5444306Z Expected 196608 but got 0. 2025-12-04T09:09:26.5444370Z Absolute difference: 196608 2025-12-04T09:09:26.5444440Z Relative difference: 1.0 2025-12-04T09:09:26.5444443Z 2025-12-04T09:09:26.5444562Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5444673Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5444676Z 2025-12-04T09:09:26.5444841Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5445059Z FAILED [0.1570s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5445063Z 2025-12-04T09:09:26.5445129Z Expected 196608 but got 0. 2025-12-04T09:09:26.5445194Z Absolute difference: 196608 2025-12-04T09:09:26.5445259Z Relative difference: 1.0 2025-12-04T09:09:26.5445262Z 2025-12-04T09:09:26.5445384Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5445498Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5445501Z 2025-12-04T09:09:26.5445653Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5445872Z FAILED [0.1583s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5445875Z 2025-12-04T09:09:26.5445939Z Expected 196608 but got 0. 2025-12-04T09:09:26.5446008Z Absolute difference: 196608 2025-12-04T09:09:26.5446072Z Relative difference: 1.0 2025-12-04T09:09:26.5446075Z 2025-12-04T09:09:26.5446193Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5446307Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5446310Z 2025-12-04T09:09:26.5446459Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5446677Z FAILED [0.1573s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5446683Z 2025-12-04T09:09:26.5446746Z Expected 196608 but got 0. 2025-12-04T09:09:26.5446812Z Absolute difference: 196608 2025-12-04T09:09:26.5446881Z Relative difference: 1.0 2025-12-04T09:09:26.5446884Z 2025-12-04T09:09:26.5447003Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5447114Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5447119Z 2025-12-04T09:09:26.5447268Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5447484Z FAILED [0.1564s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5447487Z 2025-12-04T09:09:26.5447553Z Expected 196608 but got 0. 2025-12-04T09:09:26.5447618Z Absolute difference: 196608 2025-12-04T09:09:26.5447681Z Relative difference: 1.0 2025-12-04T09:09:26.5447684Z 2025-12-04T09:09:26.5447807Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5447917Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5447992Z 2025-12-04T09:09:26.5448214Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5448437Z FAILED [0.1569s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5448440Z 2025-12-04T09:09:26.5448505Z Expected 196608 but got 0. 2025-12-04T09:09:26.5448574Z Absolute difference: 196608 2025-12-04T09:09:26.5448638Z Relative difference: 1.0 2025-12-04T09:09:26.5448641Z 2025-12-04T09:09:26.5448759Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5448871Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5448874Z 2025-12-04T09:09:26.5449021Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5449238Z FAILED [0.1577s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5449241Z 2025-12-04T09:09:26.5449307Z Expected 196608 but got 0. 2025-12-04T09:09:26.5449373Z Absolute difference: 196608 2025-12-04T09:09:26.5449442Z Relative difference: 1.0 2025-12-04T09:09:26.5449445Z 2025-12-04T09:09:26.5449563Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5449676Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5449679Z 2025-12-04T09:09:26.5449827Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5450040Z FAILED [0.1572s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5450044Z 2025-12-04T09:09:26.5450110Z Expected 196608 but got 0. 2025-12-04T09:09:26.5450173Z Absolute difference: 196608 2025-12-04T09:09:26.5450237Z Relative difference: 1.0 2025-12-04T09:09:26.5450242Z 2025-12-04T09:09:26.5450358Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5450468Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5450475Z 2025-12-04T09:09:26.5450628Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5450843Z FAILED [0.1564s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5450846Z 2025-12-04T09:09:26.5450908Z Expected 196608 but got 0. 2025-12-04T09:09:26.5450975Z Absolute difference: 196608 2025-12-04T09:09:26.5451039Z Relative difference: 1.0 2025-12-04T09:09:26.5451043Z 2025-12-04T09:09:26.5451163Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5451275Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5451278Z 2025-12-04T09:09:26.5451426Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5451645Z FAILED [0.1562s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5451648Z 2025-12-04T09:09:26.5451715Z Expected 196608 but got 0. 2025-12-04T09:09:26.5451784Z Absolute difference: 196608 2025-12-04T09:09:26.5451847Z Relative difference: 1.0 2025-12-04T09:09:26.5451853Z 2025-12-04T09:09:26.5451972Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5452085Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5452088Z 2025-12-04T09:09:26.5452236Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5452450Z FAILED [0.1562s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5452456Z 2025-12-04T09:09:26.5452519Z Expected 196608 but got 0. 2025-12-04T09:09:26.5452584Z Absolute difference: 196608 2025-12-04T09:09:26.5452651Z Relative difference: 1.0 2025-12-04T09:09:26.5452654Z 2025-12-04T09:09:26.5452772Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5452882Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5452965Z 2025-12-04T09:09:26.5453120Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5453490Z FAILED [0.1568s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5453494Z 2025-12-04T09:09:26.5453564Z Expected 196608 but got 0. 2025-12-04T09:09:26.5453630Z Absolute difference: 196608 2025-12-04T09:09:26.5453695Z Relative difference: 1.0 2025-12-04T09:09:26.5453698Z 2025-12-04T09:09:26.5453819Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5453929Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5453933Z 2025-12-04T09:09:26.5454080Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5454296Z FAILED [0.1563s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5454299Z 2025-12-04T09:09:26.5454374Z Expected 196608 but got 0. 2025-12-04T09:09:26.5454445Z Absolute difference: 196608 2025-12-04T09:09:26.5454511Z Relative difference: 1.0 2025-12-04T09:09:26.5454517Z 2025-12-04T09:09:26.5454638Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5454752Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5454755Z 2025-12-04T09:09:26.5454903Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5455119Z FAILED [0.1578s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5455122Z 2025-12-04T09:09:26.5455184Z Expected 196608 but got 0. 2025-12-04T09:09:26.5455249Z Absolute difference: 196608 2025-12-04T09:09:26.5455315Z Relative difference: 1.0 2025-12-04T09:09:26.5455318Z 2025-12-04T09:09:26.5455439Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5455550Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5455561Z 2025-12-04T09:09:26.5455709Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5455928Z FAILED [0.1572s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5455931Z 2025-12-04T09:09:26.5455998Z Expected 196608 but got 0. 2025-12-04T09:09:26.5456062Z Absolute difference: 196608 2025-12-04T09:09:26.5456127Z Relative difference: 1.0 2025-12-04T09:09:26.5456130Z 2025-12-04T09:09:26.5456256Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5456368Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5456371Z 2025-12-04T09:09:26.5456522Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5456735Z FAILED [0.1566s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5456739Z 2025-12-04T09:09:26.5456802Z Expected 196608 but got 0. 2025-12-04T09:09:26.5456877Z Absolute difference: 196608 2025-12-04T09:09:26.5456942Z Relative difference: 1.0 2025-12-04T09:09:26.5456945Z 2025-12-04T09:09:26.5457076Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5457193Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5457197Z 2025-12-04T09:09:26.5457346Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5457566Z FAILED [0.1568s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5457569Z 2025-12-04T09:09:26.5457632Z Expected 196608 but got 0. 2025-12-04T09:09:26.5457697Z Absolute difference: 196608 2025-12-04T09:09:26.5457765Z Relative difference: 1.0 2025-12-04T09:09:26.5457768Z 2025-12-04T09:09:26.5457889Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5458005Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5458097Z 2025-12-04T09:09:26.5458249Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5458531Z FAILED [0.1568s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5458538Z 2025-12-04T09:09:26.5458608Z Expected 196608 but got 0. 2025-12-04T09:09:26.5458673Z Absolute difference: 196608 2025-12-04T09:09:26.5458741Z Relative difference: 1.0 2025-12-04T09:09:26.5458745Z 2025-12-04T09:09:26.5458872Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5458982Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5458986Z 2025-12-04T09:09:26.5459138Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5459352Z FAILED [0.1568s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5459356Z 2025-12-04T09:09:26.5459421Z Expected 196608 but got 0. 2025-12-04T09:09:26.5459491Z Absolute difference: 196608 2025-12-04T09:09:26.5459555Z Relative difference: 1.0 2025-12-04T09:09:26.5459558Z 2025-12-04T09:09:26.5459682Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5459793Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5459796Z 2025-12-04T09:09:26.5459945Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5460162Z FAILED [0.1570s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5460165Z 2025-12-04T09:09:26.5460227Z Expected 196608 but got 0. 2025-12-04T09:09:26.5460293Z Absolute difference: 196608 2025-12-04T09:09:26.5460356Z Relative difference: 1.0 2025-12-04T09:09:26.5460359Z 2025-12-04T09:09:26.5460477Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5460590Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5460597Z 2025-12-04T09:09:26.5460745Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5460967Z FAILED [0.1574s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5460970Z 2025-12-04T09:09:26.5461034Z Expected 196608 but got 0. 2025-12-04T09:09:26.5461098Z Absolute difference: 196608 2025-12-04T09:09:26.5461164Z Relative difference: 1.0 2025-12-04T09:09:26.5461167Z 2025-12-04T09:09:26.5461283Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5461395Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5461398Z 2025-12-04T09:09:26.5461550Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5461766Z FAILED [0.1572s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5461769Z 2025-12-04T09:09:26.5461835Z Expected 196608 but got 0. 2025-12-04T09:09:26.5461903Z Absolute difference: 196608 2025-12-04T09:09:26.5461967Z Relative difference: 1.0 2025-12-04T09:09:26.5461970Z 2025-12-04T09:09:26.5462093Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5462205Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5462207Z 2025-12-04T09:09:26.5462359Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5462574Z FAILED [0.1563s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5462577Z 2025-12-04T09:09:26.5462639Z Expected 196608 but got 0. 2025-12-04T09:09:26.5462706Z Absolute difference: 196608 2025-12-04T09:09:26.5462770Z Relative difference: 1.0 2025-12-04T09:09:26.5462773Z 2025-12-04T09:09:26.5462891Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5463019Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5463022Z 2025-12-04T09:09:26.5463256Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5463539Z FAILED [0.1565s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5463543Z 2025-12-04T09:09:26.5463607Z Expected 196608 but got 0. 2025-12-04T09:09:26.5463672Z Absolute difference: 196608 2025-12-04T09:09:26.5463740Z Relative difference: 1.0 2025-12-04T09:09:26.5463743Z 2025-12-04T09:09:26.5463863Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5463977Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5463980Z 2025-12-04T09:09:26.5464129Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5464344Z FAILED [0.1577s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5464348Z 2025-12-04T09:09:26.5464413Z Expected 196608 but got 0. 2025-12-04T09:09:26.5464481Z Absolute difference: 196608 2025-12-04T09:09:26.5464544Z Relative difference: 1.0 2025-12-04T09:09:26.5464547Z 2025-12-04T09:09:26.5464671Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5464784Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5464787Z 2025-12-04T09:09:26.5464937Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5465150Z FAILED [0.1562s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5465153Z 2025-12-04T09:09:26.5465217Z Expected 196608 but got 0. 2025-12-04T09:09:26.5465286Z Absolute difference: 196608 2025-12-04T09:09:26.5465351Z Relative difference: 1.0 2025-12-04T09:09:26.5465354Z 2025-12-04T09:09:26.5465476Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5465588Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5465591Z 2025-12-04T09:09:26.5465743Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5465966Z FAILED [0.1565s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5465969Z 2025-12-04T09:09:26.5466032Z Expected 196608 but got 0. 2025-12-04T09:09:26.5466100Z Absolute difference: 196608 2025-12-04T09:09:26.5466164Z Relative difference: 1.0 2025-12-04T09:09:26.5466166Z 2025-12-04T09:09:26.5466286Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5466406Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5466409Z 2025-12-04T09:09:26.5466560Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5466775Z FAILED [0.1583s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5466782Z 2025-12-04T09:09:26.5466845Z Expected 196608 but got 0. 2025-12-04T09:09:26.5466911Z Absolute difference: 196608 2025-12-04T09:09:26.5466980Z Relative difference: 1.0 2025-12-04T09:09:26.5466983Z 2025-12-04T09:09:26.5467105Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5467216Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5467219Z 2025-12-04T09:09:26.5467370Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5467585Z FAILED [0.1571s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5467589Z 2025-12-04T09:09:26.5467652Z Expected 196608 but got 0. 2025-12-04T09:09:26.5467717Z Absolute difference: 196608 2025-12-04T09:09:26.5467782Z Relative difference: 1.0 2025-12-04T09:09:26.5467785Z 2025-12-04T09:09:26.5467906Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5468028Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5468031Z 2025-12-04T09:09:26.5468183Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5468536Z FAILED [0.1564s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5468540Z 2025-12-04T09:09:26.5468605Z Expected 196608 but got 0. 2025-12-04T09:09:26.5468683Z Absolute difference: 196608 2025-12-04T09:09:26.5468747Z Relative difference: 1.0 2025-12-04T09:09:26.5468750Z 2025-12-04T09:09:26.5468869Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5468982Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5468985Z 2025-12-04T09:09:26.5469134Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5469353Z FAILED [0.1575s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5469356Z 2025-12-04T09:09:26.5469420Z Expected 196608 but got 0. 2025-12-04T09:09:26.5469483Z Absolute difference: 196608 2025-12-04T09:09:26.5469553Z Relative difference: 1.0 2025-12-04T09:09:26.5469556Z 2025-12-04T09:09:26.5469678Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5469788Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5469796Z 2025-12-04T09:09:26.5469946Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5470164Z FAILED [0.1564s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5470167Z 2025-12-04T09:09:26.5470241Z Expected 196608 but got 0. 2025-12-04T09:09:26.5470306Z Absolute difference: 196608 2025-12-04T09:09:26.5470379Z Relative difference: 1.0 2025-12-04T09:09:26.5470383Z 2025-12-04T09:09:26.5470509Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5470621Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5470624Z 2025-12-04T09:09:26.5470776Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5470997Z FAILED [0.1562s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5471000Z 2025-12-04T09:09:26.5471062Z Expected 196608 but got 0. 2025-12-04T09:09:26.5471128Z Absolute difference: 196608 2025-12-04T09:09:26.5471194Z Relative difference: 1.0 2025-12-04T09:09:26.5471197Z 2025-12-04T09:09:26.5471316Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5471431Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5471434Z 2025-12-04T09:09:26.5471583Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5471802Z FAILED [0.1573s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5471805Z 2025-12-04T09:09:26.5471867Z Expected 196608 but got 0. 2025-12-04T09:09:26.5471932Z Absolute difference: 196608 2025-12-04T09:09:26.5472002Z Relative difference: 1.0 2025-12-04T09:09:26.5472005Z 2025-12-04T09:09:26.5472128Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5472242Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5472245Z 2025-12-04T09:09:26.5472395Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5472608Z FAILED [0.1573s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5472612Z 2025-12-04T09:09:26.5472677Z Expected 196608 but got 0. 2025-12-04T09:09:26.5472741Z Absolute difference: 196608 2025-12-04T09:09:26.5472805Z Relative difference: 1.0 2025-12-04T09:09:26.5472810Z 2025-12-04T09:09:26.5472929Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5473039Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5473042Z 2025-12-04T09:09:26.5473192Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5473589Z FAILED [0.1581s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5473593Z 2025-12-04T09:09:26.5473668Z Expected 196608 but got 0. 2025-12-04T09:09:26.5473740Z Absolute difference: 196608 2025-12-04T09:09:26.5473806Z Relative difference: 1.0 2025-12-04T09:09:26.5473810Z 2025-12-04T09:09:26.5473933Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5474044Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5474047Z 2025-12-04T09:09:26.5474197Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5474413Z FAILED [0.1580s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5474416Z 2025-12-04T09:09:26.5474479Z Expected 196608 but got 0. 2025-12-04T09:09:26.5474546Z Absolute difference: 196608 2025-12-04T09:09:26.5474616Z Relative difference: 1.0 2025-12-04T09:09:26.5474619Z 2025-12-04T09:09:26.5474740Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5474858Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5474861Z 2025-12-04T09:09:26.5475010Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5475224Z FAILED [0.1564s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5475231Z 2025-12-04T09:09:26.5475293Z Expected 196608 but got 0. 2025-12-04T09:09:26.5475357Z Absolute difference: 196608 2025-12-04T09:09:26.5475424Z Relative difference: 1.0 2025-12-04T09:09:26.5475427Z 2025-12-04T09:09:26.5475547Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5475659Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5475662Z 2025-12-04T09:09:26.5475811Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5476030Z FAILED [0.1620s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5476033Z 2025-12-04T09:09:26.5476097Z Expected 196608 but got 0. 2025-12-04T09:09:26.5476164Z Absolute difference: 196608 2025-12-04T09:09:26.5476227Z Relative difference: 1.0 2025-12-04T09:09:26.5476230Z 2025-12-04T09:09:26.5476350Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5476460Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5476464Z 2025-12-04T09:09:26.5476612Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5476830Z FAILED [0.1582s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5476833Z 2025-12-04T09:09:26.5476898Z Expected 196608 but got 0. 2025-12-04T09:09:26.5476965Z Absolute difference: 196608 2025-12-04T09:09:26.5477031Z Relative difference: 1.0 2025-12-04T09:09:26.5477036Z 2025-12-04T09:09:26.5477159Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5477276Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5477279Z 2025-12-04T09:09:26.5477428Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5477643Z FAILED [0.1568s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5477646Z 2025-12-04T09:09:26.5477708Z Expected 196608 but got 0. 2025-12-04T09:09:26.5477773Z Absolute difference: 196608 2025-12-04T09:09:26.5477840Z Relative difference: 1.0 2025-12-04T09:09:26.5477843Z 2025-12-04T09:09:26.5477962Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5478073Z python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:09:26.5478079Z 2025-12-04T09:09:26.5478227Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5478660Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5478665Z 2025-12-04T09:09:26.5478734Z Expected 0 but got 123. 2025-12-04T09:09:26.5478798Z Absolute difference: 123 2025-12-04T09:09:26.5478872Z Relative difference: inf 2025-12-04T09:09:26.5478875Z 2025-12-04T09:09:26.5479002Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5479154Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5479158Z 2025-12-04T09:09:26.5479309Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5479571Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5479574Z 2025-12-04T09:09:26.5479638Z Expected 0 but got 123. 2025-12-04T09:09:26.5479705Z Absolute difference: 123 2025-12-04T09:09:26.5479772Z Relative difference: inf 2025-12-04T09:09:26.5479776Z 2025-12-04T09:09:26.5479899Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5480051Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5480054Z 2025-12-04T09:09:26.5480202Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5480465Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5480468Z 2025-12-04T09:09:26.5480531Z Expected 0 but got 123. 2025-12-04T09:09:26.5480595Z Absolute difference: 123 2025-12-04T09:09:26.5480662Z Relative difference: inf 2025-12-04T09:09:26.5480665Z 2025-12-04T09:09:26.5480784Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5480935Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5480938Z 2025-12-04T09:09:26.5481088Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5481351Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5481357Z 2025-12-04T09:09:26.5481419Z Expected 0 but got 123. 2025-12-04T09:09:26.5481482Z Absolute difference: 123 2025-12-04T09:09:26.5481549Z Relative difference: inf 2025-12-04T09:09:26.5481552Z 2025-12-04T09:09:26.5481671Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5481816Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5481820Z 2025-12-04T09:09:26.5481969Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5482228Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5482232Z 2025-12-04T09:09:26.5482296Z Expected 0 but got 123. 2025-12-04T09:09:26.5482363Z Absolute difference: 123 2025-12-04T09:09:26.5482427Z Relative difference: inf 2025-12-04T09:09:26.5482431Z 2025-12-04T09:09:26.5482555Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5482701Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5482704Z 2025-12-04T09:09:26.5482851Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5483111Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5483114Z 2025-12-04T09:09:26.5483176Z Expected 0 but got 123. 2025-12-04T09:09:26.5483243Z Absolute difference: 123 2025-12-04T09:09:26.5483317Z Relative difference: inf 2025-12-04T09:09:26.5483321Z 2025-12-04T09:09:26.5483448Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5483599Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5483675Z 2025-12-04T09:09:26.5483826Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5484150Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5484154Z 2025-12-04T09:09:26.5484219Z Expected 0 but got 123. 2025-12-04T09:09:26.5484285Z Absolute difference: 123 2025-12-04T09:09:26.5484351Z Relative difference: inf 2025-12-04T09:09:26.5484354Z 2025-12-04T09:09:26.5484482Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5484634Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5484639Z 2025-12-04T09:09:26.5484785Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5485041Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5485044Z 2025-12-04T09:09:26.5485114Z Expected 0 but got 123. 2025-12-04T09:09:26.5485178Z Absolute difference: 123 2025-12-04T09:09:26.5485241Z Relative difference: inf 2025-12-04T09:09:26.5485247Z 2025-12-04T09:09:26.5485370Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5485516Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5485519Z 2025-12-04T09:09:26.5485668Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5485924Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5485927Z 2025-12-04T09:09:26.5485989Z Expected 0 but got 123. 2025-12-04T09:09:26.5486055Z Absolute difference: 123 2025-12-04T09:09:26.5486117Z Relative difference: inf 2025-12-04T09:09:26.5486121Z 2025-12-04T09:09:26.5486242Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5486388Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5486394Z 2025-12-04T09:09:26.5486545Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5486807Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5486810Z 2025-12-04T09:09:26.5486873Z Expected 0 but got 123. 2025-12-04T09:09:26.5486939Z Absolute difference: 123 2025-12-04T09:09:26.5487005Z Relative difference: inf 2025-12-04T09:09:26.5487008Z 2025-12-04T09:09:26.5487128Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5487277Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5487280Z 2025-12-04T09:09:26.5487428Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5487683Z FAILED [0.0011s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5487692Z 2025-12-04T09:09:26.5487755Z Expected 0 but got 123. 2025-12-04T09:09:26.5487819Z Absolute difference: 123 2025-12-04T09:09:26.5487888Z Relative difference: inf 2025-12-04T09:09:26.5487891Z 2025-12-04T09:09:26.5488010Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5488158Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5488162Z 2025-12-04T09:09:26.5488312Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5488567Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5488571Z 2025-12-04T09:09:26.5488643Z Expected 0 but got 123. 2025-12-04T09:09:26.5488708Z Absolute difference: 123 2025-12-04T09:09:26.5488771Z Relative difference: inf 2025-12-04T09:09:26.5488774Z 2025-12-04T09:09:26.5488897Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5489130Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5489133Z 2025-12-04T09:09:26.5489344Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5489607Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5489610Z 2025-12-04T09:09:26.5489673Z Expected 0 but got 123. 2025-12-04T09:09:26.5489740Z Absolute difference: 123 2025-12-04T09:09:26.5489803Z Relative difference: inf 2025-12-04T09:09:26.5489806Z 2025-12-04T09:09:26.5489925Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5490075Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5490078Z 2025-12-04T09:09:26.5490227Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5490487Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5490493Z 2025-12-04T09:09:26.5490555Z Expected 0 but got 123. 2025-12-04T09:09:26.5490624Z Absolute difference: 123 2025-12-04T09:09:26.5490690Z Relative difference: inf 2025-12-04T09:09:26.5490693Z 2025-12-04T09:09:26.5490813Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5490959Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5490965Z 2025-12-04T09:09:26.5491115Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5491371Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5491375Z 2025-12-04T09:09:26.5491439Z Expected 0 but got 123. 2025-12-04T09:09:26.5491505Z Absolute difference: 123 2025-12-04T09:09:26.5491569Z Relative difference: inf 2025-12-04T09:09:26.5491572Z 2025-12-04T09:09:26.5491693Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5491842Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5491848Z 2025-12-04T09:09:26.5491999Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5492254Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5492257Z 2025-12-04T09:09:26.5492319Z Expected 0 but got 123. 2025-12-04T09:09:26.5492386Z Absolute difference: 123 2025-12-04T09:09:26.5492450Z Relative difference: inf 2025-12-04T09:09:26.5492453Z 2025-12-04T09:09:26.5492575Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5492722Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5492725Z 2025-12-04T09:09:26.5492871Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5493133Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5493188Z 2025-12-04T09:09:26.5493258Z Expected 0 but got 123. 2025-12-04T09:09:26.5493326Z Absolute difference: 123 2025-12-04T09:09:26.5493390Z Relative difference: inf 2025-12-04T09:09:26.5493393Z 2025-12-04T09:09:26.5493512Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5493662Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5493665Z 2025-12-04T09:09:26.5493812Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5494070Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5494076Z 2025-12-04T09:09:26.5494139Z Expected 0 but got 123. 2025-12-04T09:09:26.5494201Z Absolute difference: 123 2025-12-04T09:09:26.5494266Z Relative difference: inf 2025-12-04T09:09:26.5494269Z 2025-12-04T09:09:26.5494464Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5494694Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5494699Z 2025-12-04T09:09:26.5494854Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5495111Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5495115Z 2025-12-04T09:09:26.5495177Z Expected 0 but got 123. 2025-12-04T09:09:26.5495240Z Absolute difference: 123 2025-12-04T09:09:26.5495303Z Relative difference: inf 2025-12-04T09:09:26.5495307Z 2025-12-04T09:09:26.5495427Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5495573Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5495576Z 2025-12-04T09:09:26.5495723Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5495986Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5495992Z 2025-12-04T09:09:26.5496054Z Expected 0 but got 123. 2025-12-04T09:09:26.5496121Z Absolute difference: 123 2025-12-04T09:09:26.5496185Z Relative difference: inf 2025-12-04T09:09:26.5496188Z 2025-12-04T09:09:26.5496306Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5496454Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5496457Z 2025-12-04T09:09:26.5496603Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5496866Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5496871Z 2025-12-04T09:09:26.5496938Z Expected 0 but got 123. 2025-12-04T09:09:26.5497004Z Absolute difference: 123 2025-12-04T09:09:26.5497070Z Relative difference: inf 2025-12-04T09:09:26.5497076Z 2025-12-04T09:09:26.5497197Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5497346Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5497353Z 2025-12-04T09:09:26.5497500Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5497758Z FAILED [0.0011s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5497761Z 2025-12-04T09:09:26.5497826Z Expected 0 but got 123. 2025-12-04T09:09:26.5497890Z Absolute difference: 123 2025-12-04T09:09:26.5497952Z Relative difference: inf 2025-12-04T09:09:26.5497955Z 2025-12-04T09:09:26.5498076Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5498225Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5498229Z 2025-12-04T09:09:26.5498379Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5498640Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5498643Z 2025-12-04T09:09:26.5498706Z Expected 0 but got 123. 2025-12-04T09:09:26.5498773Z Absolute difference: 123 2025-12-04T09:09:26.5498835Z Relative difference: inf 2025-12-04T09:09:26.5498839Z 2025-12-04T09:09:26.5498961Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5499107Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5499110Z 2025-12-04T09:09:26.5499260Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5499520Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5499523Z 2025-12-04T09:09:26.5499586Z Expected 0 but got 123. 2025-12-04T09:09:26.5499654Z Absolute difference: 123 2025-12-04T09:09:26.5499793Z Relative difference: inf 2025-12-04T09:09:26.5499796Z 2025-12-04T09:09:26.5499979Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5500129Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5500132Z 2025-12-04T09:09:26.5500289Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5500553Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5500560Z 2025-12-04T09:09:26.5500623Z Expected 0 but got 123. 2025-12-04T09:09:26.5500687Z Absolute difference: 123 2025-12-04T09:09:26.5500751Z Relative difference: inf 2025-12-04T09:09:26.5500753Z 2025-12-04T09:09:26.5500872Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5501017Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5501023Z 2025-12-04T09:09:26.5501171Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5501432Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5501435Z 2025-12-04T09:09:26.5501501Z Expected 0 but got 123. 2025-12-04T09:09:26.5501564Z Absolute difference: 123 2025-12-04T09:09:26.5501628Z Relative difference: inf 2025-12-04T09:09:26.5501631Z 2025-12-04T09:09:26.5501752Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5501899Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5501903Z 2025-12-04T09:09:26.5502050Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5502310Z FAILED [0.0011s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5502313Z 2025-12-04T09:09:26.5502376Z Expected 0 but got 123. 2025-12-04T09:09:26.5502446Z Absolute difference: 123 2025-12-04T09:09:26.5502509Z Relative difference: inf 2025-12-04T09:09:26.5502512Z 2025-12-04T09:09:26.5502635Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5502784Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5502787Z 2025-12-04T09:09:26.5502933Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5503194Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5503197Z 2025-12-04T09:09:26.5503260Z Expected 0 but got 123. 2025-12-04T09:09:26.5503323Z Absolute difference: 123 2025-12-04T09:09:26.5503391Z Relative difference: inf 2025-12-04T09:09:26.5503394Z 2025-12-04T09:09:26.5503512Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5503663Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5503668Z 2025-12-04T09:09:26.5503818Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5504077Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5504081Z 2025-12-04T09:09:26.5504144Z Expected 0 but got 123. 2025-12-04T09:09:26.5504208Z Absolute difference: 123 2025-12-04T09:09:26.5504272Z Relative difference: inf 2025-12-04T09:09:26.5504276Z 2025-12-04T09:09:26.5504397Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5504778Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5504783Z 2025-12-04T09:09:26.5504942Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5505207Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5505335Z 2025-12-04T09:09:26.5505406Z Expected 0 but got 123. 2025-12-04T09:09:26.5505474Z Absolute difference: 123 2025-12-04T09:09:26.5505633Z Relative difference: inf 2025-12-04T09:09:26.5505640Z 2025-12-04T09:09:26.5505773Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5505922Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5505925Z 2025-12-04T09:09:26.5506070Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5506331Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5506334Z 2025-12-04T09:09:26.5506397Z Expected 0 but got 123. 2025-12-04T09:09:26.5506464Z Absolute difference: 123 2025-12-04T09:09:26.5506527Z Relative difference: inf 2025-12-04T09:09:26.5506531Z 2025-12-04T09:09:26.5506650Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5506802Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5506805Z 2025-12-04T09:09:26.5506953Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5507210Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5507216Z 2025-12-04T09:09:26.5507280Z Expected 0 but got 123. 2025-12-04T09:09:26.5507344Z Absolute difference: 123 2025-12-04T09:09:26.5507409Z Relative difference: inf 2025-12-04T09:09:26.5507411Z 2025-12-04T09:09:26.5507531Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5507679Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5507682Z 2025-12-04T09:09:26.5507833Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5508090Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5508099Z 2025-12-04T09:09:26.5508163Z Expected 0 but got 123. 2025-12-04T09:09:26.5508235Z Absolute difference: 123 2025-12-04T09:09:26.5508310Z Relative difference: inf 2025-12-04T09:09:26.5508314Z 2025-12-04T09:09:26.5508437Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5508583Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5508586Z 2025-12-04T09:09:26.5508732Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5508991Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5508994Z 2025-12-04T09:09:26.5509056Z Expected 0 but got 123. 2025-12-04T09:09:26.5509123Z Absolute difference: 123 2025-12-04T09:09:26.5509187Z Relative difference: inf 2025-12-04T09:09:26.5509189Z 2025-12-04T09:09:26.5509310Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5509462Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5509468Z 2025-12-04T09:09:26.5509615Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5509874Z FAILED [0.0011s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5509877Z 2025-12-04T09:09:26.5509940Z Expected 0 but got 123. 2025-12-04T09:09:26.5510004Z Absolute difference: 123 2025-12-04T09:09:26.5510071Z Relative difference: inf 2025-12-04T09:09:26.5510073Z 2025-12-04T09:09:26.5510191Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5510340Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5510343Z 2025-12-04T09:09:26.5510490Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5510746Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5510831Z 2025-12-04T09:09:26.5510963Z Expected 0 but got 123. 2025-12-04T09:09:26.5511031Z Absolute difference: 123 2025-12-04T09:09:26.5511094Z Relative difference: inf 2025-12-04T09:09:26.5511097Z 2025-12-04T09:09:26.5511219Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5511366Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5511369Z 2025-12-04T09:09:26.5511517Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5511771Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5511775Z 2025-12-04T09:09:26.5511840Z Expected 0 but got 123. 2025-12-04T09:09:26.5511919Z Absolute difference: 123 2025-12-04T09:09:26.5511984Z Relative difference: inf 2025-12-04T09:09:26.5511987Z 2025-12-04T09:09:26.5512111Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5512259Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5512262Z 2025-12-04T09:09:26.5512409Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5512668Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5512671Z 2025-12-04T09:09:26.5512733Z Expected 0 but got 123. 2025-12-04T09:09:26.5512800Z Absolute difference: 123 2025-12-04T09:09:26.5512864Z Relative difference: inf 2025-12-04T09:09:26.5512867Z 2025-12-04T09:09:26.5512987Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5513137Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5513140Z 2025-12-04T09:09:26.5513288Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5513551Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5513559Z 2025-12-04T09:09:26.5513627Z Expected 0 but got 123. 2025-12-04T09:09:26.5513694Z Absolute difference: 123 2025-12-04T09:09:26.5513761Z Relative difference: inf 2025-12-04T09:09:26.5513764Z 2025-12-04T09:09:26.5517921Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5518148Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5518153Z 2025-12-04T09:09:26.5518330Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5518628Z FAILED [0.0011s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5518633Z 2025-12-04T09:09:26.5518708Z Expected 0 but got 123. 2025-12-04T09:09:26.5518778Z Absolute difference: 123 2025-12-04T09:09:26.5518843Z Relative difference: inf 2025-12-04T09:09:26.5518858Z 2025-12-04T09:09:26.5519003Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5519172Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5519176Z 2025-12-04T09:09:26.5519342Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5519621Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5519625Z 2025-12-04T09:09:26.5519697Z Expected 0 but got 123. 2025-12-04T09:09:26.5519764Z Absolute difference: 123 2025-12-04T09:09:26.5519829Z Relative difference: inf 2025-12-04T09:09:26.5519832Z 2025-12-04T09:09:26.5519967Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5520120Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5520123Z 2025-12-04T09:09:26.5520274Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5520742Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5520747Z 2025-12-04T09:09:26.5520816Z Expected 0 but got 123. 2025-12-04T09:09:26.5520892Z Absolute difference: 123 2025-12-04T09:09:26.5520956Z Relative difference: inf 2025-12-04T09:09:26.5520959Z 2025-12-04T09:09:26.5521087Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5521242Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5521245Z 2025-12-04T09:09:26.5521395Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5521656Z FAILED [0.0011s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5521659Z 2025-12-04T09:09:26.5521722Z Expected 0 but got 123. 2025-12-04T09:09:26.5521787Z Absolute difference: 123 2025-12-04T09:09:26.5521858Z Relative difference: inf 2025-12-04T09:09:26.5521861Z 2025-12-04T09:09:26.5521987Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5522138Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5522141Z 2025-12-04T09:09:26.5522294Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5522552Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5522555Z 2025-12-04T09:09:26.5522627Z Expected 0 but got 123. 2025-12-04T09:09:26.5522696Z Absolute difference: 123 2025-12-04T09:09:26.5522762Z Relative difference: inf 2025-12-04T09:09:26.5522766Z 2025-12-04T09:09:26.5522889Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5523037Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5523039Z 2025-12-04T09:09:26.5523193Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5523455Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5523460Z 2025-12-04T09:09:26.5523529Z Expected 0 but got 123. 2025-12-04T09:09:26.5523597Z Absolute difference: 123 2025-12-04T09:09:26.5523659Z Relative difference: inf 2025-12-04T09:09:26.5523662Z 2025-12-04T09:09:26.5523784Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5523935Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5523938Z 2025-12-04T09:09:26.5524085Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5524344Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5524347Z 2025-12-04T09:09:26.5524410Z Expected 0 but got 123. 2025-12-04T09:09:26.5524477Z Absolute difference: 123 2025-12-04T09:09:26.5524544Z Relative difference: inf 2025-12-04T09:09:26.5524548Z 2025-12-04T09:09:26.5524672Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5524826Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5524829Z 2025-12-04T09:09:26.5524978Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5525235Z FAILED [0.0012s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5525238Z 2025-12-04T09:09:26.5525307Z Expected 0 but got 123. 2025-12-04T09:09:26.5525371Z Absolute difference: 123 2025-12-04T09:09:26.5525436Z Relative difference: inf 2025-12-04T09:09:26.5525443Z 2025-12-04T09:09:26.5525564Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5525709Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5525789Z 2025-12-04T09:09:26.5525954Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5526274Z FAILED [0.0011s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:09:26.5526278Z 2025-12-04T09:09:26.5526346Z Expected 0 but got 123. 2025-12-04T09:09:26.5526410Z Absolute difference: 123 2025-12-04T09:09:26.5526473Z Relative difference: inf 2025-12-04T09:09:26.5526477Z 2025-12-04T09:09:26.5526598Z To execute this test, run the following from the base repo dir: 2025-12-04T09:09:26.5526745Z python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:09:26.5526749Z 2025-12-04T09:09:26.5526907Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:09:26.5527063Z === 97 failed, 103 passed, 50 skipped, 11200 deselected in 70.94s (0:01:10) ==== 2025-12-04T09:09:26.5527066Z 2025-12-04T09:09:26.5527292Z FINISHED PRINTING LOG FILE of test_cuda 1/1 (test/test-reports/test_cuda_1.1_c907a69791bbc315_.log) 2025-12-04T09:09:26.5527299Z 2025-12-04T09:09:26.5527458Z Finished test_cuda 1/1 ... [2025-12-04 09:09:26.458877][1002.418415708], took 1.26min 2025-12-04T09:09:26.5527821Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda/test_cuda-0b3575165da88d35.xml 2025-12-04T09:09:26.7523703Z Uploading logs for 57116084869 to S3 2025-12-04T09:09:26.7861715Z Uploading artifacts took 0.22 seconds 2025-12-04T09:09:26.7862016Z test_cuda 1/1 failed! 2025-12-04T09:09:26.7865701Z Running test_transformers 1/1 ... [2025-12-04 09:09:26.786378][1002.745923832] 2025-12-04T09:09:26.7866130Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:09:26.7869527Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_transformers.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:09:26.786690] 2025-12-04T09:09:48.9216273Z 2025-12-04T09:09:48.9217111Z test_transformers 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_transformers_1.1_553a713ded2a82f8_.log 2025-12-04T09:09:48.9217787Z Running 0 items in this shard: 2025-12-04T09:09:48.9217965Z 2025-12-04T09:09:48.9218226Z Finished test_transformers 1/1 ... [2025-12-04 09:09:48.921393][1024.880939457], took 0.37min 2025-12-04T09:09:48.9246660Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_transformers/test_transformers-0aff76852e040d8e.xml 2025-12-04T09:09:48.9974040Z Running higher_order_ops/test_local_map 1/1 ... [2025-12-04 09:09:48.997157][1024.956705301] 2025-12-04T09:09:48.9974538Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:09:48.9977660Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'higher_order_ops/test_local_map.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:09:48.997440] 2025-12-04T09:09:52.2154433Z 2025-12-04T09:09:52.2155464Z higher_order_ops/test_local_map 1/1 was successful, full logs can be found in artifacts with path test/test-reports/higher_order_ops.test_local_map_1.1_251d3402b699ff52_.log 2025-12-04T09:09:52.2156250Z Running 0 items in this shard: 2025-12-04T09:09:52.2156423Z 2025-12-04T09:09:52.2156737Z Finished higher_order_ops/test_local_map 1/1 ... [2025-12-04 09:09:52.215207][1028.174752971], took 0.05min 2025-12-04T09:09:52.2186565Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/higher_order_ops.test_local_map/higher_order_ops.test_local_map-c325eb7c48529fe1.xml 2025-12-04T09:09:52.2453036Z Running test_dataloader 1/1 ... [2025-12-04 09:09:52.245102][1028.204650618] 2025-12-04T09:09:52.2454651Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:09:52.2456588Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_dataloader.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:09:52.245405] 2025-12-04T09:09:56.0454957Z 2025-12-04T09:09:56.0455744Z test_dataloader 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_dataloader_1.1_8a6dced23b850537_.log 2025-12-04T09:09:56.0456430Z Running 0 items in this shard: 2025-12-04T09:09:56.0456617Z 2025-12-04T09:09:56.0456852Z Finished test_dataloader 1/1 ... [2025-12-04 09:09:56.045269][1032.004815339], took 0.06min 2025-12-04T09:09:56.0488365Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_dataloader/test_dataloader-c3ab56b023ec83c9.xml 2025-12-04T09:09:56.0725112Z Running test_decomp 6/22 ... [2025-12-04 09:09:56.072297][1032.031845287] 2025-12-04T09:09:56.0725652Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:09:56.0728475Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_decomp.py', '-m', 'serial', '--shard-id=6', '--num-shards=22', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:09:56.072593] 2025-12-04T09:10:12.5116370Z 2025-12-04T09:10:12.5117235Z test_decomp 6/22 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_6.22_772db1786e1c2ca3_.log 2025-12-04T09:10:12.5117870Z Running 0 items in this shard: 2025-12-04T09:10:12.5118042Z 2025-12-04T09:10:12.5118274Z Finished test_decomp 6/22 ... [2025-12-04 09:10:12.511410][1048.470955889], took 0.27min 2025-12-04T09:10:12.5150534Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-1d76070ccdaa412b.xml 2025-12-04T09:10:12.6003098Z Running test_decomp 8/22 ... [2025-12-04 09:10:12.600099][1048.559647677] 2025-12-04T09:10:12.6003515Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:10:12.6006640Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_decomp.py', '-m', 'serial', '--shard-id=8', '--num-shards=22', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:10:12.600398] 2025-12-04T09:10:29.0938273Z 2025-12-04T09:10:29.0939028Z test_decomp 8/22 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_8.22_2bb5d34331689c22_.log 2025-12-04T09:10:29.0939665Z Running 0 items in this shard: 2025-12-04T09:10:29.0939837Z 2025-12-04T09:10:29.0940065Z Finished test_decomp 8/22 ... [2025-12-04 09:10:29.093380][1065.052926274], took 0.27min 2025-12-04T09:10:29.0972366Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-d3940d286152db68.xml 2025-12-04T09:10:29.1756964Z Running test_decomp 12/22 ... [2025-12-04 09:10:29.175448][1065.134996167] 2025-12-04T09:10:29.1757370Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:10:29.1760233Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_decomp.py', '-m', 'serial', '--shard-id=12', '--num-shards=22', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:10:29.175749] 2025-12-04T09:10:45.6674786Z 2025-12-04T09:10:45.6675578Z test_decomp 12/22 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_12.22_09d4a096062ecff8_.log 2025-12-04T09:10:45.6676213Z Running 0 items in this shard: 2025-12-04T09:10:45.6676394Z 2025-12-04T09:10:45.6676621Z Finished test_decomp 12/22 ... [2025-12-04 09:10:45.666865][1081.626411758], took 0.27min 2025-12-04T09:10:45.6709783Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-7a98f01e940d9313.xml 2025-12-04T09:10:45.7425798Z Running test_decomp 18/22 ... [2025-12-04 09:10:45.742353][1081.701901307] 2025-12-04T09:10:45.7426201Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:10:45.7429127Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_decomp.py', '-m', 'serial', '--shard-id=18', '--num-shards=22', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:10:45.742651] 2025-12-04T09:11:02.2108079Z 2025-12-04T09:11:02.2108811Z test_decomp 18/22 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_18.22_b7c40422afc58245_.log 2025-12-04T09:11:02.2109454Z Running 0 items in this shard: 2025-12-04T09:11:02.2109626Z 2025-12-04T09:11:02.2109840Z Finished test_decomp 18/22 ... [2025-12-04 09:11:02.210560][1098.1701062], took 0.27min 2025-12-04T09:11:02.2146030Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-21da0f73c4549c1e.xml 2025-12-04T09:11:02.2898382Z Running test_decomp 20/22 ... [2025-12-04 09:11:02.289599][1098.249148072] 2025-12-04T09:11:02.2898774Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:11:02.2901508Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_decomp.py', '-m', 'serial', '--shard-id=20', '--num-shards=22', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:11:02.289890] 2025-12-04T09:11:19.0307015Z 2025-12-04T09:11:19.0307776Z test_decomp 20/22 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_20.22_2b2eda3179b882c5_.log 2025-12-04T09:11:19.0308417Z Running 0 items in this shard: 2025-12-04T09:11:19.0308625Z 2025-12-04T09:11:19.0308880Z Finished test_decomp 20/22 ... [2025-12-04 09:11:19.030480][1114.990024636], took 0.28min 2025-12-04T09:11:19.0346202Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-ac6cc9ee3eedc42e.xml 2025-12-04T09:11:19.1224369Z Running test_ops 2/9 ... [2025-12-04 09:11:19.122209][1115.081757342] 2025-12-04T09:11:19.1224774Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:11:19.1227749Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops.py', '-m', 'serial', '--shard-id=2', '--num-shards=9', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:11:19.122504] 2025-12-04T09:12:10.8556437Z 2025-12-04T09:12:10.8557389Z test_ops 2/9 was successful, full logs can be found in artifacts with path test/test-reports/test_ops_2.9_5593faa30198642f_.log 2025-12-04T09:12:10.8558045Z Running 0 items in this shard: 2025-12-04T09:12:10.8558220Z 2025-12-04T09:12:10.8558443Z Finished test_ops 2/9 ... [2025-12-04 09:12:10.855363][1166.814906727], took 0.86min 2025-12-04T09:12:10.8599863Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops/test_ops-6e21a7eebf8ff2d6.xml 2025-12-04T09:12:10.9394801Z Running test_ops 8/9 ... [2025-12-04 09:12:10.939226][1166.898772335] 2025-12-04T09:12:10.9395200Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:12:10.9397999Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops.py', '-m', 'serial', '--shard-id=8', '--num-shards=9', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:12:10.939536] 2025-12-04T09:13:03.0285022Z 2025-12-04T09:13:03.0285698Z test_ops 8/9 was successful, full logs can be found in artifacts with path test/test-reports/test_ops_8.9_4a7c51b6bf787854_.log 2025-12-04T09:13:03.0287489Z Running 0 items in this shard: 2025-12-04T09:13:03.0287659Z 2025-12-04T09:13:03.0288100Z Finished test_ops 8/9 ... [2025-12-04 09:13:03.028234][1218.987779939], took 0.87min 2025-12-04T09:13:03.0328444Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops/test_ops-3e23a1ec37f874be.xml 2025-12-04T09:13:03.1210939Z Running functorch/test_dims 1/1 ... [2025-12-04 09:13:03.120835][1219.080383201] 2025-12-04T09:13:03.1211372Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:13:03.1213967Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_dims.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:13:03.121125] 2025-12-04T09:13:06.0801761Z 2025-12-04T09:13:06.0802579Z functorch/test_dims 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_dims_1.1_b36a72e7d9c2bfba_.log 2025-12-04T09:13:06.0803329Z Running 0 items in this shard: 2025-12-04T09:13:06.0803514Z 2025-12-04T09:13:06.0803772Z Finished functorch/test_dims 1/1 ... [2025-12-04 09:13:06.079966][1222.039511913], took 0.05min 2025-12-04T09:13:06.0845721Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_dims/functorch.test_dims-46a4cc4d3955dce2.xml 2025-12-04T09:13:06.1121700Z Running test_torchfuzz_repros 1/1 ... [2025-12-04 09:13:06.111972][1222.071521052] 2025-12-04T09:13:06.1122139Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:13:06.1125169Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_torchfuzz_repros.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:13:06.112263] 2025-12-04T09:13:08.8346727Z 2025-12-04T09:13:08.8347643Z test_torchfuzz_repros 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_torchfuzz_repros_1.1_9f3f619c966168b6_.log 2025-12-04T09:13:08.8348384Z Running 0 items in this shard: 2025-12-04T09:13:08.8348558Z 2025-12-04T09:13:08.8348830Z Finished test_torchfuzz_repros 1/1 ... [2025-12-04 09:13:08.834461][1224.794006711], took 0.05min 2025-12-04T09:13:08.8391097Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_torchfuzz_repros/test_torchfuzz_repros-8e06ea0d070c14c7.xml 2025-12-04T09:13:08.8652192Z Running inductor/test_torchinductor 1/2 ... [2025-12-04 09:13:08.864975][1224.824523583] 2025-12-04T09:13:08.8652667Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:13:08.8655682Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor.py', '-m', 'serial', '--shard-id=1', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:13:08.865274] 2025-12-04T09:13:16.0223084Z 2025-12-04T09:13:16.0223983Z inductor/test_torchinductor 1/2 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_1.2_0c18f3f955f52c94_.log 2025-12-04T09:13:16.0224765Z Running 0 items in this shard: 2025-12-04T09:13:16.0224943Z 2025-12-04T09:13:16.0225240Z Finished inductor/test_torchinductor 1/2 ... [2025-12-04 09:13:16.022063][1231.981609789], took 0.12min 2025-12-04T09:13:16.0269070Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-9fcfc42987163c90.xml 2025-12-04T09:13:16.1036623Z Running inductor/test_torchinductor 2/2 ... [2025-12-04 09:13:16.103409][1232.062957834] 2025-12-04T09:13:16.1037078Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:13:16.1040055Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor.py', '-m', 'serial', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:13:16.103697] 2025-12-04T09:13:23.2409720Z 2025-12-04T09:13:23.2410634Z inductor/test_torchinductor 2/2 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_2.2_88a9d2ce75af7d9b_.log 2025-12-04T09:13:23.2411419Z Running 0 items in this shard: 2025-12-04T09:13:23.2411594Z 2025-12-04T09:13:23.2411900Z Finished inductor/test_torchinductor 2/2 ... [2025-12-04 09:13:23.240708][1239.200255133], took 0.12min 2025-12-04T09:13:23.2457772Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-6af558fb08ec07a5.xml 2025-12-04T09:13:23.3108970Z Running inductor/test_kernel_benchmark 1/1 ... [2025-12-04 09:13:23.310663][1239.270211348] 2025-12-04T09:13:23.3109475Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:13:23.3112565Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_kernel_benchmark.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:13:23.310958] 2025-12-04T09:13:28.5946012Z 2025-12-04T09:13:28.5947020Z inductor/test_kernel_benchmark 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_kernel_benchmark_1.1_3e5069f4126a6335_.log 2025-12-04T09:13:28.5947804Z Running 0 items in this shard: 2025-12-04T09:13:28.5947975Z 2025-12-04T09:13:28.5948275Z Finished inductor/test_kernel_benchmark 1/1 ... [2025-12-04 09:13:28.594379][1244.553925685], took 0.09min 2025-12-04T09:13:28.5994486Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_kernel_benchmark/inductor.test_kernel_benchmark-1b89aea89d8a1e3b.xml 2025-12-04T09:13:28.6267477Z Running inductor/test_torchinductor_opinfo 1/13 ... [2025-12-04 09:13:28.626536][1244.586084529] 2025-12-04T09:13:28.6267974Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:13:28.6270874Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '-m', 'serial', '--shard-id=1', '--num-shards=13', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:13:28.626833] 2025-12-04T09:13:40.3867265Z 2025-12-04T09:13:40.3868237Z inductor/test_torchinductor_opinfo 1/13 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_1.13_8df0b8edc083f301_.log 2025-12-04T09:13:40.3869098Z Running 0 items in this shard: 2025-12-04T09:13:40.3869272Z 2025-12-04T09:13:40.3869597Z Finished inductor/test_torchinductor_opinfo 1/13 ... [2025-12-04 09:13:40.386488][1256.346034169], took 0.20min 2025-12-04T09:13:40.3918226Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-c8a0664bdf5f16cf.xml 2025-12-04T09:13:40.4543664Z Running inductor/test_torchinductor_opinfo 5/13 ... [2025-12-04 09:13:40.454134][1256.413682176] 2025-12-04T09:13:40.4544183Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:13:40.4547073Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '-m', 'serial', '--shard-id=5', '--num-shards=13', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:13:40.454430] 2025-12-04T09:13:52.3140325Z 2025-12-04T09:13:52.3141638Z inductor/test_torchinductor_opinfo 5/13 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_5.13_31f48df8ddc8d1c5_.log 2025-12-04T09:13:52.3142486Z Running 0 items in this shard: 2025-12-04T09:13:52.3142680Z 2025-12-04T09:13:52.3143013Z Finished inductor/test_torchinductor_opinfo 5/13 ... [2025-12-04 09:13:52.313806][1268.273352667], took 0.20min 2025-12-04T09:13:52.3191380Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-4be68e197594ecf9.xml 2025-12-04T09:13:52.3936471Z Running inductor/test_torchinductor_opinfo 7/13 ... [2025-12-04 09:13:52.393403][1268.352951309] 2025-12-04T09:13:52.3936970Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:13:52.3939552Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '-m', 'serial', '--shard-id=7', '--num-shards=13', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:13:52.393696] 2025-12-04T09:14:04.2224864Z 2025-12-04T09:14:04.2225809Z inductor/test_torchinductor_opinfo 7/13 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_7.13_ce10eb09f3c1ae50_.log 2025-12-04T09:14:04.2226632Z Running 0 items in this shard: 2025-12-04T09:14:04.2226805Z 2025-12-04T09:14:04.2227131Z Finished inductor/test_torchinductor_opinfo 7/13 ... [2025-12-04 09:14:04.222270][1280.181816062], took 0.20min 2025-12-04T09:14:04.2277105Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-4814c98fafa7f915.xml 2025-12-04T09:14:04.3132273Z Running inductor/test_torchinductor_opinfo 13/13 ... [2025-12-04 09:14:04.312969][1280.272517465] 2025-12-04T09:14:04.3132804Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:14:04.3135416Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '-m', 'serial', '--shard-id=13', '--num-shards=13', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:14:04.313271] 2025-12-04T09:14:16.1686467Z 2025-12-04T09:14:16.1687428Z inductor/test_torchinductor_opinfo 13/13 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_13.13_23e880810105a276_.log 2025-12-04T09:14:16.1688235Z Running 0 items in this shard: 2025-12-04T09:14:16.1688417Z 2025-12-04T09:14:16.1688729Z Finished inductor/test_torchinductor_opinfo 13/13 ... [2025-12-04 09:14:16.168397][1292.12794329], took 0.20min 2025-12-04T09:14:16.1742544Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-e6a790fd2716c9ce.xml 2025-12-04T09:14:16.2517196Z Running inductor/test_layout_optim 1/1 ... [2025-12-04 09:14:16.251469][1292.211017576] 2025-12-04T09:14:16.2517660Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:14:16.2520353Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_layout_optim.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:14:16.251761] 2025-12-04T09:14:21.5159801Z 2025-12-04T09:14:21.5160642Z inductor/test_layout_optim 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_layout_optim_1.1_01519e00b3bd1c88_.log 2025-12-04T09:14:21.5161383Z Running 0 items in this shard: 2025-12-04T09:14:21.5161563Z 2025-12-04T09:14:21.5161843Z Finished inductor/test_layout_optim 1/1 ... [2025-12-04 09:14:21.515752][1297.475298515], took 0.09min 2025-12-04T09:14:21.5214988Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_layout_optim/inductor.test_layout_optim-0bd77606ca486567.xml 2025-12-04T09:14:21.5475783Z Running inductor/test_aot_inductor_arrayref 2/2 ... [2025-12-04 09:14:21.547367][1297.50691504] 2025-12-04T09:14:21.5476262Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:14:21.5479065Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_aot_inductor_arrayref.py', '-m', 'serial', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:14:21.547658] 2025-12-04T09:14:27.7507768Z 2025-12-04T09:14:27.7508739Z inductor/test_aot_inductor_arrayref 2/2 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_aot_inductor_arrayref_2.2_8efae0aa65170a95_.log 2025-12-04T09:14:27.7509572Z Running 0 items in this shard: 2025-12-04T09:14:27.7509746Z 2025-12-04T09:14:27.7510068Z Finished inductor/test_aot_inductor_arrayref 2/2 ... [2025-12-04 09:14:27.750554][1303.710100187], took 0.10min 2025-12-04T09:14:27.7564787Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor_arrayref/inductor.test_aot_inductor_arrayref-f384b91aa578f585.xml 2025-12-04T09:14:27.8241848Z Running inductor/test_torchinductor_strided_blocks 1/1 ... [2025-12-04 09:14:27.823937][1303.78348518] 2025-12-04T09:14:27.8242362Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:14:27.8245352Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_strided_blocks.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:14:27.824236] 2025-12-04T09:14:34.0089551Z 2025-12-04T09:14:34.0091046Z inductor/test_torchinductor_strided_blocks 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_strided_blocks_1.1_79b6d86fa53da74b_.log 2025-12-04T09:14:34.0091941Z Running 0 items in this shard: 2025-12-04T09:14:34.0092117Z 2025-12-04T09:14:34.0092460Z Finished inductor/test_torchinductor_strided_blocks 1/1 ... [2025-12-04 09:14:34.008706][1309.968251853], took 0.10min 2025-12-04T09:14:34.0147665Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_strided_blocks/inductor.test_torchinductor_strided_blocks-ea8814c1f4889964.xml 2025-12-04T09:14:34.0908642Z Running test_custom_ops 1/1 ... [2025-12-04 09:14:34.090624][1310.050172248] 2025-12-04T09:14:34.0909329Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:14:34.0912419Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_custom_ops.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:14:34.090919] 2025-12-04T09:14:38.3987516Z 2025-12-04T09:14:38.3988610Z test_custom_ops 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_custom_ops_1.1_57e8b64e81978c4f_.log 2025-12-04T09:14:38.3989382Z Running 0 items in this shard: 2025-12-04T09:14:38.3989556Z 2025-12-04T09:14:38.3989794Z Finished test_custom_ops 1/1 ... [2025-12-04 09:14:38.398465][1314.358011211], took 0.07min 2025-12-04T09:14:38.4047099Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_custom_ops/test_custom_ops-efe082e4ab0a44ae.xml 2025-12-04T09:14:38.4302159Z Running test_content_store 1/1 ... [2025-12-04 09:14:38.429962][1314.389510707] 2025-12-04T09:14:38.4302860Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:14:38.4306217Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_content_store.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:14:38.430257] 2025-12-04T09:14:41.5817175Z 2025-12-04T09:14:41.5818301Z test_content_store 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_content_store_1.1_abe40a032272748f_.log 2025-12-04T09:14:41.5819054Z Running 0 items in this shard: 2025-12-04T09:14:41.5819239Z 2025-12-04T09:14:41.5819495Z Finished test_content_store 1/1 ... [2025-12-04 09:14:41.581509][1317.541054956], took 0.05min 2025-12-04T09:14:41.5878315Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_content_store/test_content_store-e26a61a7c77ed768.xml 2025-12-04T09:14:41.6136174Z Running inductor/test_flex_decoding 1/3 ... [2025-12-04 09:14:41.613353][1317.572901653] 2025-12-04T09:14:41.6136970Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:14:41.6139889Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_flex_decoding.py', '-m', 'serial', '--shard-id=1', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:14:41.613649] 2025-12-04T09:14:45.5240712Z 2025-12-04T09:14:45.5241990Z inductor/test_flex_decoding 1/3 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_flex_decoding_1.3_30d64424b74ff5a2_.log 2025-12-04T09:14:45.5242787Z Running 0 items in this shard: 2025-12-04T09:14:45.5242965Z 2025-12-04T09:14:45.5243267Z Finished inductor/test_flex_decoding 1/3 ... [2025-12-04 09:14:45.523839][1321.483385797], took 0.07min 2025-12-04T09:14:45.5303005Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_flex_decoding/inductor.test_flex_decoding-7745b7464b29f22e.xml 2025-12-04T09:14:45.5557433Z Running inductor/test_flex_decoding 3/3 ... [2025-12-04 09:14:45.555492][1321.515040638] 2025-12-04T09:14:45.5558193Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:14:45.5561076Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_flex_decoding.py', '-m', 'serial', '--shard-id=3', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:14:45.555786] 2025-12-04T09:14:49.4490879Z 2025-12-04T09:14:49.4492135Z inductor/test_flex_decoding 3/3 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_flex_decoding_3.3_27519a12c6ab0e36_.log 2025-12-04T09:14:49.4492937Z Running 0 items in this shard: 2025-12-04T09:14:49.4493115Z 2025-12-04T09:14:49.4493545Z Finished inductor/test_flex_decoding 3/3 ... [2025-12-04 09:14:49.448847][1325.408393765], took 0.06min 2025-12-04T09:14:49.4554791Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_flex_decoding/inductor.test_flex_decoding-ba88f8b8af9ad9df.xml 2025-12-04T09:14:49.4817095Z Running inductor/test_deterministic 3/3 ... [2025-12-04 09:14:49.481457][1325.44100544] 2025-12-04T09:14:49.4817860Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:14:49.4820689Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_deterministic.py', '-m', 'serial', '--shard-id=3', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:14:49.481751] 2025-12-04T09:14:54.7578071Z 2025-12-04T09:14:54.7579334Z inductor/test_deterministic 3/3 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_deterministic_3.3_344aea65313e5096_.log 2025-12-04T09:14:54.7580540Z Running 0 items in this shard: 2025-12-04T09:14:54.7580884Z 2025-12-04T09:14:54.7581192Z Finished inductor/test_deterministic 3/3 ... [2025-12-04 09:14:54.757561][1330.717106603], took 0.09min 2025-12-04T09:14:54.7643161Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_deterministic/inductor.test_deterministic-5cb12759f42ba3ba.xml 2025-12-04T09:14:54.7856837Z Running inductor/test_b2b_gemm 1/1 ... [2025-12-04 09:14:54.785453][1330.745001533] 2025-12-04T09:14:54.7857563Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:14:54.7860613Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_b2b_gemm.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:14:54.785745] 2025-12-04T09:15:00.0269859Z 2025-12-04T09:15:00.0270709Z inductor/test_b2b_gemm 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_b2b_gemm_1.1_fd238e41e950af5f_.log 2025-12-04T09:15:00.0271446Z Running 0 items in this shard: 2025-12-04T09:15:00.0271629Z 2025-12-04T09:15:00.0271899Z Finished inductor/test_b2b_gemm 1/1 ... [2025-12-04 09:15:00.026758][1335.986304431], took 0.09min 2025-12-04T09:15:00.0335742Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_b2b_gemm/inductor.test_b2b_gemm-2ff610447b04c70b.xml 2025-12-04T09:15:00.0681754Z Running export/test_tree_utils 1/1 ... [2025-12-04 09:15:00.067971][1336.027519736] 2025-12-04T09:15:00.0682207Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:15:00.0685159Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'export/test_tree_utils.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:15:00.068255] 2025-12-04T09:15:02.7991339Z 2025-12-04T09:15:02.7992319Z export/test_tree_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_tree_utils_1.1_763b13f8fc426b15_.log 2025-12-04T09:15:02.7993068Z Running 0 items in this shard: 2025-12-04T09:15:02.7993250Z 2025-12-04T09:15:02.7993515Z Finished export/test_tree_utils 1/1 ... [2025-12-04 09:15:02.798900][1338.758446573], took 0.05min 2025-12-04T09:15:02.8060091Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/export.test_tree_utils/export.test_tree_utils-05f2e9a7741495f8.xml 2025-12-04T09:15:02.8355759Z Running inductor/test_triton_wrapper 1/1 ... [2025-12-04 09:15:02.835306][1338.794852901] 2025-12-04T09:15:02.8356238Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:15:02.8358897Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_triton_wrapper.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:15:02.835627] 2025-12-04T09:15:08.0758874Z 2025-12-04T09:15:08.0759742Z inductor/test_triton_wrapper 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_triton_wrapper_1.1_1526c6966d209d73_.log 2025-12-04T09:15:08.0760506Z Running 0 items in this shard: 2025-12-04T09:15:08.0760682Z 2025-12-04T09:15:08.0760968Z Finished inductor/test_triton_wrapper 1/1 ... [2025-12-04 09:15:08.075641][1344.035188431], took 0.09min 2025-12-04T09:15:08.0827334Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_triton_wrapper/inductor.test_triton_wrapper-c2ec4261f482965d.xml 2025-12-04T09:15:08.1184909Z Running inductor/test_static_cuda_launcher 1/1 ... [2025-12-04 09:15:08.118279][1344.077827258] 2025-12-04T09:15:08.1185668Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:15:08.1188423Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_static_cuda_launcher.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:15:08.118563] 2025-12-04T09:15:13.4022260Z 2025-12-04T09:15:13.4023064Z inductor/test_static_cuda_launcher 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_static_cuda_launcher_1.1_c765f37b7d99c783_.log 2025-12-04T09:15:13.4023741Z Running 0 items in this shard: 2025-12-04T09:15:13.4023876Z 2025-12-04T09:15:13.4024126Z Finished inductor/test_static_cuda_launcher 1/1 ... [2025-12-04 09:15:13.401965][1349.361511908], took 0.09min 2025-12-04T09:15:13.4092007Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_static_cuda_launcher/inductor.test_static_cuda_launcher-49f744a2d7796139.xml 2025-12-04T09:15:13.4444183Z Running export/test_dynamic_shapes 1/1 ... [2025-12-04 09:15:13.444195][1349.403743854] 2025-12-04T09:15:13.4444635Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:15:13.4447589Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'export/test_dynamic_shapes.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:15:13.444494] 2025-12-04T09:15:16.1238307Z 2025-12-04T09:15:16.1239300Z export/test_dynamic_shapes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_dynamic_shapes_1.1_09e3ef98478a54cd_.log 2025-12-04T09:15:16.1240182Z Running 0 items in this shard: 2025-12-04T09:15:16.1240500Z 2025-12-04T09:15:16.1240795Z Finished export/test_dynamic_shapes 1/1 ... [2025-12-04 09:15:16.123630][1352.083176261], took 0.04min 2025-12-04T09:15:16.1309785Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/export.test_dynamic_shapes/export.test_dynamic_shapes-66fe2c3481907496.xml 2025-12-04T09:15:16.1595680Z Running dynamo/test_sdpa 1/1 ... [2025-12-04 09:15:16.159350][1352.118898778] 2025-12-04T09:15:16.1596120Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:15:16.1599013Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_sdpa.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:15:16.159642] 2025-12-04T09:15:19.1088035Z 2025-12-04T09:15:19.1088855Z dynamo/test_sdpa 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_sdpa_1.1_6e4183b888ceacf8_.log 2025-12-04T09:15:19.1089576Z Running 0 items in this shard: 2025-12-04T09:15:19.1089753Z 2025-12-04T09:15:19.1090015Z Finished dynamo/test_sdpa 1/1 ... [2025-12-04 09:15:19.108601][1355.068146979], took 0.05min 2025-12-04T09:15:19.1161093Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_sdpa/dynamo.test_sdpa-bdc7e48df97c3ca8.xml 2025-12-04T09:15:19.1507505Z Running inductor/test_aot_inductor_package 1/1 ... [2025-12-04 09:15:19.150518][1355.110066178] 2025-12-04T09:15:19.1507990Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:15:19.1510736Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_aot_inductor_package.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:15:19.150817] 2025-12-04T09:15:24.5185523Z 2025-12-04T09:15:24.5186803Z inductor/test_aot_inductor_package 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_aot_inductor_package_1.1_2e32c0b0033ee866_.log 2025-12-04T09:15:24.5187663Z Running 0 items in this shard: 2025-12-04T09:15:24.5187838Z 2025-12-04T09:15:24.5188152Z Finished inductor/test_aot_inductor_package 1/1 ... [2025-12-04 09:15:24.518322][1360.477868799], took 0.09min 2025-12-04T09:15:24.5261135Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor_package/inductor.test_aot_inductor_package-3ed92bfad98c98b2.xml 2025-12-04T09:15:24.5537693Z Running inductor/test_compiled_optimizers 3/3 ... [2025-12-04 09:15:24.553538][1360.513085753] 2025-12-04T09:15:24.5538185Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:15:24.5541005Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_compiled_optimizers.py', '-m', 'serial', '--shard-id=3', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:15:24.553830] 2025-12-04T09:15:32.1154424Z 2025-12-04T09:15:32.1155357Z inductor/test_compiled_optimizers 3/3 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_compiled_optimizers_3.3_e91eedd79d0b2aa7_.log 2025-12-04T09:15:32.1156174Z Running 0 items in this shard: 2025-12-04T09:15:32.1156352Z 2025-12-04T09:15:32.1156673Z Finished inductor/test_compiled_optimizers 3/3 ... [2025-12-04 09:15:32.115218][1368.074763665], took 0.13min 2025-12-04T09:15:32.1230626Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_compiled_optimizers/inductor.test_compiled_optimizers-fc3fdb63f48e8ded.xml 2025-12-04T09:15:32.2025340Z Running inductor/test_aot_inductor_utils 1/1 ... [2025-12-04 09:15:32.202294][1368.161842025] 2025-12-04T09:15:32.2025874Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:15:32.2028561Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_aot_inductor_utils.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:15:32.202588] 2025-12-04T09:15:37.4654906Z 2025-12-04T09:15:37.4655708Z inductor/test_aot_inductor_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_aot_inductor_utils_1.1_3c60e8e2309ff072_.log 2025-12-04T09:15:37.4656422Z Running 0 items in this shard: 2025-12-04T09:15:37.4656592Z 2025-12-04T09:15:37.4656905Z Finished inductor/test_aot_inductor_utils 1/1 ... [2025-12-04 09:15:37.465263][1373.424809984], took 0.09min 2025-12-04T09:15:37.4732592Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor_utils/inductor.test_aot_inductor_utils-6fb3d946f8c4ffd3.xml 2025-12-04T09:15:37.4995173Z Running inductor/test_control_flow 3/5 ... [2025-12-04 09:15:37.499295][1373.458843581] 2025-12-04T09:15:37.4995635Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:15:37.4998654Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_control_flow.py', '-m', 'serial', '--shard-id=3', '--num-shards=5', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:15:37.499598] 2025-12-04T09:15:43.7993716Z 2025-12-04T09:15:43.7994588Z inductor/test_control_flow 3/5 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_control_flow_3.5_33c5414d09e1fd98_.log 2025-12-04T09:15:43.7995353Z Running 0 items in this shard: 2025-12-04T09:15:43.7995527Z 2025-12-04T09:15:43.7995808Z Finished inductor/test_control_flow 3/5 ... [2025-12-04 09:15:43.799157][1379.758703624], took 0.10min 2025-12-04T09:15:43.8073130Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_control_flow/inductor.test_control_flow-8be7394d621c6360.xml 2025-12-04T09:15:43.8786014Z Running test_mkl_verbose 1/1 ... [2025-12-04 09:15:43.878379][1379.837927073] 2025-12-04T09:15:43.8786425Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:15:43.8789246Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_mkl_verbose.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:15:43.878675] 2025-12-04T09:15:46.5843926Z 2025-12-04T09:15:46.5844695Z test_mkl_verbose 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_mkl_verbose_1.1_4ffe3bbc2c20413d_.log 2025-12-04T09:15:46.5845347Z Running 0 items in this shard: 2025-12-04T09:15:46.5845551Z 2025-12-04T09:15:46.5845795Z Finished test_mkl_verbose 1/1 ... [2025-12-04 09:15:46.584185][1382.543731523], took 0.05min 2025-12-04T09:15:46.5925165Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_mkl_verbose/test_mkl_verbose-13dd2bd43e71ccf7.xml 2025-12-04T09:15:46.6215393Z Running test_comparison_utils 1/1 ... [2025-12-04 09:15:46.621343][1382.580891517] 2025-12-04T09:15:46.6215751Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:15:46.6218670Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_comparison_utils.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:15:46.621642] 2025-12-04T09:15:49.3351336Z 2025-12-04T09:15:49.3352219Z test_comparison_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_comparison_utils_1.1_03659744aead51ab_.log 2025-12-04T09:15:49.3352942Z Running 0 items in this shard: 2025-12-04T09:15:49.3353131Z 2025-12-04T09:15:49.3353413Z Finished test_comparison_utils 1/1 ... [2025-12-04 09:15:49.334899][1385.294445778], took 0.05min 2025-12-04T09:15:49.3432746Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_comparison_utils/test_comparison_utils-54963d4d516ccde5.xml 2025-12-04T09:15:49.3658918Z Running functorch/test_ac_logging 1/1 ... [2025-12-04 09:15:49.365681][1385.325229187] 2025-12-04T09:15:49.3659377Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:15:49.3662191Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_ac_logging.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:15:49.365972] 2025-12-04T09:15:52.0817092Z 2025-12-04T09:15:52.0818006Z functorch/test_ac_logging 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_ac_logging_1.1_fe60cab7003c0ff7_.log 2025-12-04T09:15:52.0818738Z Running 0 items in this shard: 2025-12-04T09:15:52.0818911Z 2025-12-04T09:15:52.0819186Z Finished functorch/test_ac_logging 1/1 ... [2025-12-04 09:15:52.081478][1388.04102426], took 0.05min 2025-12-04T09:15:52.0900068Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_ac_logging/functorch.test_ac_logging-2dd9f98043af7cf7.xml 2025-12-04T09:15:52.1147486Z Running test_mkldnn_verbose 1/1 ... [2025-12-04 09:15:52.114521][1388.074069275] 2025-12-04T09:15:52.1147949Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:15:52.1150743Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_mkldnn_verbose.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:15:52.114814] 2025-12-04T09:15:54.8385791Z 2025-12-04T09:15:54.8386613Z test_mkldnn_verbose 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_mkldnn_verbose_1.1_24beac6fb8e8b85c_.log 2025-12-04T09:15:54.8387316Z Running 0 items in this shard: 2025-12-04T09:15:54.8387499Z 2025-12-04T09:15:54.8387772Z Finished test_mkldnn_verbose 1/1 ... [2025-12-04 09:15:54.838321][1390.797867064], took 0.05min 2025-12-04T09:15:54.8469041Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_mkldnn_verbose/test_mkldnn_verbose-d7761a02a830b7c0.xml 2025-12-04T09:15:54.8721687Z Running test_utils_config_module 1/1 ... [2025-12-04 09:15:54.871934][1390.831482312] 2025-12-04T09:15:54.8722243Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:15:54.8725349Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_utils_config_module.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:15:54.872243] 2025-12-04T09:15:57.6195806Z 2025-12-04T09:15:57.6196643Z test_utils_config_module 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_utils_config_module_1.1_2fb5daf1593c0cf5_.log 2025-12-04T09:15:57.6197376Z Running 0 items in this shard: 2025-12-04T09:15:57.6197551Z 2025-12-04T09:15:57.6197815Z Finished test_utils_config_module 1/1 ... [2025-12-04 09:15:57.619370][1393.578917195], took 0.05min 2025-12-04T09:15:57.6281960Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_utils_config_module/test_utils_config_module-472c75b33322e8a7.xml 2025-12-04T09:15:57.6588247Z Running test_hop_infra 1/1 ... [2025-12-04 09:15:57.658592][1393.618140949] 2025-12-04T09:15:57.6588707Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:15:57.6591678Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_hop_infra.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:15:57.658891] 2025-12-04T09:16:01.0941206Z 2025-12-04T09:16:01.0941969Z test_hop_infra 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_hop_infra_1.1_8ee2aa3c7e32f90d_.log 2025-12-04T09:16:01.0942619Z Running 0 items in this shard: 2025-12-04T09:16:01.0942801Z 2025-12-04T09:16:01.0943021Z Finished test_hop_infra 1/1 ... [2025-12-04 09:16:01.093923][1397.053469307], took 0.06min 2025-12-04T09:16:01.1028778Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_hop_infra/test_hop_infra-2ef307420583b1a3.xml 2025-12-04T09:16:01.1291681Z Running test_appending_byte_serializer 1/1 ... [2025-12-04 09:16:01.128932][1397.08848067] 2025-12-04T09:16:01.1292174Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:16:01.1295080Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_appending_byte_serializer.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:16:01.129229] 2025-12-04T09:16:03.8429103Z 2025-12-04T09:16:03.8430008Z test_appending_byte_serializer 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_appending_byte_serializer_1.1_64533d4e3bf3cda4_.log 2025-12-04T09:16:03.8430807Z Running 0 items in this shard: 2025-12-04T09:16:03.8430983Z 2025-12-04T09:16:03.8431278Z Finished test_appending_byte_serializer 1/1 ... [2025-12-04 09:16:03.842692][1399.802238517], took 0.05min 2025-12-04T09:16:03.8519056Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_appending_byte_serializer/test_appending_byte_serializer-7b18d6c4f7206cf9.xml 2025-12-04T09:16:03.8811109Z Running test_license 1/1 ... [2025-12-04 09:16:03.880888][1399.840436258] 2025-12-04T09:16:03.8811517Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:16:03.8814537Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_license.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:16:03.881183] 2025-12-04T09:16:06.5876554Z 2025-12-04T09:16:06.5877267Z test_license 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_license_1.1_702ac49c8c00b783_.log 2025-12-04T09:16:06.5878004Z Running 0 items in this shard: 2025-12-04T09:16:06.5878179Z 2025-12-04T09:16:06.5878400Z Finished test_license 1/1 ... [2025-12-04 09:16:06.587449][1402.54699571], took 0.05min 2025-12-04T09:16:06.5966803Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_license/test_license-787c69b7e1011532.xml 2025-12-04T09:16:06.6192784Z Running test_ao_sparsity 1/1 ... [2025-12-04 09:16:06.619066][1402.578615059] 2025-12-04T09:16:06.6193213Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:16:06.6196060Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ao_sparsity.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:16:06.619356] 2025-12-04T09:16:09.8632344Z 2025-12-04T09:16:09.8633177Z test_ao_sparsity 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_ao_sparsity_1.1_fc9d1e285ec3ec39_.log 2025-12-04T09:16:09.8633855Z Running 0 items in this shard: 2025-12-04T09:16:09.8634058Z 2025-12-04T09:16:09.8634307Z Finished test_ao_sparsity 1/1 ... [2025-12-04 09:16:09.862993][1405.822539134], took 0.05min 2025-12-04T09:16:09.8724274Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ao_sparsity/test_ao_sparsity-ce7eaf33f7dbaec9.xml 2025-12-04T09:16:09.9011745Z Running test_autoload 1/1 ... [2025-12-04 09:16:09.900936][1405.860483768] 2025-12-04T09:16:09.9012158Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:16:09.9015208Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_autoload.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:16:09.901232] 2025-12-04T09:16:12.6055516Z 2025-12-04T09:16:12.6056293Z test_autoload 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_autoload_1.1_b3134511ed88d0ac_.log 2025-12-04T09:16:12.6056987Z Running 0 items in this shard: 2025-12-04T09:16:12.6057191Z 2025-12-04T09:16:12.6057438Z Finished test_autoload 1/1 ... [2025-12-04 09:16:12.605353][1408.564899588], took 0.05min 2025-12-04T09:16:12.6147949Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_autoload/test_autoload-e264c515ca0f679c.xml 2025-12-04T09:16:12.6362680Z Running nn/attention/test_open_registry 1/1 ... [2025-12-04 09:16:12.636057][1408.595605066] 2025-12-04T09:16:12.6363157Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:16:12.6366119Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/attention/test_open_registry.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:16:12.636344] 2025-12-04T09:16:15.3531398Z 2025-12-04T09:16:15.3532823Z nn/attention/test_open_registry 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.attention.test_open_registry_1.1_da6b3f4cb0dc7340_.log 2025-12-04T09:16:15.3533749Z Running 0 items in this shard: 2025-12-04T09:16:15.3533922Z 2025-12-04T09:16:15.3534230Z Finished nn/attention/test_open_registry 1/1 ... [2025-12-04 09:16:15.352936][1411.312483363], took 0.05min 2025-12-04T09:16:15.3625150Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.attention.test_open_registry/nn.attention.test_open_registry-96e12afd5c2c5834.xml 2025-12-04T09:16:15.3902895Z Running test_as_strided 1/1 ... [2025-12-04 09:16:15.390082][1411.34963056] 2025-12-04T09:16:15.3903307Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:16:15.3906577Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_as_strided.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:16:15.390373] 2025-12-04T09:16:18.0778266Z 2025-12-04T09:16:18.0779051Z test_as_strided 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_as_strided_1.1_96a2fa8a410b6049_.log 2025-12-04T09:16:18.0779723Z Running 0 items in this shard: 2025-12-04T09:16:18.0779895Z 2025-12-04T09:16:18.0780146Z Finished test_as_strided 1/1 ... [2025-12-04 09:16:18.077601][1414.037147507], took 0.04min 2025-12-04T09:16:18.0875071Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_as_strided/test_as_strided-498f7726c4fd5386.xml 2025-12-04T09:16:18.1136945Z Running test_foreach 1/1 ... [2025-12-04 09:16:18.113485][1414.073033023] 2025-12-04T09:16:18.1137364Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:16:18.1140288Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_foreach.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:16:18.113774] 2025-12-04T09:16:29.7546063Z 2025-12-04T09:16:29.7546787Z test_foreach 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_foreach_1.1_4b853a43bd6e725e_.log 2025-12-04T09:16:29.7547430Z Running 0 items in this shard: 2025-12-04T09:16:29.7547602Z 2025-12-04T09:16:29.7547823Z Finished test_foreach 1/1 ... [2025-12-04 09:16:29.754387][1425.713934023], took 0.19min 2025-12-04T09:16:29.7643796Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_foreach/test_foreach-7dea0e9af6e9ef83.xml 2025-12-04T09:16:29.8427960Z Running xpu/test_gemm 1/1 ... [2025-12-04 09:16:29.842543][1425.802091502] 2025-12-04T09:16:29.8428635Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:16:29.8431488Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'xpu/test_gemm.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:16:29.842831] 2025-12-04T09:16:33.1823558Z 2025-12-04T09:16:33.1824613Z xpu/test_gemm 1/1 was successful, full logs can be found in artifacts with path test/test-reports/xpu.test_gemm_1.1_709617076e2e7b44_.log 2025-12-04T09:16:33.1825300Z Running 0 items in this shard: 2025-12-04T09:16:33.1825476Z 2025-12-04T09:16:33.1825698Z Finished xpu/test_gemm 1/1 ... [2025-12-04 09:16:33.182127][1429.141674463], took 0.06min 2025-12-04T09:16:33.1922884Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/xpu.test_gemm/xpu.test_gemm-e12f724e557e6d57.xml 2025-12-04T09:16:33.2192371Z Running higher_order_ops/test_print 1/1 ... [2025-12-04 09:16:33.218988][1429.178535632] 2025-12-04T09:16:33.2193104Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:16:33.2196168Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'higher_order_ops/test_print.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:16:33.219289] 2025-12-04T09:16:36.1824673Z 2025-12-04T09:16:36.1825911Z higher_order_ops/test_print 1/1 was successful, full logs can be found in artifacts with path test/test-reports/higher_order_ops.test_print_1.1_afd6752b8d2cc0f6_.log 2025-12-04T09:16:36.1826637Z Running 0 items in this shard: 2025-12-04T09:16:36.1826778Z 2025-12-04T09:16:36.1827005Z Finished higher_order_ops/test_print 1/1 ... [2025-12-04 09:16:36.182221][1432.141767357], took 0.05min 2025-12-04T09:16:36.1924821Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/higher_order_ops.test_print/higher_order_ops.test_print-4bc80d64e1280602.xml 2025-12-04T09:16:36.2183664Z Running test_per_overload_api 1/1 ... [2025-12-04 09:16:36.218119][1432.177667272] 2025-12-04T09:16:36.2184391Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:16:36.2187075Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_per_overload_api.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:16:36.218398] 2025-12-04T09:16:38.9188369Z 2025-12-04T09:16:38.9189546Z test_per_overload_api 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_per_overload_api_1.1_d91a23c6a925aea0_.log 2025-12-04T09:16:38.9190271Z Running 0 items in this shard: 2025-12-04T09:16:38.9190448Z 2025-12-04T09:16:38.9190710Z Finished test_per_overload_api 1/1 ... [2025-12-04 09:16:38.918628][1434.878174542], took 0.05min 2025-12-04T09:16:38.9289468Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_per_overload_api/test_per_overload_api-21ff1c6aab58c94d.xml 2025-12-04T09:16:38.9736158Z Running torch_np/numpy_tests/core/test_einsum 1/1 ... [2025-12-04 09:16:38.973364][1434.932912106] 2025-12-04T09:16:38.9736958Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:16:38.9739874Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_einsum.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:16:38.973659] 2025-12-04T09:16:41.7595978Z 2025-12-04T09:16:41.7597341Z torch_np/numpy_tests/core/test_einsum 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_einsum_1.1_720e264a46a74885_.log 2025-12-04T09:16:41.7598174Z Running 0 items in this shard: 2025-12-04T09:16:41.7598375Z 2025-12-04T09:16:41.7598706Z Finished torch_np/numpy_tests/core/test_einsum 1/1 ... [2025-12-04 09:16:41.759379][1437.7189257], took 0.05min 2025-12-04T09:16:41.7698392Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.core.test_einsum/torch_np.numpy_tests.core.test_einsum-e842edbf10eb231f.xml 2025-12-04T09:16:41.7937662Z Running test_out_dtype_op 1/1 ... [2025-12-04 09:16:41.793537][1437.753085153] 2025-12-04T09:16:41.7938331Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:16:41.7941569Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_out_dtype_op.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:16:41.793828] 2025-12-04T09:16:45.0985562Z 2025-12-04T09:16:45.0986599Z test_out_dtype_op 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_out_dtype_op_1.1_f2a6f18e42db3cfb_.log 2025-12-04T09:16:45.0987881Z Running 0 items in this shard: 2025-12-04T09:16:45.0988059Z 2025-12-04T09:16:45.0988303Z Finished test_out_dtype_op 1/1 ... [2025-12-04 09:16:45.098359][1441.057905959], took 0.06min 2025-12-04T09:16:45.1091042Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_out_dtype_op/test_out_dtype_op-000e0deed8d272bf.xml 2025-12-04T09:16:45.1332125Z Running torch_np/test_ufuncs_basic 1/1 ... [2025-12-04 09:16:45.132904][1441.092448082] 2025-12-04T09:16:45.1332862Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:16:45.1335998Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/test_ufuncs_basic.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:16:45.133258] 2025-12-04T09:16:48.3325359Z 2025-12-04T09:16:48.3326643Z torch_np/test_ufuncs_basic 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.test_ufuncs_basic_1.1_bfaee78e6ce1e6e9_.log 2025-12-04T09:16:48.3327393Z Running 0 items in this shard: 2025-12-04T09:16:48.3327574Z 2025-12-04T09:16:48.3327861Z Finished torch_np/test_ufuncs_basic 1/1 ... [2025-12-04 09:16:48.332309][1444.291855194], took 0.05min 2025-12-04T09:16:48.3433595Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.test_ufuncs_basic/torch_np.test_ufuncs_basic-f7d6a1f02b6919cc.xml 2025-12-04T09:16:48.3716220Z Running lazy/test_step_closures 1/1 ... [2025-12-04 09:16:48.371385][1444.330933538] 2025-12-04T09:16:48.3716925Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:16:48.3720201Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'lazy/test_step_closures.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:16:48.371683] 2025-12-04T09:16:51.1104105Z 2025-12-04T09:16:51.1105490Z lazy/test_step_closures 1/1 was successful, full logs can be found in artifacts with path test/test-reports/lazy.test_step_closures_1.1_868a341d9b1e0a2d_.log 2025-12-04T09:16:51.1106221Z Running 0 items in this shard: 2025-12-04T09:16:51.1106395Z 2025-12-04T09:16:51.1106670Z Finished lazy/test_step_closures 1/1 ... [2025-12-04 09:16:51.110192][1447.06973942], took 0.05min 2025-12-04T09:16:51.1210551Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/lazy.test_step_closures/lazy.test_step_closures-b9a26ea3a1d388cc.xml 2025-12-04T09:16:51.1477043Z Running functorch/dim/test_getsetitem 1/1 ... [2025-12-04 09:16:51.147466][1447.10701463] 2025-12-04T09:16:51.1477813Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:16:51.1480631Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/dim/test_getsetitem.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:16:51.147756] 2025-12-04T09:16:53.8961012Z 2025-12-04T09:16:53.8962303Z functorch/dim/test_getsetitem 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.dim.test_getsetitem_1.1_38009191dd39b63e_.log 2025-12-04T09:16:53.8963082Z Running 0 items in this shard: 2025-12-04T09:16:53.8963254Z 2025-12-04T09:16:53.8963549Z Finished functorch/dim/test_getsetitem 1/1 ... [2025-12-04 09:16:53.895869][1449.855415826], took 0.05min 2025-12-04T09:16:53.9069411Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.dim.test_getsetitem/functorch.dim.test_getsetitem-54dfd23e9b8e2df1.xml 2025-12-04T09:16:53.9338109Z Running test_ops_fwd_gradients 1/3 ... [2025-12-04 09:16:53.933564][1449.893112496] 2025-12-04T09:16:53.9339118Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:16:53.9341645Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops_fwd_gradients.py', '-m', 'serial', '--shard-id=1', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:16:53.933855] 2025-12-04T09:17:02.3774989Z 2025-12-04T09:17:02.3775826Z test_ops_fwd_gradients 1/3 was successful, full logs can be found in artifacts with path test/test-reports/test_ops_fwd_gradients_1.3_f370258098a4a3d8_.log 2025-12-04T09:17:02.3776514Z Running 0 items in this shard: 2025-12-04T09:17:02.3776691Z 2025-12-04T09:17:02.3776950Z Finished test_ops_fwd_gradients 1/3 ... [2025-12-04 09:17:02.377219][1458.33676593], took 0.14min 2025-12-04T09:17:02.3884317Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-4bb6f0d5c3842f89.xml 2025-12-04T09:17:02.4601830Z Running test_ops_fwd_gradients 2/3 ... [2025-12-04 09:17:02.459939][1458.419487156] 2025-12-04T09:17:02.4602251Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:17:02.4605559Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops_fwd_gradients.py', '-m', 'serial', '--shard-id=2', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:17:02.460236] 2025-12-04T09:17:10.9206562Z 2025-12-04T09:17:10.9207386Z test_ops_fwd_gradients 2/3 was successful, full logs can be found in artifacts with path test/test-reports/test_ops_fwd_gradients_2.3_222494559a713d2f_.log 2025-12-04T09:17:10.9208079Z Running 0 items in this shard: 2025-12-04T09:17:10.9208265Z 2025-12-04T09:17:10.9208531Z Finished test_ops_fwd_gradients 2/3 ... [2025-12-04 09:17:10.920387][1466.879933797], took 0.14min 2025-12-04T09:17:10.9315770Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-bc1bcd823497d313.xml 2025-12-04T09:17:11.0075296Z Running test_meta 4/4 ... [2025-12-04 09:17:11.007266][1466.966813513] 2025-12-04T09:17:11.0075691Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:17:11.0078374Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_meta.py', '-m', 'serial', '--shard-id=4', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:17:11.007564] 2025-12-04T09:18:10.0064832Z 2025-12-04T09:18:10.0067400Z test_meta 4/4 was successful, full logs can be found in artifacts with path test/test-reports/test_meta_4.4_42623e8fef4d35a7_.log 2025-12-04T09:18:10.0067939Z Running 0 items in this shard: 2025-12-04T09:18:10.0068121Z 2025-12-04T09:18:10.0068315Z Finished test_meta 4/4 ... [2025-12-04 09:18:10.006233][1525.965775407], took 0.98min 2025-12-04T09:18:10.0180585Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_meta/test_meta-b802ca42a17a2408.xml 2025-12-04T09:18:10.1081320Z Running test_nestedtensor 3/4 ... [2025-12-04 09:18:10.107906][1526.067453134] 2025-12-04T09:18:10.1081756Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:18:10.1084992Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_nestedtensor.py', '-m', 'serial', '--shard-id=3', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:18:10.108202] 2025-12-04T09:18:16.6751014Z 2025-12-04T09:18:16.6751828Z test_nestedtensor 3/4 was successful, full logs can be found in artifacts with path test/test-reports/test_nestedtensor_3.4_7d556cc06aefdeb1_.log 2025-12-04T09:18:16.6752934Z Running 0 items in this shard: 2025-12-04T09:18:16.6753116Z 2025-12-04T09:18:16.6753542Z Finished test_nestedtensor 3/4 ... [2025-12-04 09:18:16.674872][1532.634414697], took 0.11min 2025-12-04T09:18:16.6865386Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-49efc715bdff07e3.xml 2025-12-04T09:18:16.7620707Z Running test_nestedtensor 4/4 ... [2025-12-04 09:18:16.761817][1532.721362805] 2025-12-04T09:18:16.7621148Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:18:16.7624086Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_nestedtensor.py', '-m', 'serial', '--shard-id=4', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:18:16.762129] 2025-12-04T09:18:23.3676071Z 2025-12-04T09:18:23.3676888Z test_nestedtensor 4/4 was successful, full logs can be found in artifacts with path test/test-reports/test_nestedtensor_4.4_1ca35917d202dd6d_.log 2025-12-04T09:18:23.3677584Z Running 0 items in this shard: 2025-12-04T09:18:23.3677761Z 2025-12-04T09:18:23.3678016Z Finished test_nestedtensor 4/4 ... [2025-12-04 09:18:23.367363][1539.326907874], took 0.11min 2025-12-04T09:18:23.3793862Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-66fecba3f80d413a.xml 2025-12-04T09:18:23.4550592Z Running test_modules 2/2 ... [2025-12-04 09:18:23.454815][1539.414361254] 2025-12-04T09:18:23.4550999Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:18:23.4554158Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_modules.py', '-m', 'serial', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:18:23.455120] 2025-12-04T09:18:32.7058276Z 2025-12-04T09:18:32.7059010Z test_modules 2/2 was successful, full logs can be found in artifacts with path test/test-reports/test_modules_2.2_8119bdbc5009a9d0_.log 2025-12-04T09:18:32.7059655Z Running 0 items in this shard: 2025-12-04T09:18:32.7059831Z 2025-12-04T09:18:32.7060060Z Finished test_modules 2/2 ... [2025-12-04 09:18:32.705590][1548.665134214], took 0.15min 2025-12-04T09:18:32.7174358Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_modules/test_modules-dedf31149329b85c.xml 2025-12-04T09:18:32.7852818Z Running test_tensorboard 1/1 ... [2025-12-04 09:18:32.785021][1548.744567129] 2025-12-04T09:18:32.7853348Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:18:32.7856238Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_tensorboard.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:18:32.785330] 2025-12-04T09:18:35.8841050Z 2025-12-04T09:18:35.8841824Z test_tensorboard 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_tensorboard_1.1_cb20da95004ddae1_.log 2025-12-04T09:18:35.8842481Z Running 0 items in this shard: 2025-12-04T09:18:35.8842646Z 2025-12-04T09:18:35.8842882Z Finished test_tensorboard 1/1 ... [2025-12-04 09:18:35.883876][1551.843418965], took 0.05min 2025-12-04T09:18:35.8958977Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_tensorboard/test_tensorboard-2f994ca1a93f319e.xml 2025-12-04T09:18:35.9261655Z Running torch_np/numpy_tests/core/test_indexing 1/1 ... [2025-12-04 09:18:35.925936][1551.885482489] 2025-12-04T09:18:35.9262150Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:18:35.9265351Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_indexing.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:18:35.926234] 2025-12-04T09:18:38.7300966Z 2025-12-04T09:18:38.7301942Z torch_np/numpy_tests/core/test_indexing 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_indexing_1.1_ae4b7321e401d129_.log 2025-12-04T09:18:38.7302784Z Running 0 items in this shard: 2025-12-04T09:18:38.7302954Z 2025-12-04T09:18:38.7303282Z Finished torch_np/numpy_tests/core/test_indexing 1/1 ... [2025-12-04 09:18:38.729878][1554.689421559], took 0.05min 2025-12-04T09:18:38.7419719Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.core.test_indexing/torch_np.numpy_tests.core.test_indexing-273e2720b4753f9b.xml 2025-12-04T09:18:38.7681363Z Running test_futures 1/1 ... [2025-12-04 09:18:38.767919][1554.727465056] 2025-12-04T09:18:38.7681805Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:18:38.7684854Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_futures.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:18:38.768218] 2025-12-04T09:18:41.5193239Z 2025-12-04T09:18:41.5193990Z test_futures 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_futures_1.1_ab361dc5625ed4a3_.log 2025-12-04T09:18:41.5194632Z Running 0 items in this shard: 2025-12-04T09:18:41.5194815Z 2025-12-04T09:18:41.5195044Z Finished test_futures 1/1 ... [2025-12-04 09:18:41.519142][1557.478685911], took 0.05min 2025-12-04T09:18:41.5314827Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_futures/test_futures-e1ab9c636dcc5536.xml 2025-12-04T09:18:41.5637497Z Running nn/test_dropout 1/1 ... [2025-12-04 09:18:41.563512][1557.523058934] 2025-12-04T09:18:41.5637917Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:18:41.5640901Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/test_dropout.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:18:41.563817] 2025-12-04T09:18:44.7557905Z 2025-12-04T09:18:44.7558834Z nn/test_dropout 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.test_dropout_1.1_397a2054f8bff0d6_.log 2025-12-04T09:18:44.7559513Z Running 0 items in this shard: 2025-12-04T09:18:44.7560784Z 2025-12-04T09:18:44.7561309Z Finished nn/test_dropout 1/1 ... [2025-12-04 09:18:44.755583][1560.715127258], took 0.05min 2025-12-04T09:18:44.7679071Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.test_dropout/nn.test_dropout-3a6a2fbf43a8e461.xml 2025-12-04T09:18:44.7955662Z Running functorch/dim/test_split 1/1 ... [2025-12-04 09:18:44.795326][1560.754872107] 2025-12-04T09:18:44.7956128Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:18:44.7958952Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/dim/test_split.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:18:44.795612] 2025-12-04T09:18:47.5497672Z 2025-12-04T09:18:47.5498638Z functorch/dim/test_split 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.dim.test_split_1.1_822ccb8ad215b24b_.log 2025-12-04T09:18:47.5499384Z Running 0 items in this shard: 2025-12-04T09:18:47.5499568Z 2025-12-04T09:18:47.5499848Z Finished functorch/dim/test_split 1/1 ... [2025-12-04 09:18:47.549553][1563.509097315], took 0.05min 2025-12-04T09:18:47.5620841Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.dim.test_split/functorch.dim.test_split-8adecf644c26bf14.xml 2025-12-04T09:18:47.5879094Z Running torch_np/numpy_tests/lib/test_type_check 1/1 ... [2025-12-04 09:18:47.587685][1563.547230746] 2025-12-04T09:18:47.5879593Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:18:47.5882478Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/lib/test_type_check.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:18:47.587966] 2025-12-04T09:18:50.3700658Z 2025-12-04T09:18:50.3701847Z torch_np/numpy_tests/lib/test_type_check 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.lib.test_type_check_1.1_ccc3d1df2287267e_.log 2025-12-04T09:18:50.3702694Z Running 0 items in this shard: 2025-12-04T09:18:50.3702855Z 2025-12-04T09:18:50.3703129Z Finished torch_np/numpy_tests/lib/test_type_check 1/1 ... [2025-12-04 09:18:50.369854][1566.329398905], took 0.05min 2025-12-04T09:18:50.3825440Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.lib.test_type_check/torch_np.numpy_tests.lib.test_type_check-8efd39d4632ebedf.xml 2025-12-04T09:18:50.4069430Z Running cpp_extensions/test_libtorch_agnostic 1/1 ... [2025-12-04 09:18:50.406732][1566.366278644] 2025-12-04T09:18:50.4069937Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:18:50.4072795Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'cpp_extensions/test_libtorch_agnostic.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:18:50.407028] 2025-12-04T09:19:08.1740918Z 2025-12-04T09:19:08.1741837Z cpp_extensions/test_libtorch_agnostic 1/1 was successful, full logs can be found in artifacts with path test/test-reports/cpp_extensions.test_libtorch_agnostic_1.1_db47f5251d8f5d77_.log 2025-12-04T09:19:08.1742550Z Running 0 items in this shard: 2025-12-04T09:19:08.1742698Z 2025-12-04T09:19:08.1742966Z Finished cpp_extensions/test_libtorch_agnostic 1/1 ... [2025-12-04 09:19:08.174031][1584.133578319], took 0.30min 2025-12-04T09:19:08.1868716Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/cpp_extensions.test_libtorch_agnostic/cpp_extensions.test_libtorch_agnostic-9de1e21a730c2263.xml 2025-12-04T09:19:08.2655581Z Running profiler/test_execution_trace 1/1 ... [2025-12-04 09:19:08.265300][1584.224848489] 2025-12-04T09:19:08.2656090Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:19:08.2659054Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'profiler/test_execution_trace.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:19:08.265614] 2025-12-04T09:19:11.4422421Z 2025-12-04T09:19:11.4423337Z profiler/test_execution_trace 1/1 was successful, full logs can be found in artifacts with path test/test-reports/profiler.test_execution_trace_1.1_6be3fb6173e553da_.log 2025-12-04T09:19:11.4424133Z Running 0 items in this shard: 2025-12-04T09:19:11.4424311Z 2025-12-04T09:19:11.4424617Z Finished profiler/test_execution_trace 1/1 ... [2025-12-04 09:19:11.442004][1587.401549737], took 0.05min 2025-12-04T09:19:11.4552815Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/profiler.test_execution_trace/profiler.test_execution_trace-852e4822f8230313.xml 2025-12-04T09:19:11.4894670Z Running test_jit 1/1 ... [2025-12-04 09:19:11.489230][1587.448777848] 2025-12-04T09:19:11.4895434Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:19:11.4898885Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_jit.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:19:11.489588] 2025-12-04T09:19:19.5007600Z 2025-12-04T09:19:19.5008367Z test_jit 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_jit_1.1_235772887b92014d_.log 2025-12-04T09:19:19.5008902Z Running 0 items in this shard: 2025-12-04T09:19:19.5009052Z 2025-12-04T09:19:19.5009222Z Finished test_jit 1/1 ... [2025-12-04 09:19:19.500447][1595.459993144], took 0.13min 2025-12-04T09:19:19.5136932Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_jit/test_jit-711c2ccc279f7210.xml 2025-12-04T09:19:19.5759109Z Running test_datapipe 1/1 ... [2025-12-04 09:19:19.575670][1595.535218533] 2025-12-04T09:19:19.5759553Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:19:19.5762302Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_datapipe.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:19:19.575962] 2025-12-04T09:19:22.4037721Z 2025-12-04T09:19:22.4038521Z test_datapipe 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_datapipe_1.1_eeeeff3cb1cf07c0_.log 2025-12-04T09:19:22.4039185Z Running 0 items in this shard: 2025-12-04T09:19:22.4039362Z 2025-12-04T09:19:22.4039592Z Finished test_datapipe 1/1 ... [2025-12-04 09:19:22.403534][1598.363079565], took 0.05min 2025-12-04T09:19:22.4167462Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_datapipe/test_datapipe-0e1964a956d0a480.xml 2025-12-04T09:19:22.4462653Z Running test_numba_integration 1/1 ... [2025-12-04 09:19:22.446028][1598.405576949] 2025-12-04T09:19:22.4463132Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:19:22.4465607Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_numba_integration.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:19:22.446320] 2025-12-04T09:19:25.4598204Z 2025-12-04T09:19:25.4599034Z test_numba_integration 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_numba_integration_1.1_4717bde06fdb8313_.log 2025-12-04T09:19:25.4599746Z Running 0 items in this shard: 2025-12-04T09:19:25.4599926Z 2025-12-04T09:19:25.4600190Z Finished test_numba_integration 1/1 ... [2025-12-04 09:19:25.459608][1601.419154985], took 0.05min 2025-12-04T09:19:25.4728699Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_numba_integration/test_numba_integration-844fbe0ccaacabee.xml 2025-12-04T09:19:25.5002949Z Running test_functional_optim 1/1 ... [2025-12-04 09:19:25.500086][1601.45963445] 2025-12-04T09:19:25.5003381Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:19:25.5007081Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_functional_optim.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:19:25.500375] 2025-12-04T09:19:28.3597970Z 2025-12-04T09:19:28.3598746Z test_functional_optim 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_functional_optim_1.1_f4eff2a83f847c20_.log 2025-12-04T09:19:28.3599354Z Running 0 items in this shard: 2025-12-04T09:19:28.3599837Z 2025-12-04T09:19:28.3600201Z Finished test_functional_optim 1/1 ... [2025-12-04 09:19:28.359573][1604.319120169], took 0.05min 2025-12-04T09:19:28.3727213Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_functional_optim/test_functional_optim-182d833e6572f213.xml 2025-12-04T09:19:28.3981394Z Running test_maskedtensor 1/1 ... [2025-12-04 09:19:28.397941][1604.357489545] 2025-12-04T09:19:28.3981836Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:19:28.3984848Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_maskedtensor.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:19:28.398228] 2025-12-04T09:19:33.8987715Z 2025-12-04T09:19:33.8988440Z test_maskedtensor 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_maskedtensor_1.1_b12b8ab6e2307bad_.log 2025-12-04T09:19:33.8989024Z Running 0 items in this shard: 2025-12-04T09:19:33.8989187Z 2025-12-04T09:19:33.8989390Z Finished test_maskedtensor 1/1 ... [2025-12-04 09:19:33.898544][1609.858090246], took 0.09min 2025-12-04T09:19:33.9121261Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_maskedtensor/test_maskedtensor-2e832fb0545eaaab.xml 2025-12-04T09:19:33.9385779Z Running benchmark_utils/test_benchmark_utils 1/1 ... [2025-12-04 09:19:33.938328][1609.897876388] 2025-12-04T09:19:33.9386329Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:19:33.9389141Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'benchmark_utils/test_benchmark_utils.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:19:33.938618] 2025-12-04T09:19:36.6644690Z 2025-12-04T09:19:36.6645644Z benchmark_utils/test_benchmark_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/benchmark_utils.test_benchmark_utils_1.1_772184f487491c3a_.log 2025-12-04T09:19:36.6646483Z Running 0 items in this shard: 2025-12-04T09:19:36.6646669Z 2025-12-04T09:19:36.6646991Z Finished benchmark_utils/test_benchmark_utils 1/1 ... [2025-12-04 09:19:36.664244][1612.623790325], took 0.05min 2025-12-04T09:19:36.6782246Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/benchmark_utils.test_benchmark_utils/benchmark_utils.test_benchmark_utils-af96c8a6347bcdb1.xml 2025-12-04T09:19:36.7324345Z Running torch_np/numpy_tests/core/test_scalarmath 1/1 ... [2025-12-04 09:19:36.732193][1612.69174045] 2025-12-04T09:19:36.7324873Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:19:36.7327783Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_scalarmath.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:19:36.732484] 2025-12-04T09:19:39.7204857Z 2025-12-04T09:19:39.7205923Z torch_np/numpy_tests/core/test_scalarmath 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_scalarmath_1.1_6cf1169adf67ed5b_.log 2025-12-04T09:19:39.7206792Z Running 0 items in this shard: 2025-12-04T09:19:39.7206969Z 2025-12-04T09:19:39.7207309Z Finished torch_np/numpy_tests/core/test_scalarmath 1/1 ... [2025-12-04 09:19:39.720251][1615.679798008], took 0.05min 2025-12-04T09:19:39.7342883Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.core.test_scalarmath/torch_np.numpy_tests.core.test_scalarmath-b099d6ab8284837f.xml 2025-12-04T09:19:39.7656497Z Running test_scaled_matmul_cuda 1/1 ... [2025-12-04 09:19:39.765423][1615.72497075] 2025-12-04T09:19:39.7657293Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:19:39.7659883Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_scaled_matmul_cuda.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:19:39.765708] 2025-12-04T09:19:44.2848886Z 2025-12-04T09:19:44.2849853Z test_scaled_matmul_cuda 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_scaled_matmul_cuda_1.1_e3dc30c781dc2515_.log 2025-12-04T09:19:44.2850573Z Running 0 items in this shard: 2025-12-04T09:19:44.2850752Z 2025-12-04T09:19:44.2851025Z Finished test_scaled_matmul_cuda 1/1 ... [2025-12-04 09:19:44.284615][1620.24416107], took 0.08min 2025-12-04T09:19:44.2985588Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_scaled_matmul_cuda/test_scaled_matmul_cuda-5972d960cfe7ec4e.xml 2025-12-04T09:19:44.3259134Z Running torch_np/numpy_tests/core/test_shape_base 1/1 ... [2025-12-04 09:19:44.325675][1620.285222962] 2025-12-04T09:19:44.3259626Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:19:44.3262264Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_shape_base.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:19:44.325962] 2025-12-04T09:19:47.2449957Z 2025-12-04T09:19:47.2451074Z torch_np/numpy_tests/core/test_shape_base 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_shape_base_1.1_fdd19f4d65e5ee90_.log 2025-12-04T09:19:47.2451927Z Running 0 items in this shard: 2025-12-04T09:19:47.2452115Z 2025-12-04T09:19:47.2452443Z Finished torch_np/numpy_tests/core/test_shape_base 1/1 ... [2025-12-04 09:19:47.244766][1623.204313131], took 0.05min 2025-12-04T09:19:47.2588986Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.core.test_shape_base/torch_np.numpy_tests.core.test_shape_base-ea32bc17692c1ce8.xml 2025-12-04T09:19:47.2876212Z Running test_vulkan 1/1 ... [2025-12-04 09:19:47.287396][1623.246944623] 2025-12-04T09:19:47.2876620Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:19:47.2879601Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_vulkan.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:19:47.287687] 2025-12-04T09:19:50.0036382Z 2025-12-04T09:19:50.0037136Z test_vulkan 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_vulkan_1.1_610e90772bff2bbd_.log 2025-12-04T09:19:50.0037825Z Running 0 items in this shard: 2025-12-04T09:19:50.0038016Z 2025-12-04T09:19:50.0038255Z Finished test_vulkan 1/1 ... [2025-12-04 09:19:50.003451][1625.962993894], took 0.05min 2025-12-04T09:19:50.0180282Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_vulkan/test_vulkan-8b9c4d8ede98ee3d.xml 2025-12-04T09:19:50.0398314Z Running lazy/test_generator 1/1 ... [2025-12-04 09:19:50.039623][1625.999169806] 2025-12-04T09:19:50.0398768Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:19:50.0401769Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'lazy/test_generator.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:19:50.039931] 2025-12-04T09:19:52.7561849Z 2025-12-04T09:19:52.7563189Z lazy/test_generator 1/1 was successful, full logs can be found in artifacts with path test/test-reports/lazy.test_generator_1.1_08428f43a23dddc2_.log 2025-12-04T09:19:52.7563890Z Running 0 items in this shard: 2025-12-04T09:19:52.7564063Z 2025-12-04T09:19:52.7564331Z Finished lazy/test_generator 1/1 ... [2025-12-04 09:19:52.755967][1628.715510247], took 0.05min 2025-12-04T09:19:52.7706045Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/lazy.test_generator/lazy.test_generator-3b115040434e772f.xml 2025-12-04T09:19:52.8008703Z Running torch_np/numpy_tests/linalg/test_linalg 1/1 ... [2025-12-04 09:19:52.800627][1628.760173748] 2025-12-04T09:19:52.8009204Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:19:52.8012110Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/linalg/test_linalg.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:19:52.800921] 2025-12-04T09:19:55.9470232Z 2025-12-04T09:19:55.9471201Z torch_np/numpy_tests/linalg/test_linalg 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.linalg.test_linalg_1.1_2ae6c059e7041367_.log 2025-12-04T09:19:55.9472038Z Running 0 items in this shard: 2025-12-04T09:19:55.9472208Z 2025-12-04T09:19:55.9472549Z Finished torch_np/numpy_tests/linalg/test_linalg 1/1 ... [2025-12-04 09:19:55.946826][1631.906369576], took 0.05min 2025-12-04T09:19:55.9616098Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.linalg.test_linalg/torch_np.numpy_tests.linalg.test_linalg-40abae860675e681.xml 2025-12-04T09:19:55.9883415Z Running torch_np/numpy_tests/core/test_dtype 1/1 ... [2025-12-04 09:19:55.988106][1631.947652644] 2025-12-04T09:19:55.9883915Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:19:55.9886826Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_dtype.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:19:55.988407] 2025-12-04T09:19:58.8302817Z 2025-12-04T09:19:58.8303794Z torch_np/numpy_tests/core/test_dtype 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_dtype_1.1_ce81285daa6e127c_.log 2025-12-04T09:19:58.8304931Z Running 0 items in this shard: 2025-12-04T09:19:58.8305119Z 2025-12-04T09:19:58.8305443Z Finished torch_np/numpy_tests/core/test_dtype 1/1 ... [2025-12-04 09:19:58.830073][1634.789614552], took 0.05min 2025-12-04T09:19:58.8451716Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.core.test_dtype/torch_np.numpy_tests.core.test_dtype-fa2df09a508d13dc.xml 2025-12-04T09:19:58.8712690Z Running lazy/test_debug_util 1/1 ... [2025-12-04 09:19:58.871021][1634.830567971] 2025-12-04T09:19:58.8713138Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:19:58.8716100Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'lazy/test_debug_util.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:19:58.871322] 2025-12-04T09:20:01.5781243Z 2025-12-04T09:20:01.5782113Z lazy/test_debug_util 1/1 was successful, full logs can be found in artifacts with path test/test-reports/lazy.test_debug_util_1.1_ce59c472ce92b8ab_.log 2025-12-04T09:20:01.5782830Z Running 0 items in this shard: 2025-12-04T09:20:01.5783018Z 2025-12-04T09:20:01.5783277Z Finished lazy/test_debug_util 1/1 ... [2025-12-04 09:20:01.577875][1637.537417681], took 0.05min 2025-12-04T09:20:01.5930215Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/lazy.test_debug_util/lazy.test_debug_util-4dae4a778d170bc1.xml 2025-12-04T09:20:01.6216298Z Running nn/test_load_state_dict 1/1 ... [2025-12-04 09:20:01.621412][1637.580958794] 2025-12-04T09:20:01.6216741Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:20:01.6219695Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/test_load_state_dict.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:20:01.621724] 2025-12-04T09:20:04.6636723Z 2025-12-04T09:20:04.6637546Z nn/test_load_state_dict 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.test_load_state_dict_1.1_74b8c756583770f2_.log 2025-12-04T09:20:04.6638259Z Running 0 items in this shard: 2025-12-04T09:20:04.6638428Z 2025-12-04T09:20:04.6638717Z Finished nn/test_load_state_dict 1/1 ... [2025-12-04 09:20:04.663462][1640.623005155], took 0.05min 2025-12-04T09:20:04.6784245Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.test_load_state_dict/nn.test_load_state_dict-91536496375f24b6.xml 2025-12-04T09:20:04.6995669Z Running test_shape_ops 1/1 ... [2025-12-04 09:20:04.699352][1640.658898742] 2025-12-04T09:20:04.6996084Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:20:04.6999005Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_shape_ops.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:20:04.699654] 2025-12-04T09:20:07.9906547Z 2025-12-04T09:20:07.9907322Z test_shape_ops 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_shape_ops_1.1_55833b361d0f983c_.log 2025-12-04T09:20:07.9907992Z Running 0 items in this shard: 2025-12-04T09:20:07.9908175Z 2025-12-04T09:20:07.9908418Z Finished test_shape_ops 1/1 ... [2025-12-04 09:20:07.990392][1643.949934998], took 0.05min 2025-12-04T09:20:08.0056523Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_shape_ops/test_shape_ops-012d0fcc882a88fd.xml 2025-12-04T09:20:08.0325066Z Running nn/test_module_hooks 1/1 ... [2025-12-04 09:20:08.032282][1643.991828129] 2025-12-04T09:20:08.0325495Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:20:08.0328300Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/test_module_hooks.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:20:08.032571] 2025-12-04T09:20:11.1321316Z 2025-12-04T09:20:11.1322144Z nn/test_module_hooks 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.test_module_hooks_1.1_69958f2e5e5f5e6f_.log 2025-12-04T09:20:11.1322905Z Running 0 items in this shard: 2025-12-04T09:20:11.1323085Z 2025-12-04T09:20:11.1323340Z Finished nn/test_module_hooks 1/1 ... [2025-12-04 09:20:11.131721][1647.09126633], took 0.05min 2025-12-04T09:20:11.1470081Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.test_module_hooks/nn.test_module_hooks-46ca7ec97ec9318c.xml 2025-12-04T09:20:11.1744605Z Running torch_np/numpy_tests/lib/test_twodim_base 1/1 ... [2025-12-04 09:20:11.174227][1647.133775176] 2025-12-04T09:20:11.1745115Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:20:11.1747951Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/lib/test_twodim_base.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:20:11.174531] 2025-12-04T09:20:13.9520872Z 2025-12-04T09:20:13.9522176Z torch_np/numpy_tests/lib/test_twodim_base 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.lib.test_twodim_base_1.1_af867f8373093dbc_.log 2025-12-04T09:20:13.9523070Z Running 0 items in this shard: 2025-12-04T09:20:13.9523248Z 2025-12-04T09:20:13.9523600Z Finished torch_np/numpy_tests/lib/test_twodim_base 1/1 ... [2025-12-04 09:20:13.951891][1649.91143716], took 0.05min 2025-12-04T09:20:13.9671763Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.lib.test_twodim_base/torch_np.numpy_tests.lib.test_twodim_base-2accadefabd28fef.xml 2025-12-04T09:20:13.9939308Z Running profiler/test_memory_profiler 1/1 ... [2025-12-04 09:20:13.993710][1649.95325862] 2025-12-04T09:20:13.9939794Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:20:13.9942585Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'profiler/test_memory_profiler.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:20:13.994003] 2025-12-04T09:20:17.2302841Z 2025-12-04T09:20:17.2303771Z profiler/test_memory_profiler 1/1 was successful, full logs can be found in artifacts with path test/test-reports/profiler.test_memory_profiler_1.1_ea25b5650767ab1b_.log 2025-12-04T09:20:17.2305160Z Running 0 items in this shard: 2025-12-04T09:20:17.2305366Z 2025-12-04T09:20:17.2305674Z Finished profiler/test_memory_profiler 1/1 ... [2025-12-04 09:20:17.230035][1653.189581784], took 0.05min 2025-12-04T09:20:17.2458467Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/profiler.test_memory_profiler/profiler.test_memory_profiler-d6dc2fc9f3dab1f2.xml 2025-12-04T09:20:17.2776380Z Running test_jit_llga_fuser 1/1 ... [2025-12-04 09:20:17.277409][1653.236957937] 2025-12-04T09:20:17.2776832Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:20:17.2779687Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_jit_llga_fuser.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:20:17.277701] 2025-12-04T09:20:20.5684979Z 2025-12-04T09:20:20.5685772Z test_jit_llga_fuser 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_jit_llga_fuser_1.1_7b4085049eb12443_.log 2025-12-04T09:20:20.5686449Z Running 0 items in this shard: 2025-12-04T09:20:20.5686629Z 2025-12-04T09:20:20.5686876Z Finished test_jit_llga_fuser 1/1 ... [2025-12-04 09:20:20.568322][1656.527868975], took 0.05min 2025-12-04T09:20:20.5841789Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_jit_llga_fuser/test_jit_llga_fuser-644d80dd48236bd4.xml 2025-12-04T09:20:20.6080304Z Running test_serialization 1/1 ... [2025-12-04 09:20:20.607811][1656.567359235] 2025-12-04T09:20:20.6080730Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:20:20.6083520Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_serialization.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:20:20.608102] 2025-12-04T09:20:24.2027199Z 2025-12-04T09:20:24.2028051Z test_serialization 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_serialization_1.1_9875b20eb7f175eb_.log 2025-12-04T09:20:24.2028758Z Running 0 items in this shard: 2025-12-04T09:20:24.2028938Z 2025-12-04T09:20:24.2029217Z Finished test_serialization 1/1 ... [2025-12-04 09:20:24.202465][1660.16201175], took 0.06min 2025-12-04T09:20:24.2185212Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_serialization/test_serialization-9d6f2802fe66a8b4.xml 2025-12-04T09:20:24.2452261Z Running test_sparse_csr 1/3 ... [2025-12-04 09:20:24.244998][1660.204547024] 2025-12-04T09:20:24.2452676Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:20:24.2455650Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_sparse_csr.py', '-m', 'serial', '--shard-id=1', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:20:24.245294] 2025-12-04T09:20:35.4098821Z 2025-12-04T09:20:35.4099593Z test_sparse_csr 1/3 was successful, full logs can be found in artifacts with path test/test-reports/test_sparse_csr_1.3_feec296a088cb196_.log 2025-12-04T09:20:35.4100247Z Running 0 items in this shard: 2025-12-04T09:20:35.4100415Z 2025-12-04T09:20:35.4100682Z Finished test_sparse_csr 1/3 ... [2025-12-04 09:20:35.409612][1671.369157251], took 0.19min 2025-12-04T09:20:35.4257623Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_sparse_csr/test_sparse_csr-1e945673d9d0e64f.xml 2025-12-04T09:20:35.5015101Z Running functorch/test_ops 6/9 ... [2025-12-04 09:20:35.501247][1671.460794825] 2025-12-04T09:20:35.5015536Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:20:35.5018305Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_ops.py', '-m', 'serial', '--shard-id=6', '--num-shards=9', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:20:35.501564] 2025-12-04T09:20:53.5322929Z 2025-12-04T09:20:53.5323935Z functorch/test_ops 6/9 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_ops_6.9_a93c0a53b4ef93ed_.log 2025-12-04T09:20:53.5324662Z Running 0 items in this shard: 2025-12-04T09:20:53.5324850Z 2025-12-04T09:20:53.5325135Z Finished functorch/test_ops 6/9 ... [2025-12-04 09:20:53.532001][1689.491547505], took 0.30min 2025-12-04T09:20:53.5482176Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-258fb62f6ed3ad5d.xml 2025-12-04T09:20:53.6142978Z Running functorch/test_ops 8/9 ... [2025-12-04 09:20:53.614047][1689.573595237] 2025-12-04T09:20:53.6143418Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:20:53.6145905Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_ops.py', '-m', 'serial', '--shard-id=8', '--num-shards=9', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:20:53.614340] 2025-12-04T09:21:11.8881452Z 2025-12-04T09:21:11.8882360Z functorch/test_ops 8/9 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_ops_8.9_8bf2c8a6adbd0bd1_.log 2025-12-04T09:21:11.8883221Z Running 0 items in this shard: 2025-12-04T09:21:11.8883399Z 2025-12-04T09:21:11.8883657Z Finished functorch/test_ops 8/9 ... [2025-12-04 09:21:11.887863][1707.847409589], took 0.30min 2025-12-04T09:21:11.9044521Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-a4489a5f35bbfe70.xml 2025-12-04T09:21:11.9698061Z Running test_mkldnn 1/1 ... [2025-12-04 09:21:11.969476][1707.929019048] 2025-12-04T09:21:11.9698480Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:21:11.9701222Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_mkldnn.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:21:11.969866] 2025-12-04T09:21:15.1119925Z 2025-12-04T09:21:15.1121139Z test_mkldnn 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_mkldnn_1.1_e0499fc9c340a32c_.log 2025-12-04T09:21:15.1121810Z Running 0 items in this shard: 2025-12-04T09:21:15.1121995Z 2025-12-04T09:21:15.1122218Z Finished test_mkldnn 1/1 ... [2025-12-04 09:21:15.111775][1711.071321719], took 0.05min 2025-12-04T09:21:15.1283543Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_mkldnn/test_mkldnn-96ab41951b468da7.xml 2025-12-04T09:21:18.0789403Z Running inductor/test_aot_inductor 3/4 ... [2025-12-04 09:21:18.078389][1714.037933708] 2025-12-04T09:21:18.0789907Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:21:18.0791318Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_aot_inductor.py', '-m', 'not serial', '--shard-id=3', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:21:18.078804] 2025-12-04T09:21:18.0830277Z Running dynamo/test_model_output 1/1 ... [2025-12-04 09:21:18.082744][1714.042288694] 2025-12-04T09:21:18.0830731Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:21:18.0834021Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_model_output.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:21:18.083130] 2025-12-04T09:21:18.0882529Z Running inductor/test_extension_backend 1/1 ... [2025-12-04 09:21:18.087981][1714.047525282] 2025-12-04T09:21:18.0883003Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:21:18.0886476Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_extension_backend.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:21:18.088367] 2025-12-04T09:21:21.9168663Z 2025-12-04T09:21:21.9169477Z dynamo/test_model_output 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_model_output_1.1_4f24511bc5d419f7_.log 2025-12-04T09:21:21.9182253Z Running 50 items in this shard: test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr 2025-12-04T09:21:21.9194097Z 2025-12-04T09:21:21.9194323Z Finished dynamo/test_model_output 1/1 ... [2025-12-04 09:21:21.916890][1717.876431944], took 0.06min 2025-12-04T09:21:21.9230717Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_model_output/dynamo.test_model_output-0d371bc3cd364085.xml 2025-12-04T09:21:25.4650905Z Running inductor/test_cooperative_reductions 1/1 ... [2025-12-04 09:21:25.464642][1721.424186629] 2025-12-04T09:21:25.4651660Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:21:25.4653599Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_cooperative_reductions.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:21:25.465029] 2025-12-04T09:21:25.7382148Z 2025-12-04T09:21:25.7383201Z inductor/test_aot_inductor 3/4 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_aot_inductor_3.4_67cbdcc1e4ecfe70_.log 2025-12-04T09:21:25.7383990Z Running 0 items in this shard: 2025-12-04T09:21:25.7384161Z 2025-12-04T09:21:25.7384440Z Finished inductor/test_aot_inductor 3/4 ... [2025-12-04 09:21:25.738107][1721.69765227], took 0.13min 2025-12-04T09:21:25.7429110Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-863ed9428b922f00.xml 2025-12-04T09:21:29.1832412Z 2025-12-04T09:21:29.1834029Z PRINTING LOG FILE of inductor/test_extension_backend 1/1 (test/test-reports/inductor.test_extension_backend_1.1_c635b12d7c8d4f3e_.log) 2025-12-04T09:21:29.1835743Z W1204 09:21:24.788000 7756 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T09:21:29.1837312Z Test results will be stored in test-reports/python-pytest/inductor.test_extension_backend/inductor.test_extension_backend-184a28db46f4749e.xml 2025-12-04T09:21:29.1838285Z ============================= test session starts ============================== 2025-12-04T09:21:29.1839015Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:21:29.1839648Z cachedir: .pytest_cache 2025-12-04T09:21:29.1840438Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:21:29.1841319Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:21:29.1841723Z configfile: pytest.ini 2025-12-04T09:21:29.1842441Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:21:29.1843253Z collecting ... collected 1 item 2025-12-04T09:21:29.1843528Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:21:29.1857868Z Running 50 items in this shard: test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration, test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration 2025-12-04T09:21:29.1871970Z 2025-12-04T09:21:29.1873595Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration [1/2] c++ -MMD -MF extension_device.o.d -DTORCH_EXTENSION_NAME=extension_device -DTORCH_API_INCLUDE_EXTENSION_H -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -isystem /opt/conda/envs/py_3.10/include/python3.10 -fPIC -std=c++17 -g -c /var/lib/jenkins/workspace/test/inductor/extension_backends/cpp/extension_device.cpp -o extension_device.o 2025-12-04T09:21:29.1875668Z [2/2] c++ extension_device.o -shared -L/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o extension_device.so 2025-12-04T09:21:29.1876225Z FAILED [0.0514s] [ 2%] 2025-12-04T09:21:29.1876634Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0026s] [ 2%] 2025-12-04T09:21:29.1877267Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0025s] [ 2%] 2025-12-04T09:21:29.1877872Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1878478Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1879084Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0023s] [ 2%] 2025-12-04T09:21:29.1879685Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0026s] [ 2%] 2025-12-04T09:21:29.1880297Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1880997Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0023s] [ 2%] 2025-12-04T09:21:29.1882542Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1883161Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1883771Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0024s] [ 2%] 2025-12-04T09:21:29.1884374Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1884990Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1885595Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1886201Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1886821Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0024s] [ 2%] 2025-12-04T09:21:29.1887433Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1888033Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1888642Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1889247Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1889855Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0023s] [ 2%] 2025-12-04T09:21:29.1890469Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1891088Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1891695Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1892303Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1892903Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0023s] [ 2%] 2025-12-04T09:21:29.1893905Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1894667Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1895432Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1896073Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1896691Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0024s] [ 2%] 2025-12-04T09:21:29.1897313Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1897922Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0025s] [ 2%] 2025-12-04T09:21:29.1898530Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1899144Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1899753Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0024s] [ 2%] 2025-12-04T09:21:29.1900490Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1901173Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1901783Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1902392Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1902996Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0023s] [ 2%] 2025-12-04T09:21:29.1903597Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1904202Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1905092Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1905705Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1906303Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0023s] [ 2%] 2025-12-04T09:21:29.1906911Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1907517Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0021s] [ 2%] 2025-12-04T09:21:29.1908120Z inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration FAILED [0.0022s] [ 2%] 2025-12-04T09:21:29.1908458Z 2025-12-04T09:21:29.1908567Z =================================== FAILURES =================================== 2025-12-04T09:21:29.1908913Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.1909249Z Traceback (most recent call last): 2025-12-04T09:21:29.1909710Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 161, in test_open_device_registration 2025-12-04T09:21:29.1910194Z _, code = run_and_get_cpp_code(opt_fn, x, y, z) 2025-12-04T09:21:29.1910676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 3199, in run_and_get_cpp_code 2025-12-04T09:21:29.1911125Z result = fn(*args, **kwargs) 2025-12-04T09:21:29.1911535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 926, in compile_wrapper 2025-12-04T09:21:29.1911962Z return fn(*args, **kwargs) 2025-12-04T09:21:29.1912375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/convert_frame.py", line 2194, in __call__ 2025-12-04T09:21:29.1912810Z result = self._torchdynamo_orig_backend( 2025-12-04T09:21:29.1913241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/convert_frame.py", line 1937, in __call__ 2025-12-04T09:21:29.1913671Z result = self._inner_convert( 2025-12-04T09:21:29.1914077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/convert_frame.py", line 706, in __call__ 2025-12-04T09:21:29.1914482Z result = _compile( 2025-12-04T09:21:29.1914858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/convert_frame.py", line 1807, in _compile 2025-12-04T09:21:29.1915303Z raise InternalTorchDynamoError( 2025-12-04T09:21:29.1915718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/convert_frame.py", line 1744, in _compile 2025-12-04T09:21:29.1916224Z guarded_code, tracer_output = compile_inner(code, one_graph, hooks) 2025-12-04T09:21:29.1916728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_utils_internal.py", line 97, in wrapper_function 2025-12-04T09:21:29.1917156Z return function(*args, **kwargs) 2025-12-04T09:21:29.1917910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/convert_frame.py", line 1425, in compile_inner 2025-12-04T09:21:29.1918383Z return _compile_inner(code, one_graph, hooks) 2025-12-04T09:21:29.1918851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/convert_frame.py", line 1459, in _compile_inner 2025-12-04T09:21:29.1919307Z dynamo_output = compile_frame( 2025-12-04T09:21:29.1919731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/convert_frame.py", line 1341, in compile_frame 2025-12-04T09:21:29.1920238Z bytecode, tracer_output = transform_code_object(code, transform) 2025-12-04T09:21:29.1920817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/bytecode_transformation.py", line 1600, in transform_code_object 2025-12-04T09:21:29.1921377Z tracer_output = transformations(instructions, code_options) 2025-12-04T09:21:29.1921860Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/convert_frame.py", line 1313, in transform 2025-12-04T09:21:29.1922305Z tracer_output = trace_frame( 2025-12-04T09:21:29.1922695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/convert_frame.py", line 328, in _fn 2025-12-04T09:21:29.1923095Z return fn(*args, **kwargs) 2025-12-04T09:21:29.1923499Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/convert_frame.py", line 794, in trace_frame 2025-12-04T09:21:29.1923933Z tracer = InstructionTranslator( 2025-12-04T09:21:29.1924359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 4586, in __init__ 2025-12-04T09:21:29.1924821Z self.symbolic_stream_state = SymbolicStreamState() 2025-12-04T09:21:29.1925287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/variables/streams.py", line 215, in __init__ 2025-12-04T09:21:29.1925724Z if torch.accelerator.is_available(): 2025-12-04T09:21:29.1926152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/accelerator/__init__.py", line 95, in is_available 2025-12-04T09:21:29.1926578Z return mod.is_available() 2025-12-04T09:21:29.1927006Z torch._dynamo.exc.InternalTorchDynamoError: AttributeError: module 'extension_device' has no attribute 'is_available' 2025-12-04T09:21:29.1927382Z 2025-12-04T09:21:29.1927811Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:21:29.1928297Z 2025-12-04T09:21:29.1928300Z 2025-12-04T09:21:29.1928441Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.1928904Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.1929239Z 2025-12-04T09:21:29.1929403Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.1929788Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.1930062Z frames [('total', 1)] 2025-12-04T09:21:29.1930344Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.1930668Z Traceback (most recent call last): 2025-12-04T09:21:29.1931121Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.1931636Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.1932130Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.1932551Z raise RuntimeError( 2025-12-04T09:21:29.1933307Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.1934036Z 2025-12-04T09:21:29.1934170Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.1934718Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.1935050Z 2025-12-04T09:21:29.1935219Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.1935586Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.1935863Z frames [('total', 1)] 2025-12-04T09:21:29.1936102Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.1937027Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.1937876Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.1938174Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.1938508Z Traceback (most recent call last): 2025-12-04T09:21:29.1938963Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.1939479Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.1939975Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.1940397Z raise RuntimeError( 2025-12-04T09:21:29.1941057Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.1941689Z 2025-12-04T09:21:29.1941821Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.1942284Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.1942621Z 2025-12-04T09:21:29.1942785Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.1943357Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.1943642Z frames [('total', 1)] 2025-12-04T09:21:29.1943889Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.1944701Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.1945460Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.1945758Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.1946083Z Traceback (most recent call last): 2025-12-04T09:21:29.1946534Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.1947068Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.1947568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.1947993Z raise RuntimeError( 2025-12-04T09:21:29.1948654Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.1949268Z 2025-12-04T09:21:29.1949397Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.1949853Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.1950185Z 2025-12-04T09:21:29.1950347Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.1950828Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.1951164Z frames [('total', 1)] 2025-12-04T09:21:29.1951399Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.1952194Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.1952934Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.1953224Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.1953544Z Traceback (most recent call last): 2025-12-04T09:21:29.1953988Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.1954499Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.1955003Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.1955430Z raise RuntimeError( 2025-12-04T09:21:29.1956081Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.1956694Z 2025-12-04T09:21:29.1956822Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.1957275Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.1957605Z 2025-12-04T09:21:29.1957777Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.1958149Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.1958415Z frames [('total', 1)] 2025-12-04T09:21:29.1958646Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.1959456Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.1960194Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.1960489Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.1960811Z Traceback (most recent call last): 2025-12-04T09:21:29.1961250Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.1961782Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.1962298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.1962723Z raise RuntimeError( 2025-12-04T09:21:29.1963398Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.1964020Z 2025-12-04T09:21:29.1964155Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.1964640Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.1964969Z 2025-12-04T09:21:29.1965133Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.1965510Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.1965792Z frames [('total', 1)] 2025-12-04T09:21:29.1966033Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.1966833Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.1967757Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.1968066Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.1968392Z Traceback (most recent call last): 2025-12-04T09:21:29.1968856Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.1969379Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.1969867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.1970282Z raise RuntimeError( 2025-12-04T09:21:29.1970935Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.1971560Z 2025-12-04T09:21:29.1971695Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.1972148Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.1972473Z 2025-12-04T09:21:29.1972634Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.1973002Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.1973393Z frames [('total', 1)] 2025-12-04T09:21:29.1973630Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.1974421Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.1975165Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.1975469Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.1975790Z Traceback (most recent call last): 2025-12-04T09:21:29.1976224Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.1976744Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.1977232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.1977643Z raise RuntimeError( 2025-12-04T09:21:29.1978299Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.1978921Z 2025-12-04T09:21:29.1979052Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.1979506Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.1979831Z 2025-12-04T09:21:29.1979994Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.1980362Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.1980638Z frames [('total', 1)] 2025-12-04T09:21:29.1980869Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.1981664Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.1982408Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.1982709Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.1983030Z Traceback (most recent call last): 2025-12-04T09:21:29.1983558Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.1984141Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.1984631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.1985042Z raise RuntimeError( 2025-12-04T09:21:29.1985699Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.1986318Z 2025-12-04T09:21:29.1986445Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.1986899Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.1987221Z 2025-12-04T09:21:29.1987380Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.1987777Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.1988060Z frames [('total', 1)] 2025-12-04T09:21:29.1988291Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.1989087Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.1989830Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.1990132Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.1990457Z Traceback (most recent call last): 2025-12-04T09:21:29.1990898Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.1991411Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.1991906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.1992319Z raise RuntimeError( 2025-12-04T09:21:29.1992972Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.1993586Z 2025-12-04T09:21:29.1993725Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.1994179Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.1994497Z 2025-12-04T09:21:29.1994655Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.1995020Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.1995299Z frames [('total', 1)] 2025-12-04T09:21:29.1995526Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.1996316Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.1997051Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.1997343Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.1997661Z Traceback (most recent call last): 2025-12-04T09:21:29.1998097Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.1998607Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.1999094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.1999593Z raise RuntimeError( 2025-12-04T09:21:29.2000313Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2000939Z 2025-12-04T09:21:29.2001071Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2001520Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2001838Z 2025-12-04T09:21:29.2002002Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2002362Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2002638Z frames [('total', 1)] 2025-12-04T09:21:29.2002872Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2003666Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2004410Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2005070Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2005399Z Traceback (most recent call last): 2025-12-04T09:21:29.2005843Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2006365Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2006857Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2007284Z raise RuntimeError( 2025-12-04T09:21:29.2007941Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2008574Z 2025-12-04T09:21:29.2008710Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2009172Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2009497Z 2025-12-04T09:21:29.2009663Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2010033Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2010310Z frames [('total', 1)] 2025-12-04T09:21:29.2010543Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2011340Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2012086Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2012389Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2012718Z Traceback (most recent call last): 2025-12-04T09:21:29.2013154Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2013794Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2014287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2014704Z raise RuntimeError( 2025-12-04T09:21:29.2015353Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2015967Z 2025-12-04T09:21:29.2016092Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2016877Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2017216Z 2025-12-04T09:21:29.2017388Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2017762Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2018037Z frames [('total', 1)] 2025-12-04T09:21:29.2018269Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2019071Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2019815Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2020115Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2020441Z Traceback (most recent call last): 2025-12-04T09:21:29.2020883Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2021413Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2021904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2022327Z raise RuntimeError( 2025-12-04T09:21:29.2022977Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2023594Z 2025-12-04T09:21:29.2023721Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2024174Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2024502Z 2025-12-04T09:21:29.2024669Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2025037Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2025315Z frames [('total', 1)] 2025-12-04T09:21:29.2025548Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2026339Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2027091Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2027390Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2027714Z Traceback (most recent call last): 2025-12-04T09:21:29.2028149Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2028670Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2029170Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2029592Z raise RuntimeError( 2025-12-04T09:21:29.2030241Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2030862Z 2025-12-04T09:21:29.2030991Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2031450Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2031783Z 2025-12-04T09:21:29.2031950Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2032310Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2032717Z frames [('total', 1)] 2025-12-04T09:21:29.2033046Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2033839Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2034576Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2034874Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2035193Z Traceback (most recent call last): 2025-12-04T09:21:29.2035639Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2036163Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2036649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2037080Z raise RuntimeError( 2025-12-04T09:21:29.2037728Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2038344Z 2025-12-04T09:21:29.2038471Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2038918Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2039242Z 2025-12-04T09:21:29.2039407Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2039765Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2040037Z frames [('total', 1)] 2025-12-04T09:21:29.2040264Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2041058Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2041793Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2042087Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2042405Z Traceback (most recent call last): 2025-12-04T09:21:29.2042837Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2043362Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2043915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2044576Z raise RuntimeError( 2025-12-04T09:21:29.2045353Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2045976Z 2025-12-04T09:21:29.2046105Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2046598Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2046945Z 2025-12-04T09:21:29.2047111Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2047470Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2047748Z frames [('total', 1)] 2025-12-04T09:21:29.2047977Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2048765Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2049619Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2049998Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2050325Z Traceback (most recent call last): 2025-12-04T09:21:29.2050758Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2051273Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2051762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2052181Z raise RuntimeError( 2025-12-04T09:21:29.2052892Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2053701Z 2025-12-04T09:21:29.2053836Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2054298Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2054625Z 2025-12-04T09:21:29.2054791Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2055160Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2055452Z frames [('total', 1)] 2025-12-04T09:21:29.2067527Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2068465Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2069234Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2069558Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2069920Z Traceback (most recent call last): 2025-12-04T09:21:29.2070403Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2070929Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2071446Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2071884Z raise RuntimeError( 2025-12-04T09:21:29.2072563Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2073184Z 2025-12-04T09:21:29.2073319Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2073785Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2074115Z 2025-12-04T09:21:29.2074289Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2074668Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2074952Z frames [('total', 1)] 2025-12-04T09:21:29.2075197Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2076011Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2076748Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2077065Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2077394Z Traceback (most recent call last): 2025-12-04T09:21:29.2077850Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2078591Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2079091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2079517Z raise RuntimeError( 2025-12-04T09:21:29.2080190Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2080817Z 2025-12-04T09:21:29.2080953Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2081425Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2081751Z 2025-12-04T09:21:29.2081923Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2082299Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2082584Z frames [('total', 1)] 2025-12-04T09:21:29.2082828Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2083632Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2084362Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2084664Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2084991Z Traceback (most recent call last): 2025-12-04T09:21:29.2085462Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2086001Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2086509Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2086944Z raise RuntimeError( 2025-12-04T09:21:29.2087611Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2088233Z 2025-12-04T09:21:29.2088365Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2088826Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2089150Z 2025-12-04T09:21:29.2089320Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2089696Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2089971Z frames [('total', 1)] 2025-12-04T09:21:29.2090209Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2091020Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2091755Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2092062Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2092390Z Traceback (most recent call last): 2025-12-04T09:21:29.2092844Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2093466Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2093963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2094391Z raise RuntimeError( 2025-12-04T09:21:29.2095240Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2095864Z 2025-12-04T09:21:29.2095998Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2096456Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2096782Z 2025-12-04T09:21:29.2096953Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2097324Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2097598Z frames [('total', 1)] 2025-12-04T09:21:29.2097835Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2098632Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2099377Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2099683Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2100009Z Traceback (most recent call last): 2025-12-04T09:21:29.2100456Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2100984Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2101478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2101900Z raise RuntimeError( 2025-12-04T09:21:29.2102551Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2103170Z 2025-12-04T09:21:29.2103304Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2103760Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2104085Z 2025-12-04T09:21:29.2104252Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2104917Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2105194Z frames [('total', 1)] 2025-12-04T09:21:29.2105443Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2106254Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2106991Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2107300Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2107629Z Traceback (most recent call last): 2025-12-04T09:21:29.2108072Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2108590Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2109079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2109503Z raise RuntimeError( 2025-12-04T09:21:29.2110154Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2110773Z 2025-12-04T09:21:29.2110901Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2111566Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2111908Z 2025-12-04T09:21:29.2112179Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2112550Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2112830Z frames [('total', 1)] 2025-12-04T09:21:29.2113072Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2113874Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2114609Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2114922Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2115244Z Traceback (most recent call last): 2025-12-04T09:21:29.2115700Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2116219Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2116709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2117131Z raise RuntimeError( 2025-12-04T09:21:29.2117772Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2118394Z 2025-12-04T09:21:29.2118521Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2118986Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2119310Z 2025-12-04T09:21:29.2119478Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2119854Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2120128Z frames [('total', 1)] 2025-12-04T09:21:29.2120364Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2121157Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2121887Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2122187Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2122509Z Traceback (most recent call last): 2025-12-04T09:21:29.2122953Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2123471Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2123972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2124394Z raise RuntimeError( 2025-12-04T09:21:29.2125042Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2125659Z 2025-12-04T09:21:29.2125790Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2126244Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2126569Z 2025-12-04T09:21:29.2126736Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2127112Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2127477Z frames [('total', 1)] 2025-12-04T09:21:29.2127709Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2128590Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2129325Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2129627Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2129953Z Traceback (most recent call last): 2025-12-04T09:21:29.2130403Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2130915Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2131402Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2131830Z raise RuntimeError( 2025-12-04T09:21:29.2132483Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2133106Z 2025-12-04T09:21:29.2133310Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2133773Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2134108Z 2025-12-04T09:21:29.2134275Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2134643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2134917Z frames [('total', 1)] 2025-12-04T09:21:29.2135151Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2135953Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2136691Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2136999Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2137334Z Traceback (most recent call last): 2025-12-04T09:21:29.2137787Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2138300Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2138788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2139205Z raise RuntimeError( 2025-12-04T09:21:29.2139861Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2140485Z 2025-12-04T09:21:29.2140617Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2141065Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2141387Z 2025-12-04T09:21:29.2141552Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2141920Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2142186Z frames [('total', 1)] 2025-12-04T09:21:29.2142415Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2143203Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2144031Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2144425Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2144751Z Traceback (most recent call last): 2025-12-04T09:21:29.2145191Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2145702Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2146186Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2146603Z raise RuntimeError( 2025-12-04T09:21:29.2147263Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2147875Z 2025-12-04T09:21:29.2148004Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2148465Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2148792Z 2025-12-04T09:21:29.2148956Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2149324Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2149592Z frames [('total', 1)] 2025-12-04T09:21:29.2149822Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2150610Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2151353Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2151654Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2151982Z Traceback (most recent call last): 2025-12-04T09:21:29.2152428Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2152939Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2153430Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2153850Z raise RuntimeError( 2025-12-04T09:21:29.2154502Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2155113Z 2025-12-04T09:21:29.2155242Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2155694Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2156019Z 2025-12-04T09:21:29.2156187Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2156558Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2156826Z frames [('total', 1)] 2025-12-04T09:21:29.2157069Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2157872Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2158607Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2158902Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2159224Z Traceback (most recent call last): 2025-12-04T09:21:29.2159670Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2160262Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2160810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2161238Z raise RuntimeError( 2025-12-04T09:21:29.2161898Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2162508Z 2025-12-04T09:21:29.2162636Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2163103Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2163423Z 2025-12-04T09:21:29.2163590Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2163954Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2164230Z frames [('total', 1)] 2025-12-04T09:21:29.2164464Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2165255Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2165988Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2166280Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2166605Z Traceback (most recent call last): 2025-12-04T09:21:29.2167048Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2167552Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2168035Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2168648Z raise RuntimeError( 2025-12-04T09:21:29.2169689Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2170386Z 2025-12-04T09:21:29.2170522Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2170983Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2171317Z 2025-12-04T09:21:29.2171482Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2171860Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2172179Z frames [('total', 1)] 2025-12-04T09:21:29.2172592Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2173680Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2174449Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2174752Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2175081Z Traceback (most recent call last): 2025-12-04T09:21:29.2175534Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2176050Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2176545Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2176968Z raise RuntimeError( 2025-12-04T09:21:29.2177730Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2178417Z 2025-12-04T09:21:29.2178548Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2179004Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2179332Z 2025-12-04T09:21:29.2179493Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2179861Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2180136Z frames [('total', 1)] 2025-12-04T09:21:29.2180376Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2181173Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2181919Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2182229Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2182555Z Traceback (most recent call last): 2025-12-04T09:21:29.2183002Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2183527Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2184023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2184446Z raise RuntimeError( 2025-12-04T09:21:29.2185111Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2185735Z 2025-12-04T09:21:29.2185870Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2186340Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2186675Z 2025-12-04T09:21:29.2186844Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2187228Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2187503Z frames [('total', 1)] 2025-12-04T09:21:29.2187742Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2188546Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2189289Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2189588Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2189921Z Traceback (most recent call last): 2025-12-04T09:21:29.2190379Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2190902Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2191390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2191813Z raise RuntimeError( 2025-12-04T09:21:29.2192475Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2193089Z 2025-12-04T09:21:29.2193227Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2193680Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2194108Z 2025-12-04T09:21:29.2194338Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2194713Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2194995Z frames [('total', 1)] 2025-12-04T09:21:29.2195223Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2196026Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2196766Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2197084Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2197413Z Traceback (most recent call last): 2025-12-04T09:21:29.2197862Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2198398Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2198887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2199309Z raise RuntimeError( 2025-12-04T09:21:29.2199970Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2200582Z 2025-12-04T09:21:29.2200715Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2201160Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2201488Z 2025-12-04T09:21:29.2201647Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2202019Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2202303Z frames [('total', 1)] 2025-12-04T09:21:29.2202529Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2203319Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2204056Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2204352Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2204920Z Traceback (most recent call last): 2025-12-04T09:21:29.2205361Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2205878Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2206364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2206787Z raise RuntimeError( 2025-12-04T09:21:29.2207463Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2208077Z 2025-12-04T09:21:29.2208210Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2208656Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2208987Z 2025-12-04T09:21:29.2209146Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2209519Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2209797Z frames [('total', 1)] 2025-12-04T09:21:29.2210034Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2211155Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2211906Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2212206Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2212535Z Traceback (most recent call last): 2025-12-04T09:21:29.2212990Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2213587Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2214078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2214500Z raise RuntimeError( 2025-12-04T09:21:29.2215170Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2215786Z 2025-12-04T09:21:29.2215923Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2216371Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2216701Z 2025-12-04T09:21:29.2216867Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2217242Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2217525Z frames [('total', 1)] 2025-12-04T09:21:29.2217755Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2218549Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2219313Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2219612Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2219940Z Traceback (most recent call last): 2025-12-04T09:21:29.2220407Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2220948Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2221449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2221881Z raise RuntimeError( 2025-12-04T09:21:29.2222551Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2223174Z 2025-12-04T09:21:29.2223316Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2223769Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2224103Z 2025-12-04T09:21:29.2224264Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2224634Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2224908Z frames [('total', 1)] 2025-12-04T09:21:29.2225134Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2225941Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2226677Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2227068Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2227450Z Traceback (most recent call last): 2025-12-04T09:21:29.2227918Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2228438Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2228922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2229344Z raise RuntimeError( 2025-12-04T09:21:29.2229994Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2230611Z 2025-12-04T09:21:29.2230747Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2231198Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2231528Z 2025-12-04T09:21:29.2231695Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2232065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2232339Z frames [('total', 1)] 2025-12-04T09:21:29.2232568Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2233364Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2234103Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2234404Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2234725Z Traceback (most recent call last): 2025-12-04T09:21:29.2235177Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2235702Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2236184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2236604Z raise RuntimeError( 2025-12-04T09:21:29.2237290Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2238249Z 2025-12-04T09:21:29.2238398Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2238853Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2239188Z 2025-12-04T09:21:29.2239359Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2239750Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2240035Z frames [('total', 1)] 2025-12-04T09:21:29.2240269Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2241074Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2241814Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2242122Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2242447Z Traceback (most recent call last): 2025-12-04T09:21:29.2242907Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2243444Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2244109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2244538Z raise RuntimeError( 2025-12-04T09:21:29.2245200Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2245815Z 2025-12-04T09:21:29.2245954Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2246432Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2246759Z 2025-12-04T09:21:29.2246921Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2247304Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2247595Z frames [('total', 1)] 2025-12-04T09:21:29.2247828Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2248636Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2249391Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2249711Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2250037Z Traceback (most recent call last): 2025-12-04T09:21:29.2250500Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2251029Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2251529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2251951Z raise RuntimeError( 2025-12-04T09:21:29.2252620Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2253357Z 2025-12-04T09:21:29.2253492Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2253950Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2254277Z 2025-12-04T09:21:29.2254437Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2254808Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2255086Z frames [('total', 1)] 2025-12-04T09:21:29.2255315Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2256120Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2256864Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2257169Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2257494Z Traceback (most recent call last): 2025-12-04T09:21:29.2257946Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2258468Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2258970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2259391Z raise RuntimeError( 2025-12-04T09:21:29.2260052Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2260761Z 2025-12-04T09:21:29.2260956Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2261419Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2261745Z 2025-12-04T09:21:29.2261907Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2262278Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2262559Z frames [('total', 1)] 2025-12-04T09:21:29.2262787Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2263587Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2264333Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2264644Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2264966Z Traceback (most recent call last): 2025-12-04T09:21:29.2265415Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2265938Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2266431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2266848Z raise RuntimeError( 2025-12-04T09:21:29.2267515Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2268134Z 2025-12-04T09:21:29.2268264Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2268729Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2269060Z 2025-12-04T09:21:29.2269222Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2269592Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2269871Z frames [('total', 1)] 2025-12-04T09:21:29.2270119Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2270914Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2271660Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2271966Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2272292Z Traceback (most recent call last): 2025-12-04T09:21:29.2272751Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2273275Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2273771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2274190Z raise RuntimeError( 2025-12-04T09:21:29.2274850Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2275468Z 2025-12-04T09:21:29.2275597Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2276052Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2276485Z 2025-12-04T09:21:29.2276647Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2277124Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2277409Z frames [('total', 1)] 2025-12-04T09:21:29.2277647Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2278440Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2279172Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2279471Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2279788Z Traceback (most recent call last): 2025-12-04T09:21:29.2280232Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2280760Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2281256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2281673Z raise RuntimeError( 2025-12-04T09:21:29.2282336Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2282957Z 2025-12-04T09:21:29.2283087Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2283556Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2283880Z 2025-12-04T09:21:29.2284042Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2284416Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2284698Z frames [('total', 1)] 2025-12-04T09:21:29.2284937Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2285732Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2286471Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2286775Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2287104Z Traceback (most recent call last): 2025-12-04T09:21:29.2287561Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2288080Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2288568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2288991Z raise RuntimeError( 2025-12-04T09:21:29.2289674Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2290301Z 2025-12-04T09:21:29.2290432Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2290888Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2291216Z 2025-12-04T09:21:29.2291375Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2291749Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2292030Z frames [('total', 1)] 2025-12-04T09:21:29.2292268Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2293313Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2294096Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2294402Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2294739Z Traceback (most recent call last): 2025-12-04T09:21:29.2295198Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2295722Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2296217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2296638Z raise RuntimeError( 2025-12-04T09:21:29.2297315Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2297948Z 2025-12-04T09:21:29.2298082Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2298542Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2298870Z 2025-12-04T09:21:29.2299029Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2299407Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2299687Z frames [('total', 1)] 2025-12-04T09:21:29.2299916Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2300712Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2301471Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2301780Z _____________ ExtensionBackendTests.test_open_device_registration ______________ 2025-12-04T09:21:29.2302105Z Traceback (most recent call last): 2025-12-04T09:21:29.2302549Z File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 121, in test_open_device_registration 2025-12-04T09:21:29.2303065Z torch._register_device_module("extension_device", self.module) 2025-12-04T09:21:29.2303554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 2773, in _register_device_module 2025-12-04T09:21:29.2303970Z raise RuntimeError( 2025-12-04T09:21:29.2304914Z RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2305548Z 2025-12-04T09:21:29.2305678Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2306132Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2306458Z 2025-12-04T09:21:29.2306618Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2306991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:21:29.2307268Z frames [('total', 1)] 2025-12-04T09:21:29.2307499Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:21:29.2308305Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/random.py:42: UserWarning: Set seed for `extension_device` device does not take effect, please add API's `_is_in_bad_fork` and `manual_seed_all` to `extension_device` device module. 2025-12-04T09:21:29.2309047Z return _manual_seed_impl(seed) 2025-12-04T09:21:29.2309802Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_extension_backend/inductor.test_extension_backend-184a28db46f4749e.xml - 2025-12-04T09:21:29.2310556Z =========================== short test summary info ============================ 2025-12-04T09:21:29.2311382Z FAILED [0.0514s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - torch._dynamo.exc.InternalTorchDynamoError: AttributeError: module 'extension_device' has no attribute 'is_available' 2025-12-04T09:21:29.2312068Z 2025-12-04T09:21:29.2312490Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:21:29.2312983Z 2025-12-04T09:21:29.2312986Z 2025-12-04T09:21:29.2313117Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2313581Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2313915Z 2025-12-04T09:21:29.2314092Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2315198Z FAILED [0.0026s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2316124Z 2025-12-04T09:21:29.2316256Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2316717Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2317047Z 2025-12-04T09:21:29.2317212Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2318308Z FAILED [0.0025s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2319227Z 2025-12-04T09:21:29.2319357Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2319809Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2320143Z 2025-12-04T09:21:29.2320304Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2321389Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2322300Z 2025-12-04T09:21:29.2322434Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2322889Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2323220Z 2025-12-04T09:21:29.2323382Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2324466Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2325381Z 2025-12-04T09:21:29.2325531Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2325976Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2326303Z 2025-12-04T09:21:29.2326459Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2327687Z FAILED [0.0023s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2328603Z 2025-12-04T09:21:29.2328735Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2329182Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2329504Z 2025-12-04T09:21:29.2329659Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2330740Z FAILED [0.0026s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2331660Z 2025-12-04T09:21:29.2331790Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2332235Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2332554Z 2025-12-04T09:21:29.2332708Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2333922Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2334836Z 2025-12-04T09:21:29.2334962Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2335404Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2335727Z 2025-12-04T09:21:29.2335891Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2336971Z FAILED [0.0023s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2337888Z 2025-12-04T09:21:29.2338012Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2338451Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2338769Z 2025-12-04T09:21:29.2338933Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2340015Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2340939Z 2025-12-04T09:21:29.2341064Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2341513Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2341839Z 2025-12-04T09:21:29.2341995Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2343069Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2344087Z 2025-12-04T09:21:29.2344219Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2344731Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2345074Z 2025-12-04T09:21:29.2345234Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2346311Z FAILED [0.0024s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2347223Z 2025-12-04T09:21:29.2347353Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2347793Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2348141Z 2025-12-04T09:21:29.2348299Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2349381Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2350302Z 2025-12-04T09:21:29.2350430Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2350874Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2351196Z 2025-12-04T09:21:29.2351354Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2352439Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2353364Z 2025-12-04T09:21:29.2353487Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2353933Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2354255Z 2025-12-04T09:21:29.2354435Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2355515Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2356439Z 2025-12-04T09:21:29.2356565Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2357019Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2357344Z 2025-12-04T09:21:29.2357508Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2358581Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2359499Z 2025-12-04T09:21:29.2359625Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2360074Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2360395Z 2025-12-04T09:21:29.2360559Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2361781Z FAILED [0.0024s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2362700Z 2025-12-04T09:21:29.2362825Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2363271Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2363596Z 2025-12-04T09:21:29.2363749Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2364840Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2365758Z 2025-12-04T09:21:29.2365893Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2366331Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2366658Z 2025-12-04T09:21:29.2366813Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2367889Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2368798Z 2025-12-04T09:21:29.2368927Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2369372Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2369700Z 2025-12-04T09:21:29.2369858Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2370928Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2371844Z 2025-12-04T09:21:29.2371970Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2372417Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2372738Z 2025-12-04T09:21:29.2372891Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2374080Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2375003Z 2025-12-04T09:21:29.2375130Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2375575Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2375895Z 2025-12-04T09:21:29.2376057Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2377147Z FAILED [0.0023s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2378068Z 2025-12-04T09:21:29.2378277Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2378786Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2379112Z 2025-12-04T09:21:29.2379275Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2380352Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2381265Z 2025-12-04T09:21:29.2381391Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2381838Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2382166Z 2025-12-04T09:21:29.2382323Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2383407Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2384322Z 2025-12-04T09:21:29.2384455Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2384897Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2385226Z 2025-12-04T09:21:29.2385382Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2386473Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2387393Z 2025-12-04T09:21:29.2387524Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2387968Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2388295Z 2025-12-04T09:21:29.2388454Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2389537Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2390459Z 2025-12-04T09:21:29.2390593Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2391049Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2391375Z 2025-12-04T09:21:29.2391538Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2392622Z FAILED [0.0023s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2393540Z 2025-12-04T09:21:29.2393665Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2394123Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2394454Z 2025-12-04T09:21:29.2394611Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2395771Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2396780Z 2025-12-04T09:21:29.2396908Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2397356Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2397686Z 2025-12-04T09:21:29.2397847Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2398920Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2399849Z 2025-12-04T09:21:29.2399982Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2400431Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2400750Z 2025-12-04T09:21:29.2400915Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2401992Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2402904Z 2025-12-04T09:21:29.2403027Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2403481Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2403811Z 2025-12-04T09:21:29.2403966Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2405390Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2406303Z 2025-12-04T09:21:29.2406432Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2406870Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2407207Z 2025-12-04T09:21:29.2407359Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2408436Z FAILED [0.0024s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2409359Z 2025-12-04T09:21:29.2409490Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2409928Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2410258Z 2025-12-04T09:21:29.2410413Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2411490Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2412408Z 2025-12-04T09:21:29.2412533Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2413319Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2413656Z 2025-12-04T09:21:29.2413813Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2414889Z FAILED [0.0025s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2415876Z 2025-12-04T09:21:29.2415999Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2416445Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2416765Z 2025-12-04T09:21:29.2416925Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2418011Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2418926Z 2025-12-04T09:21:29.2419050Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2419503Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2419828Z 2025-12-04T09:21:29.2419983Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2421061Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2421986Z 2025-12-04T09:21:29.2422124Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2422565Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2422890Z 2025-12-04T09:21:29.2423042Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2424123Z FAILED [0.0024s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2425041Z 2025-12-04T09:21:29.2425171Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2425612Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2425945Z 2025-12-04T09:21:29.2426107Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2427189Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2428105Z 2025-12-04T09:21:29.2428229Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2428678Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2429000Z 2025-12-04T09:21:29.2429154Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2430325Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2431314Z 2025-12-04T09:21:29.2431443Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2431894Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2432217Z 2025-12-04T09:21:29.2432381Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2433457Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2434385Z 2025-12-04T09:21:29.2434520Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2434981Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2435303Z 2025-12-04T09:21:29.2435464Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2436532Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2437449Z 2025-12-04T09:21:29.2437571Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2438018Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2438339Z 2025-12-04T09:21:29.2438505Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2439585Z FAILED [0.0023s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2440495Z 2025-12-04T09:21:29.2440623Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2441062Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2441386Z 2025-12-04T09:21:29.2441541Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2442621Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2443538Z 2025-12-04T09:21:29.2443667Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2444106Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2444434Z 2025-12-04T09:21:29.2444588Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2445669Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2446573Z 2025-12-04T09:21:29.2446704Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2447023Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2447092Z 2025-12-04T09:21:29.2447254Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2448098Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2448102Z 2025-12-04T09:21:29.2448229Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2448469Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2448472Z 2025-12-04T09:21:29.2448628Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2449483Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2449487Z 2025-12-04T09:21:29.2449611Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2449853Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2449856Z 2025-12-04T09:21:29.2450009Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2450857Z FAILED [0.0023s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2450864Z 2025-12-04T09:21:29.2450992Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2451232Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2451236Z 2025-12-04T09:21:29.2451398Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2452236Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2452240Z 2025-12-04T09:21:29.2452370Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2452611Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2452619Z 2025-12-04T09:21:29.2452782Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2453711Z FAILED [0.0021s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2453715Z 2025-12-04T09:21:29.2453840Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2454085Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2454088Z 2025-12-04T09:21:29.2454239Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2455155Z FAILED [0.0022s] inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration - RuntimeError: The runtime module of 'extension_device' has already been registered with '' 2025-12-04T09:21:29.2455243Z 2025-12-04T09:21:29.2455379Z To execute this test, run the following from the base repo dir: 2025-12-04T09:21:29.2455634Z python test/inductor/test_extension_backend.py ExtensionBackendTests.test_open_device_registration 2025-12-04T09:21:29.2455637Z 2025-12-04T09:21:29.2455792Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:21:29.2455895Z ============================== 50 failed in 3.03s ============================== 2025-12-04T09:21:29.2455900Z 2025-12-04T09:21:29.2456273Z FINISHED PRINTING LOG FILE of inductor/test_extension_backend 1/1 (test/test-reports/inductor.test_extension_backend_1.1_c635b12d7c8d4f3e_.log) 2025-12-04T09:21:29.2456277Z 2025-12-04T09:21:29.2456506Z Finished inductor/test_extension_backend 1/1 ... [2025-12-04 09:21:29.184190][1725.143734133], took 0.18min 2025-12-04T09:21:29.2457026Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_extension_backend/inductor.test_extension_backend-184a28db46f4749e.xml 2025-12-04T09:21:29.2457219Z Running dynamo/test_fx_graph_runnable 1/1 ... [2025-12-04 09:21:29.216692][1725.176235076] 2025-12-04T09:21:29.2457299Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:21:29.2458062Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_fx_graph_runnable.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:21:29.217133] 2025-12-04T09:21:29.6216089Z Uploading logs for 57116084869 to S3 2025-12-04T09:21:29.6583079Z Uploading artifacts took 0.42 seconds 2025-12-04T09:21:29.6583481Z inductor/test_extension_backend 1/1 failed! 2025-12-04T09:21:31.7085331Z 2025-12-04T09:21:31.7086346Z inductor/test_cooperative_reductions 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_cooperative_reductions_1.1_07ed6b958ddfd437_.log 2025-12-04T09:21:31.7087201Z Running 0 items in this shard: 2025-12-04T09:21:31.7087389Z 2025-12-04T09:21:31.7087721Z Finished inductor/test_cooperative_reductions 1/1 ... [2025-12-04 09:21:31.708422][1727.667965769], took 0.10min 2025-12-04T09:21:31.7163406Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cooperative_reductions/inductor.test_cooperative_reductions-61458c2618582136.xml 2025-12-04T09:21:32.8339258Z Running dynamo/test_backends 1/1 ... [2025-12-04 09:21:32.833456][1728.792997383] 2025-12-04T09:21:32.8339906Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:21:32.8344302Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_backends.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:21:32.833980] 2025-12-04T09:21:35.2544252Z Running dynamo/test_functions 1/1 ... [2025-12-04 09:21:35.253973][1731.21351662] 2025-12-04T09:21:35.2544738Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:21:35.2547225Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_functions.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:21:35.254401] 2025-12-04T09:21:38.9032573Z 2025-12-04T09:21:38.9033411Z dynamo/test_backends 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_backends_1.1_732b26763f9c8444_.log 2025-12-04T09:21:38.9034545Z Running 0 items in this shard: 2025-12-04T09:21:38.9034725Z 2025-12-04T09:21:38.9035145Z Finished dynamo/test_backends 1/1 ... [2025-12-04 09:21:38.903122][1734.862664992], took 0.10min 2025-12-04T09:21:38.9113879Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_backends/dynamo.test_backends-2bf853a9ce3082ab.xml 2025-12-04T09:21:42.0670987Z 2025-12-04T09:21:42.0671934Z dynamo/test_functions 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_functions_1.1_e9fec77e4c6ac578_.log 2025-12-04T09:21:42.0672530Z Running 0 items in this shard: 2025-12-04T09:21:42.0672671Z 2025-12-04T09:21:42.0672881Z Finished dynamo/test_functions 1/1 ... [2025-12-04 09:21:42.066935][1738.026480944], took 0.11min 2025-12-04T09:21:42.0746485Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_functions/dynamo.test_functions-74b8016243633845.xml 2025-12-04T09:21:42.5290237Z Running inductor/test_mix_order_reduction 1/1 ... [2025-12-04 09:21:42.528547][1738.488091064] 2025-12-04T09:21:42.5290815Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:21:42.5292919Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_mix_order_reduction.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:21:42.528988] 2025-12-04T09:21:45.6100321Z Running inductor/test_padding 1/1 ... [2025-12-04 09:21:45.609598][1741.569142539] 2025-12-04T09:21:45.6100812Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:21:45.6103016Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_padding.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:21:45.610012] 2025-12-04T09:21:49.3113002Z 2025-12-04T09:21:49.3114229Z inductor/test_mix_order_reduction 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_mix_order_reduction_1.1_737090c6800dd0c8_.log 2025-12-04T09:21:49.3128786Z Running 50 items in this shard: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction 2025-12-04T09:21:49.3142670Z 2025-12-04T09:21:49.3142917Z Finished inductor/test_mix_order_reduction 1/1 ... [2025-12-04 09:21:49.311251][1745.270795068], took 0.11min 2025-12-04T09:21:49.3176547Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7d44f20bb1c46c3f.xml 2025-12-04T09:21:51.7615424Z 2025-12-04T09:21:51.7616179Z inductor/test_padding 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_padding_1.1_dd27bd85980e083b_.log 2025-12-04T09:21:51.7616785Z Running 0 items in this shard: 2025-12-04T09:21:51.7616936Z 2025-12-04T09:21:51.7617163Z Finished inductor/test_padding 1/1 ... [2025-12-04 09:21:51.761390][1747.720934792], took 0.10min 2025-12-04T09:21:51.7679401Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_padding/inductor.test_padding-764f2c6729c80dee.xml 2025-12-04T09:21:52.8874642Z Running dynamo/test_dicts 1/1 ... [2025-12-04 09:21:52.887006][1748.846548987] 2025-12-04T09:21:52.8876460Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:21:52.8879659Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_dicts.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:21:52.887619] 2025-12-04T09:21:55.2505833Z Running dynamo/test_aot_compile 1/1 ... [2025-12-04 09:21:55.250044][1751.209588074] 2025-12-04T09:21:55.2506549Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:21:55.2508390Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_aot_compile.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:21:55.250440] 2025-12-04T09:21:56.6471084Z 2025-12-04T09:21:56.6471967Z dynamo/test_dicts 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_dicts_1.1_de613a189e3e1504_.log 2025-12-04T09:21:56.6472643Z Running 0 items in this shard: 2025-12-04T09:21:56.6472833Z 2025-12-04T09:21:56.6473092Z Finished dynamo/test_dicts 1/1 ... [2025-12-04 09:21:56.646928][1752.606473016], took 0.06min 2025-12-04T09:21:56.6536478Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_dicts/dynamo.test_dicts-2ba8c58b73d80421.xml 2025-12-04T09:21:58.7919762Z 2025-12-04T09:21:58.7920712Z dynamo/test_aot_compile 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_aot_compile_1.1_d4d7319297f0b64f_.log 2025-12-04T09:21:58.7921603Z Running 0 items in this shard: 2025-12-04T09:21:58.7921872Z 2025-12-04T09:21:58.7922150Z Finished dynamo/test_aot_compile 1/1 ... [2025-12-04 09:21:58.791879][1754.751424934], took 0.06min 2025-12-04T09:21:58.7986652Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_aot_compile/dynamo.test_aot_compile-603f3f2f0f262772.xml 2025-12-04T09:22:00.2495309Z Running dynamo/test_sets 1/1 ... [2025-12-04 09:22:00.249021][1756.208564005] 2025-12-04T09:22:00.2495788Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:22:00.2497424Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_sets.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:22:00.249429] 2025-12-04T09:22:02.4140571Z Running dynamo/test_callback 1/1 ... [2025-12-04 09:22:02.413587][1758.37313087] 2025-12-04T09:22:02.4141141Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:22:02.4143227Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_callback.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:22:02.414012] 2025-12-04T09:22:03.8736968Z 2025-12-04T09:22:03.8737780Z dynamo/test_sets 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_sets_1.1_11315a33e392b1af_.log 2025-12-04T09:22:03.8738463Z Running 0 items in this shard: 2025-12-04T09:22:03.8738652Z 2025-12-04T09:22:03.8738891Z Finished dynamo/test_sets 1/1 ... [2025-12-04 09:22:03.873526][1759.83307257], took 0.06min 2025-12-04T09:22:03.8855535Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_sets/dynamo.test_sets-03077bc01df815c9.xml 2025-12-04T09:22:07.5041158Z Running inductor/test_cudagraph_trees_expandable_segments 1/1 ... [2025-12-04 09:22:07.503683][1763.463227504] 2025-12-04T09:22:07.5042275Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:22:07.5044230Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_cudagraph_trees_expandable_segments.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:22:07.504111] 2025-12-04T09:22:08.4521310Z 2025-12-04T09:22:08.4522306Z dynamo/test_callback 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_callback_1.1_46f26d924e4a214f_.log 2025-12-04T09:22:08.4523171Z Running 0 items in this shard: 2025-12-04T09:22:08.4523469Z 2025-12-04T09:22:08.4523764Z Finished dynamo/test_callback 1/1 ... [2025-12-04 09:22:08.452048][1764.411593236], took 0.10min 2025-12-04T09:22:08.4598102Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_callback/dynamo.test_callback-9aebdb8d0ae9ab09.xml 2025-12-04T09:22:11.9853900Z Running dynamo/test_exceptions 1/1 ... [2025-12-04 09:22:11.984914][1767.944457143] 2025-12-04T09:22:11.9854658Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:22:11.9856623Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_exceptions.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:22:11.985331] 2025-12-04T09:22:15.6627861Z 2025-12-04T09:22:15.6628808Z dynamo/test_exceptions 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_exceptions_1.1_c4596b16d2ddbb72_.log 2025-12-04T09:22:15.6629621Z Running 0 items in this shard: 2025-12-04T09:22:15.6629913Z 2025-12-04T09:22:15.6630191Z Finished dynamo/test_exceptions 1/1 ... [2025-12-04 09:22:15.662622][1771.622168984], took 0.06min 2025-12-04T09:22:15.6706499Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_exceptions/dynamo.test_exceptions-d0179208da9e5b37.xml 2025-12-04T09:22:19.2829150Z Running test_cuda 1/1 ... [2025-12-04 09:22:19.282431][1775.241973706] 2025-12-04T09:22:19.2829605Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:22:19.2831165Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_cuda.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:22:19.282818] 2025-12-04T09:22:29.1916539Z 2025-12-04T09:22:29.1918295Z inductor/test_cudagraph_trees_expandable_segments 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_cudagraph_trees_expandable_segments_1.1_944dfa8853de3b32_.log 2025-12-04T09:22:29.1936408Z Running 50 items in this shard: test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial 2025-12-04T09:22:29.1950969Z 2025-12-04T09:22:29.1951274Z Finished inductor/test_cudagraph_trees_expandable_segments 1/1 ... [2025-12-04 09:22:29.191643][1785.151186538], took 0.36min 2025-12-04T09:22:29.1994100Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cudagraph_trees_expandable_segments/inductor.test_cudagraph_trees_expandable_segments-06ab583be2814889.xml 2025-12-04T09:22:32.7168751Z Running test_transformers 1/1 ... [2025-12-04 09:22:32.716431][1788.675975364] 2025-12-04T09:22:32.7169233Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:22:32.7172038Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_transformers.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:22:32.716873] 2025-12-04T09:22:38.6266260Z 2025-12-04T09:22:38.6267149Z PRINTING LOG FILE of test_cuda 1/1 (test/test-reports/test_cuda_1.1_1d7d8eb7c1053f49_.log) 2025-12-04T09:22:38.6268368Z Test results will be stored in test-reports/python-pytest/test_cuda/test_cuda-8b00a0d732cb0448.xml 2025-12-04T09:22:38.6269334Z ============================= test session starts ============================== 2025-12-04T09:22:38.6270317Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:22:38.6271122Z cachedir: .pytest_cache 2025-12-04T09:22:38.6271932Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:22:38.6272736Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:22:38.6273112Z configfile: pytest.ini 2025-12-04T09:22:38.6273877Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:22:38.6274819Z collecting ... collected 252 items / 1400 deselected / -1148 selected 2025-12-04T09:22:38.6275389Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:22:38.6328633Z Running 300 items in this shard: test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator 2025-12-04T09:22:38.6378248Z 2025-12-04T09:22:38.6378732Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 0%] 2025-12-04T09:22:38.6379441Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0034s] [ 0%] 2025-12-04T09:22:38.6380155Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 1%] 2025-12-04T09:22:38.6380850Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot FAILED [0.0810s] [ 1%] 2025-12-04T09:22:38.6381564Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp [W1204 09:22:23.615469045 unwind.cpp:219] Warning: Unsupported unwinding pattern: Address not in range (function unwinderFor) 2025-12-04T09:22:38.6382164Z FAILED [9.2600s] [ 1%] 2025-12-04T09:22:38.6383741Z test_cuda.py::TestMemPool::test_mempool_with_allocator [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=dummy_allocator -DTORCH_API_INCLUDE_EXTENSION_H -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -isystem /usr/local/cuda/include -isystem /opt/conda/envs/py_3.10/include/python3.10 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py310_cu128/dummy_allocator/main.cpp -o main.o 2025-12-04T09:22:38.6385785Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib -lc10 -lc10_cuda -ltorch_cpu -ltorch_cuda -ltorch -ltorch_python -L/usr/local/cuda/lib64 -lcudart -o dummy_allocator.so 2025-12-04T09:22:38.6386449Z PASSED [2.2950s] [ 2%] 2025-12-04T09:22:38.6387019Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6387949Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6388873Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6389797Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6390723Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6391641Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6392567Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6393490Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6394576Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6395510Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6396430Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6397345Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6398256Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6399183Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6400099Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6401019Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6401933Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6402848Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6403762Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6405024Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6405967Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6406897Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6407829Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6408746Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6409661Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6410587Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6411511Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6412429Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6413726Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6414670Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6415602Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6416519Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6417430Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6418488Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6419436Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6420355Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6421277Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6422183Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6423112Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6424041Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6424966Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6425884Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6426796Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6427722Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6428651Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6429569Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6430486Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6431505Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6432494Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6433172Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0011s] [ 2%] 2025-12-04T09:22:38.6433601Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0008s] [ 2%] 2025-12-04T09:22:38.6434002Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0009s] [ 2%] 2025-12-04T09:22:38.6434412Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0008s] [ 2%] 2025-12-04T09:22:38.6434956Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6435367Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6435773Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6436185Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6436593Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6436991Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6437392Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6437793Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6438191Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0014s] [ 2%] 2025-12-04T09:22:38.6438585Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6438988Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6439394Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6439798Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6440205Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6440658Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6441138Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6441609Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6442081Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6442528Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6442926Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6443326Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6443728Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6444131Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6444534Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6444930Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6445344Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6445750Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6446146Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6446549Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6447057Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6447522Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6447929Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6448332Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6448736Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6449130Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6449532Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6449932Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6450334Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6450728Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6451150Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6451552Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6451958Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0011s] [ 2%] 2025-12-04T09:22:38.6452357Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0008s] [ 2%] 2025-12-04T09:22:38.6452773Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0008s] [ 2%] 2025-12-04T09:22:38.6453293Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0008s] [ 2%] 2025-12-04T09:22:38.6454172Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6455134Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6456083Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6457015Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6457956Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6458889Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6459834Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6460774Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6461721Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6462669Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6463608Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6464711Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6465654Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6466589Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6467542Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6468481Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6469427Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6470377Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6471324Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6472267Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6473204Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6474151Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6475088Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6476019Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6476950Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6477897Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6478841Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6479778Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6480706Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6481649Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6482727Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6483666Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6484606Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6485542Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6486479Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6487415Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6488353Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6489291Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6490224Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6491167Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6492108Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6493043Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6494099Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6495035Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6495978Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6496915Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6497853Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6498795Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6499737Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:22:38.6500522Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0319s] [ 2%] 2025-12-04T09:22:38.6501013Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0304s] [ 2%] 2025-12-04T09:22:38.6501457Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0297s] [ 2%] 2025-12-04T09:22:38.6501864Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0289s] [ 2%] 2025-12-04T09:22:38.6502271Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0296s] [ 2%] 2025-12-04T09:22:38.6502677Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0287s] [ 2%] 2025-12-04T09:22:38.6503086Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0288s] [ 2%] 2025-12-04T09:22:38.6503490Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0351s] [ 2%] 2025-12-04T09:22:38.6503900Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0763s] [ 2%] 2025-12-04T09:22:38.6504316Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0325s] [ 2%] 2025-12-04T09:22:38.6505113Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0286s] [ 2%] 2025-12-04T09:22:38.6505564Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0283s] [ 2%] 2025-12-04T09:22:38.6505988Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0280s] [ 2%] 2025-12-04T09:22:38.6506401Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0282s] [ 2%] 2025-12-04T09:22:38.6506809Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0281s] [ 2%] 2025-12-04T09:22:38.6507217Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0282s] [ 2%] 2025-12-04T09:22:38.6507627Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0303s] [ 2%] 2025-12-04T09:22:38.6508023Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0300s] [ 2%] 2025-12-04T09:22:38.6508437Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0297s] [ 2%] 2025-12-04T09:22:38.6508853Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0285s] [ 2%] 2025-12-04T09:22:38.6509259Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0281s] [ 2%] 2025-12-04T09:22:38.6509660Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0287s] [ 2%] 2025-12-04T09:22:38.6510064Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0290s] [ 2%] 2025-12-04T09:22:38.6510469Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0295s] [ 2%] 2025-12-04T09:22:38.6510871Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0289s] [ 2%] 2025-12-04T09:22:38.6511286Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0283s] [ 2%] 2025-12-04T09:22:38.6511694Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0281s] [ 2%] 2025-12-04T09:22:38.6512101Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0283s] [ 2%] 2025-12-04T09:22:38.6512505Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0285s] [ 2%] 2025-12-04T09:22:38.6512911Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0288s] [ 2%] 2025-12-04T09:22:38.6513318Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0300s] [ 2%] 2025-12-04T09:22:38.6513722Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0291s] [ 2%] 2025-12-04T09:22:38.6514121Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0288s] [ 2%] 2025-12-04T09:22:38.6514525Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0287s] [ 2%] 2025-12-04T09:22:38.6514929Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0308s] [ 2%] 2025-12-04T09:22:38.6515341Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0308s] [ 2%] 2025-12-04T09:22:38.6515745Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0306s] [ 2%] 2025-12-04T09:22:38.6516154Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0299s] [ 2%] 2025-12-04T09:22:38.6516828Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0289s] [ 2%] 2025-12-04T09:22:38.6517234Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0286s] [ 2%] 2025-12-04T09:22:38.6517639Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0287s] [ 2%] 2025-12-04T09:22:38.6518045Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0288s] [ 2%] 2025-12-04T09:22:38.6518446Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0285s] [ 2%] 2025-12-04T09:22:38.6518859Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0284s] [ 2%] 2025-12-04T09:22:38.6519266Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0288s] [ 2%] 2025-12-04T09:22:38.6519682Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0307s] [ 2%] 2025-12-04T09:22:38.6520085Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0298s] [ 2%] 2025-12-04T09:22:38.6520494Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0287s] [ 2%] 2025-12-04T09:22:38.6520906Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0284s] [ 2%] 2025-12-04T09:22:38.6521347Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0015s] [ 2%] 2025-12-04T09:22:38.6521810Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6522267Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6522721Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6523184Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6523650Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6524107Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6524574Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6525027Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6525484Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6525944Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6526400Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6526850Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6527309Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6527771Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6528223Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6528699Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6529161Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6529618Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6530066Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6530528Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6530992Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6531446Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6531894Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6532437Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6532959Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6533551Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6534009Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6534461Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6534924Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6535375Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6535824Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6536274Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6536730Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6537180Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6537635Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6538091Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6538543Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6538990Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6539446Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6539897Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6540344Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0013s] [ 2%] 2025-12-04T09:22:38.6540805Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6541256Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6541710Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6542156Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6542607Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0012s] [ 2%] 2025-12-04T09:22:38.6543059Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:22:38.6543510Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:22:38.6543951Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0018s] [ 2%] 2025-12-04T09:22:38.6544365Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0014s] [ 2%] 2025-12-04T09:22:38.6544770Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6545157Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6545555Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:22:38.6545947Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6546340Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6546727Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0014s] [ 2%] 2025-12-04T09:22:38.6547120Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6547512Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0016s] [ 2%] 2025-12-04T09:22:38.6547895Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6548464Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6548868Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6549259Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6549644Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6550039Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6550431Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6550820Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6551206Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6551596Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6551990Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:22:38.6552381Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6552785Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6553181Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6553573Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6553960Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:22:38.6554351Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6554743Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6555125Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6555659Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6556061Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6556450Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6556833Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6557225Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:22:38.6557611Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6557994Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6558384Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:22:38.6558770Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6559163Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6559551Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6559945Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6560334Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:22:38.6560736Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6561123Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6561515Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6561906Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6562292Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6562683Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6563169Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0013s] [ 2%] 2025-12-04T09:22:38.6563403Z 2025-12-04T09:22:38.6563566Z =================================== FAILURES =================================== 2025-12-04T09:22:38.6563891Z ___________________ TestCudaMallocAsync.test_memory_snapshot ___________________ 2025-12-04T09:22:38.6564207Z Traceback (most recent call last): 2025-12-04T09:22:38.6564579Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 4091, in test_memory_snapshot 2025-12-04T09:22:38.6564966Z torch.cuda.memory._save_segment_usage(f.name) 2025-12-04T09:22:38.6565436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/memory.py", line 1288, in _save_segment_usage 2025-12-04T09:22:38.6565876Z f.write(_segments(snapshot)) 2025-12-04T09:22:38.6566268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/_memory_viz.py", line 158, in segments 2025-12-04T09:22:38.6566680Z return format_flamegraph(f.getvalue()) 2025-12-04T09:22:38.6567119Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/_memory_viz.py", line 100, in format_flamegraph 2025-12-04T09:22:38.6567631Z with tempfile.NamedTemporaryFile(mode="wb", suffix=".pl") as f: 2025-12-04T09:22:38.6568037Z File "/opt/conda/envs/py_3.10/lib/python3.10/tempfile.py", line 518, in __exit__ 2025-12-04T09:22:38.6568350Z self.close() 2025-12-04T09:22:38.6568610Z File "/opt/conda/envs/py_3.10/lib/python3.10/tempfile.py", line 525, in close 2025-12-04T09:22:38.6568920Z self._closer.close() 2025-12-04T09:22:38.6569190Z File "/opt/conda/envs/py_3.10/lib/python3.10/tempfile.py", line 462, in close 2025-12-04T09:22:38.6569491Z unlink(self.name) 2025-12-04T09:22:38.6569778Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmputo3q17i.pl' 2025-12-04T09:22:38.6570161Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:22:38.6570531Z Downloading flamegraph.pl to: /var/lib/jenkins/.cache//flamegraph.pl 2025-12-04T09:22:38.6570945Z ______________ TestCudaMallocAsync.test_memory_snapshot_with_cpp _______________ 2025-12-04T09:22:38.6571292Z Traceback (most recent call last): 2025-12-04T09:22:38.6571659Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 4131, in test_memory_snapshot_with_cpp 2025-12-04T09:22:38.6572069Z self.assertTrue("::rand" in str(b["frames"])) 2025-12-04T09:22:38.6572430Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 687, in assertTrue 2025-12-04T09:22:38.6572772Z raise self.failureException(msg) 2025-12-04T09:22:38.6572993Z AssertionError: False is not true 2025-12-04T09:22:38.6573135Z 2025-12-04T09:22:38.6573390Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6573769Z python test/test_cuda.py TestCudaMallocAsync.test_memory_snapshot_with_cpp 2025-12-04T09:22:38.6574014Z 2025-12-04T09:22:38.6574172Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6574560Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6574862Z Traceback (most recent call last): 2025-12-04T09:22:38.6575228Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6575620Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6576101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6576619Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6576895Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6577052Z 2025-12-04T09:22:38.6577119Z Expected 0 but got 123. 2025-12-04T09:22:38.6577313Z Absolute difference: 123 2025-12-04T09:22:38.6577492Z Relative difference: inf 2025-12-04T09:22:38.6577611Z 2025-12-04T09:22:38.6577737Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6578181Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6578398Z 2025-12-04T09:22:38.6578636Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6579025Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6579326Z Traceback (most recent call last): 2025-12-04T09:22:38.6579688Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6580080Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6580552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6581069Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6581349Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6581495Z 2025-12-04T09:22:38.6581562Z Expected 0 but got 123. 2025-12-04T09:22:38.6581744Z Absolute difference: 123 2025-12-04T09:22:38.6581931Z Relative difference: inf 2025-12-04T09:22:38.6582039Z 2025-12-04T09:22:38.6582170Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6582514Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6582747Z 2025-12-04T09:22:38.6582910Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6583293Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6583585Z Traceback (most recent call last): 2025-12-04T09:22:38.6583945Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6584336Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6584815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6585315Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6585598Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6585745Z 2025-12-04T09:22:38.6585825Z Expected 0 but got 123. 2025-12-04T09:22:38.6586010Z Absolute difference: 123 2025-12-04T09:22:38.6586193Z Relative difference: inf 2025-12-04T09:22:38.6586304Z 2025-12-04T09:22:38.6586435Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6586780Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6586996Z 2025-12-04T09:22:38.6587155Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6587534Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6587835Z Traceback (most recent call last): 2025-12-04T09:22:38.6588180Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6588567Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6589050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6589552Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6589821Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6589971Z 2025-12-04T09:22:38.6590037Z Expected 0 but got 123. 2025-12-04T09:22:38.6590215Z Absolute difference: 123 2025-12-04T09:22:38.6590389Z Relative difference: inf 2025-12-04T09:22:38.6590500Z 2025-12-04T09:22:38.6590624Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6590975Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6591193Z 2025-12-04T09:22:38.6591354Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6591725Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6592019Z Traceback (most recent call last): 2025-12-04T09:22:38.6592463Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6592910Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6593389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6593891Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6594169Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6594314Z 2025-12-04T09:22:38.6594379Z Expected 0 but got 123. 2025-12-04T09:22:38.6594555Z Absolute difference: 123 2025-12-04T09:22:38.6594735Z Relative difference: inf 2025-12-04T09:22:38.6594842Z 2025-12-04T09:22:38.6594965Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6595305Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6595527Z 2025-12-04T09:22:38.6595694Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6596081Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6596373Z Traceback (most recent call last): 2025-12-04T09:22:38.6596723Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6597110Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6597581Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6598087Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6598361Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6598508Z 2025-12-04T09:22:38.6598581Z Expected 0 but got 123. 2025-12-04T09:22:38.6598750Z Absolute difference: 123 2025-12-04T09:22:38.6598927Z Relative difference: inf 2025-12-04T09:22:38.6599035Z 2025-12-04T09:22:38.6599177Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6599535Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6599764Z 2025-12-04T09:22:38.6599922Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6600308Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6600611Z Traceback (most recent call last): 2025-12-04T09:22:38.6600960Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6601353Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6601832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6602332Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6602612Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6602762Z 2025-12-04T09:22:38.6602831Z Expected 0 but got 123. 2025-12-04T09:22:38.6603008Z Absolute difference: 123 2025-12-04T09:22:38.6603183Z Relative difference: inf 2025-12-04T09:22:38.6603302Z 2025-12-04T09:22:38.6603429Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6603769Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6603982Z 2025-12-04T09:22:38.6604135Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6604830Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6605170Z Traceback (most recent call last): 2025-12-04T09:22:38.6605529Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6605907Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6606385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6607088Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6607477Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6607640Z 2025-12-04T09:22:38.6607706Z Expected 0 but got 123. 2025-12-04T09:22:38.6607885Z Absolute difference: 123 2025-12-04T09:22:38.6608062Z Relative difference: inf 2025-12-04T09:22:38.6608170Z 2025-12-04T09:22:38.6608293Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6608645Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6608858Z 2025-12-04T09:22:38.6609020Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6609402Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6609696Z Traceback (most recent call last): 2025-12-04T09:22:38.6610051Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6610441Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6610916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6611426Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6611704Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6611848Z 2025-12-04T09:22:38.6611924Z Expected 0 but got 123. 2025-12-04T09:22:38.6612105Z Absolute difference: 123 2025-12-04T09:22:38.6612283Z Relative difference: inf 2025-12-04T09:22:38.6612390Z 2025-12-04T09:22:38.6612519Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6612853Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6613071Z 2025-12-04T09:22:38.6613337Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6613723Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6614022Z Traceback (most recent call last): 2025-12-04T09:22:38.6614372Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6614757Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6615235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6615736Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6616022Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6616167Z 2025-12-04T09:22:38.6616239Z Expected 0 but got 123. 2025-12-04T09:22:38.6616416Z Absolute difference: 123 2025-12-04T09:22:38.6616590Z Relative difference: inf 2025-12-04T09:22:38.6616710Z 2025-12-04T09:22:38.6616839Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6617185Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6617403Z 2025-12-04T09:22:38.6617561Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6617943Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6618243Z Traceback (most recent call last): 2025-12-04T09:22:38.6618598Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6618981Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6619460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6619960Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6620236Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6620392Z 2025-12-04T09:22:38.6620456Z Expected 0 but got 123. 2025-12-04T09:22:38.6620637Z Absolute difference: 123 2025-12-04T09:22:38.6620817Z Relative difference: inf 2025-12-04T09:22:38.6621005Z 2025-12-04T09:22:38.6621129Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6621534Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6621747Z 2025-12-04T09:22:38.6621923Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6622300Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6622594Z Traceback (most recent call last): 2025-12-04T09:22:38.6622941Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6623337Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6623806Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6624308Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6624579Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6624726Z 2025-12-04T09:22:38.6624791Z Expected 0 but got 123. 2025-12-04T09:22:38.6624970Z Absolute difference: 123 2025-12-04T09:22:38.6625150Z Relative difference: inf 2025-12-04T09:22:38.6625259Z 2025-12-04T09:22:38.6625389Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6625724Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6625945Z 2025-12-04T09:22:38.6626100Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6626483Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6626777Z Traceback (most recent call last): 2025-12-04T09:22:38.6627130Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6627526Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6628009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6628512Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6628786Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6628931Z 2025-12-04T09:22:38.6629002Z Expected 0 but got 123. 2025-12-04T09:22:38.6629175Z Absolute difference: 123 2025-12-04T09:22:38.6629355Z Relative difference: inf 2025-12-04T09:22:38.6629462Z 2025-12-04T09:22:38.6629592Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6629933Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6630150Z 2025-12-04T09:22:38.6630307Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6630693Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6631001Z Traceback (most recent call last): 2025-12-04T09:22:38.6631348Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6631740Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6632217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6632725Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6633006Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6633164Z 2025-12-04T09:22:38.6633234Z Expected 0 but got 123. 2025-12-04T09:22:38.6633427Z Absolute difference: 123 2025-12-04T09:22:38.6633610Z Relative difference: inf 2025-12-04T09:22:38.6633729Z 2025-12-04T09:22:38.6633861Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6634216Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6634438Z 2025-12-04T09:22:38.6634603Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6635078Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6635450Z Traceback (most recent call last): 2025-12-04T09:22:38.6635814Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6636202Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6636685Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6637197Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6637476Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6637628Z 2025-12-04T09:22:38.6637695Z Expected 0 but got 123. 2025-12-04T09:22:38.6637874Z Absolute difference: 123 2025-12-04T09:22:38.6638053Z Relative difference: inf 2025-12-04T09:22:38.6638162Z 2025-12-04T09:22:38.6638288Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6638636Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6638859Z 2025-12-04T09:22:38.6639023Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6639408Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6639705Z Traceback (most recent call last): 2025-12-04T09:22:38.6640062Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6640456Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6640936Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6641442Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6641719Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6641866Z 2025-12-04T09:22:38.6641941Z Expected 0 but got 123. 2025-12-04T09:22:38.6642134Z Absolute difference: 123 2025-12-04T09:22:38.6642316Z Relative difference: inf 2025-12-04T09:22:38.6642425Z 2025-12-04T09:22:38.6642563Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6642900Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6643123Z 2025-12-04T09:22:38.6643280Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6643662Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6643967Z Traceback (most recent call last): 2025-12-04T09:22:38.6644315Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6644705Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6645184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6645687Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6645962Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6646113Z 2025-12-04T09:22:38.6646182Z Expected 0 but got 123. 2025-12-04T09:22:38.6646362Z Absolute difference: 123 2025-12-04T09:22:38.6646537Z Relative difference: inf 2025-12-04T09:22:38.6646652Z 2025-12-04T09:22:38.6646780Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6647126Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6647341Z 2025-12-04T09:22:38.6647502Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6647887Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6648192Z Traceback (most recent call last): 2025-12-04T09:22:38.6648548Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6648931Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6649557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6650070Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6650346Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6650495Z 2025-12-04T09:22:38.6650564Z Expected 0 but got 123. 2025-12-04T09:22:38.6650746Z Absolute difference: 123 2025-12-04T09:22:38.6650935Z Relative difference: inf 2025-12-04T09:22:38.6651044Z 2025-12-04T09:22:38.6651169Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6651513Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6651729Z 2025-12-04T09:22:38.6651893Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6652279Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6652579Z Traceback (most recent call last): 2025-12-04T09:22:38.6652944Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6653412Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6653887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6654394Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6654666Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6654810Z 2025-12-04T09:22:38.6654881Z Expected 0 but got 123. 2025-12-04T09:22:38.6655051Z Absolute difference: 123 2025-12-04T09:22:38.6655230Z Relative difference: inf 2025-12-04T09:22:38.6655341Z 2025-12-04T09:22:38.6655474Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6655813Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6656034Z 2025-12-04T09:22:38.6656197Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6656579Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6656876Z Traceback (most recent call last): 2025-12-04T09:22:38.6657225Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6657624Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6658105Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6658608Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6658885Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6659036Z 2025-12-04T09:22:38.6659105Z Expected 0 but got 123. 2025-12-04T09:22:38.6659283Z Absolute difference: 123 2025-12-04T09:22:38.6659456Z Relative difference: inf 2025-12-04T09:22:38.6659571Z 2025-12-04T09:22:38.6659702Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6660046Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6660261Z 2025-12-04T09:22:38.6660418Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6660805Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6661103Z Traceback (most recent call last): 2025-12-04T09:22:38.6661458Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6661854Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6662342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6662852Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6663121Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6663274Z 2025-12-04T09:22:38.6674144Z Expected 0 but got 123. 2025-12-04T09:22:38.6674434Z Absolute difference: 123 2025-12-04T09:22:38.6674783Z Relative difference: inf 2025-12-04T09:22:38.6674917Z 2025-12-04T09:22:38.6675070Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6675465Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6675696Z 2025-12-04T09:22:38.6675873Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6676273Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6676604Z Traceback (most recent call last): 2025-12-04T09:22:38.6677010Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6677438Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6677945Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6678508Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6678805Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6678959Z 2025-12-04T09:22:38.6679031Z Expected 0 but got 123. 2025-12-04T09:22:38.6679222Z Absolute difference: 123 2025-12-04T09:22:38.6679418Z Relative difference: inf 2025-12-04T09:22:38.6679534Z 2025-12-04T09:22:38.6679667Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6680028Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6680256Z 2025-12-04T09:22:38.6680421Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6680816Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6681127Z Traceback (most recent call last): 2025-12-04T09:22:38.6681496Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6681914Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6682411Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6682923Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6683209Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6683359Z 2025-12-04T09:22:38.6683436Z Expected 0 but got 123. 2025-12-04T09:22:38.6683615Z Absolute difference: 123 2025-12-04T09:22:38.6683801Z Relative difference: inf 2025-12-04T09:22:38.6683912Z 2025-12-04T09:22:38.6684050Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6684407Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6684630Z 2025-12-04T09:22:38.6684791Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6685184Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6685507Z Traceback (most recent call last): 2025-12-04T09:22:38.6685873Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6686273Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6686767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6687290Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6687571Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6687728Z 2025-12-04T09:22:38.6687797Z Expected 0 but got 123. 2025-12-04T09:22:38.6687995Z Absolute difference: 123 2025-12-04T09:22:38.6688175Z Relative difference: inf 2025-12-04T09:22:38.6688292Z 2025-12-04T09:22:38.6688422Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6688773Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6689083Z 2025-12-04T09:22:38.6689253Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6689703Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6690012Z Traceback (most recent call last): 2025-12-04T09:22:38.6690379Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6690783Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6691276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6691793Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6692077Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6692229Z 2025-12-04T09:22:38.6692297Z Expected 0 but got 123. 2025-12-04T09:22:38.6692481Z Absolute difference: 123 2025-12-04T09:22:38.6692671Z Relative difference: inf 2025-12-04T09:22:38.6692787Z 2025-12-04T09:22:38.6692925Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6693401Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6693628Z 2025-12-04T09:22:38.6693792Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6694191Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6694492Z Traceback (most recent call last): 2025-12-04T09:22:38.6694863Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6695264Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6695750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6696264Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6696549Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6696698Z 2025-12-04T09:22:38.6696772Z Expected 0 but got 123. 2025-12-04T09:22:38.6696953Z Absolute difference: 123 2025-12-04T09:22:38.6697147Z Relative difference: inf 2025-12-04T09:22:38.6697260Z 2025-12-04T09:22:38.6697393Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6697740Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6697961Z 2025-12-04T09:22:38.6698116Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6698494Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6698791Z Traceback (most recent call last): 2025-12-04T09:22:38.6699145Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6699539Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6700016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6700530Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6700819Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6700964Z 2025-12-04T09:22:38.6701037Z Expected 0 but got 123. 2025-12-04T09:22:38.6701209Z Absolute difference: 123 2025-12-04T09:22:38.6701391Z Relative difference: inf 2025-12-04T09:22:38.6701507Z 2025-12-04T09:22:38.6701634Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6701976Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6702191Z 2025-12-04T09:22:38.6702356Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6702738Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6703043Z Traceback (most recent call last): 2025-12-04T09:22:38.6703393Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6703930Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6704411Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6705286Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6705565Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6705718Z 2025-12-04T09:22:38.6705789Z Expected 0 but got 123. 2025-12-04T09:22:38.6705969Z Absolute difference: 123 2025-12-04T09:22:38.6706144Z Relative difference: inf 2025-12-04T09:22:38.6706261Z 2025-12-04T09:22:38.6706397Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6706740Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6706956Z 2025-12-04T09:22:38.6707120Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6707502Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6707806Z Traceback (most recent call last): 2025-12-04T09:22:38.6708163Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6708554Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6709033Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6709543Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6709823Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6709970Z 2025-12-04T09:22:38.6710036Z Expected 0 but got 123. 2025-12-04T09:22:38.6710215Z Absolute difference: 123 2025-12-04T09:22:38.6710414Z Relative difference: inf 2025-12-04T09:22:38.6710525Z 2025-12-04T09:22:38.6710658Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6711000Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6711236Z 2025-12-04T09:22:38.6711397Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6711784Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6712084Z Traceback (most recent call last): 2025-12-04T09:22:38.6712432Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6712825Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6713308Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6713813Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6714092Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6714239Z 2025-12-04T09:22:38.6714307Z Expected 0 but got 123. 2025-12-04T09:22:38.6714474Z Absolute difference: 123 2025-12-04T09:22:38.6714654Z Relative difference: inf 2025-12-04T09:22:38.6714765Z 2025-12-04T09:22:38.6714898Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6715243Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6715459Z 2025-12-04T09:22:38.6715616Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6715991Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6716292Z Traceback (most recent call last): 2025-12-04T09:22:38.6716641Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6717026Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6717506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6718008Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6718519Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6718669Z 2025-12-04T09:22:38.6718840Z Expected 0 but got 123. 2025-12-04T09:22:38.6719024Z Absolute difference: 123 2025-12-04T09:22:38.6719203Z Relative difference: inf 2025-12-04T09:22:38.6719324Z 2025-12-04T09:22:38.6719452Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6719795Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6720017Z 2025-12-04T09:22:38.6720178Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6720558Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6720863Z Traceback (most recent call last): 2025-12-04T09:22:38.6721213Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6721594Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6722086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6722600Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6722881Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6723028Z 2025-12-04T09:22:38.6723095Z Expected 0 but got 123. 2025-12-04T09:22:38.6723287Z Absolute difference: 123 2025-12-04T09:22:38.6723472Z Relative difference: inf 2025-12-04T09:22:38.6723581Z 2025-12-04T09:22:38.6723708Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6724058Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6724281Z 2025-12-04T09:22:38.6724438Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6724825Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6725128Z Traceback (most recent call last): 2025-12-04T09:22:38.6725487Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6725880Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6726360Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6726868Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6727154Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6727303Z 2025-12-04T09:22:38.6727381Z Expected 0 but got 123. 2025-12-04T09:22:38.6727555Z Absolute difference: 123 2025-12-04T09:22:38.6727739Z Relative difference: inf 2025-12-04T09:22:38.6727849Z 2025-12-04T09:22:38.6727981Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6728321Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6728545Z 2025-12-04T09:22:38.6728708Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6729109Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6729420Z Traceback (most recent call last): 2025-12-04T09:22:38.6729771Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6730166Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6730648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6731152Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6731428Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6731577Z 2025-12-04T09:22:38.6731643Z Expected 0 but got 123. 2025-12-04T09:22:38.6731820Z Absolute difference: 123 2025-12-04T09:22:38.6731991Z Relative difference: inf 2025-12-04T09:22:38.6732105Z 2025-12-04T09:22:38.6732345Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6732761Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6732985Z 2025-12-04T09:22:38.6733235Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6733624Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6733924Z Traceback (most recent call last): 2025-12-04T09:22:38.6734280Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6734665Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6735141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6735666Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6735944Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6736104Z 2025-12-04T09:22:38.6736173Z Expected 0 but got 123. 2025-12-04T09:22:38.6736354Z Absolute difference: 123 2025-12-04T09:22:38.6736546Z Relative difference: inf 2025-12-04T09:22:38.6736657Z 2025-12-04T09:22:38.6736782Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6737131Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6737347Z 2025-12-04T09:22:38.6737511Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6737897Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6738195Z Traceback (most recent call last): 2025-12-04T09:22:38.6738556Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6738948Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6739433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6739948Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6740231Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6740378Z 2025-12-04T09:22:38.6740450Z Expected 0 but got 123. 2025-12-04T09:22:38.6740626Z Absolute difference: 123 2025-12-04T09:22:38.6740814Z Relative difference: inf 2025-12-04T09:22:38.6740928Z 2025-12-04T09:22:38.6741067Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6741424Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6741655Z 2025-12-04T09:22:38.6741816Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6742209Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6742515Z Traceback (most recent call last): 2025-12-04T09:22:38.6742873Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6743282Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6743772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6744285Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6744580Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6744730Z 2025-12-04T09:22:38.6744806Z Expected 0 but got 123. 2025-12-04T09:22:38.6744997Z Absolute difference: 123 2025-12-04T09:22:38.6745180Z Relative difference: inf 2025-12-04T09:22:38.6745300Z 2025-12-04T09:22:38.6745431Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6745782Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6746001Z 2025-12-04T09:22:38.6746164Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6746562Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6746964Z Traceback (most recent call last): 2025-12-04T09:22:38.6747387Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6747783Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6748266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6748779Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6749054Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6749206Z 2025-12-04T09:22:38.6749274Z Expected 0 but got 123. 2025-12-04T09:22:38.6749457Z Absolute difference: 123 2025-12-04T09:22:38.6749639Z Relative difference: inf 2025-12-04T09:22:38.6749750Z 2025-12-04T09:22:38.6749875Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6750224Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6750446Z 2025-12-04T09:22:38.6750627Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6751024Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6751327Z Traceback (most recent call last): 2025-12-04T09:22:38.6751683Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6752076Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6752558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6753070Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6753354Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6753499Z 2025-12-04T09:22:38.6753565Z Expected 0 but got 123. 2025-12-04T09:22:38.6753746Z Absolute difference: 123 2025-12-04T09:22:38.6753932Z Relative difference: inf 2025-12-04T09:22:38.6754044Z 2025-12-04T09:22:38.6754177Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6754528Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6754754Z 2025-12-04T09:22:38.6754911Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6755298Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6755595Z Traceback (most recent call last): 2025-12-04T09:22:38.6755954Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6756346Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6756830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6757333Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6757617Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6757779Z 2025-12-04T09:22:38.6757851Z Expected 0 but got 123. 2025-12-04T09:22:38.6758029Z Absolute difference: 123 2025-12-04T09:22:38.6758211Z Relative difference: inf 2025-12-04T09:22:38.6758323Z 2025-12-04T09:22:38.6758458Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6758807Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6759028Z 2025-12-04T09:22:38.6759186Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6759573Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6759880Z Traceback (most recent call last): 2025-12-04T09:22:38.6760228Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6760622Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6761112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6761783Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6762065Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6762220Z 2025-12-04T09:22:38.6762289Z Expected 0 but got 123. 2025-12-04T09:22:38.6762469Z Absolute difference: 123 2025-12-04T09:22:38.6762647Z Relative difference: inf 2025-12-04T09:22:38.6762767Z 2025-12-04T09:22:38.6762894Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6763243Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6763459Z 2025-12-04T09:22:38.6763621Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6764002Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6764306Z Traceback (most recent call last): 2025-12-04T09:22:38.6764659Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6765071Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6765552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6766064Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6766344Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6766490Z 2025-12-04T09:22:38.6766559Z Expected 0 but got 123. 2025-12-04T09:22:38.6766751Z Absolute difference: 123 2025-12-04T09:22:38.6766937Z Relative difference: inf 2025-12-04T09:22:38.6767047Z 2025-12-04T09:22:38.6767180Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6767521Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6767749Z 2025-12-04T09:22:38.6767910Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6768303Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6768601Z Traceback (most recent call last): 2025-12-04T09:22:38.6768964Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6769357Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6769838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6770346Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6770639Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6770787Z 2025-12-04T09:22:38.6770862Z Expected 0 but got 123. 2025-12-04T09:22:38.6771036Z Absolute difference: 123 2025-12-04T09:22:38.6771228Z Relative difference: inf 2025-12-04T09:22:38.6771338Z 2025-12-04T09:22:38.6771467Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6771817Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6772032Z 2025-12-04T09:22:38.6772190Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6772576Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6772877Z Traceback (most recent call last): 2025-12-04T09:22:38.6773312Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6773707Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6774187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6774695Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6774966Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6775117Z 2025-12-04T09:22:38.6775184Z Expected 0 but got 123. 2025-12-04T09:22:38.6775446Z Absolute difference: 123 2025-12-04T09:22:38.6775622Z Relative difference: inf 2025-12-04T09:22:38.6775741Z 2025-12-04T09:22:38.6775960Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6776314Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6776529Z 2025-12-04T09:22:38.6776691Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6777070Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6777382Z Traceback (most recent call last): 2025-12-04T09:22:38.6777743Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6778131Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6778617Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6779139Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6779430Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6779579Z 2025-12-04T09:22:38.6779648Z Expected 0 but got 123. 2025-12-04T09:22:38.6779826Z Absolute difference: 123 2025-12-04T09:22:38.6780007Z Relative difference: inf 2025-12-04T09:22:38.6780118Z 2025-12-04T09:22:38.6780243Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6780590Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6780821Z 2025-12-04T09:22:38.6780979Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6781366Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6781664Z Traceback (most recent call last): 2025-12-04T09:22:38.6782026Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6782420Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6782919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6783436Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6783714Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6783859Z 2025-12-04T09:22:38.6783933Z Expected 0 but got 123. 2025-12-04T09:22:38.6784104Z Absolute difference: 123 2025-12-04T09:22:38.6784289Z Relative difference: inf 2025-12-04T09:22:38.6784401Z 2025-12-04T09:22:38.6784535Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6784869Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6785094Z 2025-12-04T09:22:38.6785251Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6785638Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6785941Z Traceback (most recent call last): 2025-12-04T09:22:38.6786293Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6786684Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6787164Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6787675Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6787947Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6788097Z 2025-12-04T09:22:38.6788162Z Expected 0 but got 123. 2025-12-04T09:22:38.6788346Z Absolute difference: 123 2025-12-04T09:22:38.6788527Z Relative difference: inf 2025-12-04T09:22:38.6788642Z 2025-12-04T09:22:38.6788768Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6789111Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6789326Z 2025-12-04T09:22:38.6789570Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6790011Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6790319Z Traceback (most recent call last): 2025-12-04T09:22:38.6790680Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6791065Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6791548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6792059Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6792339Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6792485Z 2025-12-04T09:22:38.6792555Z Expected 0 but got 123. 2025-12-04T09:22:38.6792734Z Absolute difference: 123 2025-12-04T09:22:38.6792917Z Relative difference: inf 2025-12-04T09:22:38.6793027Z 2025-12-04T09:22:38.6793155Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6793514Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6793732Z 2025-12-04T09:22:38.6793896Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6794281Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:22:38.6794582Z Traceback (most recent call last): 2025-12-04T09:22:38.6794941Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:22:38.6795342Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:22:38.6795824Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:22:38.6796335Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:22:38.6796616Z AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6796770Z 2025-12-04T09:22:38.6796846Z Expected 0 but got 123. 2025-12-04T09:22:38.6797021Z Absolute difference: 123 2025-12-04T09:22:38.6797209Z Relative difference: inf 2025-12-04T09:22:38.6797322Z 2025-12-04T09:22:38.6797454Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6797793Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6798017Z 2025-12-04T09:22:38.6798173Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6798749Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda/test_cuda-8b00a0d732cb0448.xml - 2025-12-04T09:22:38.6799283Z =========================== short test summary info ============================ 2025-12-04T09:22:38.6799886Z FAILED [0.0810s] test_cuda.py::TestCudaMallocAsync::test_memory_snapshot - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmputo3q17i.pl' 2025-12-04T09:22:38.6800616Z FAILED [9.2600s] test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp - AssertionError: False is not true 2025-12-04T09:22:38.6800979Z 2025-12-04T09:22:38.6801111Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6801508Z python test/test_cuda.py TestCudaMallocAsync.test_memory_snapshot_with_cpp 2025-12-04T09:22:38.6801758Z 2025-12-04T09:22:38.6801918Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6802410Z FAILED [0.0018s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6802753Z 2025-12-04T09:22:38.6802822Z Expected 0 but got 123. 2025-12-04T09:22:38.6803017Z Absolute difference: 123 2025-12-04T09:22:38.6803196Z Relative difference: inf 2025-12-04T09:22:38.6803318Z 2025-12-04T09:22:38.6803445Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6803792Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6804105Z 2025-12-04T09:22:38.6804271Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6806225Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6806605Z 2025-12-04T09:22:38.6806677Z Expected 0 but got 123. 2025-12-04T09:22:38.6806862Z Absolute difference: 123 2025-12-04T09:22:38.6807039Z Relative difference: inf 2025-12-04T09:22:38.6807155Z 2025-12-04T09:22:38.6807286Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6807643Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6807867Z 2025-12-04T09:22:38.6808039Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6808535Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6808871Z 2025-12-04T09:22:38.6808938Z Expected 0 but got 123. 2025-12-04T09:22:38.6809122Z Absolute difference: 123 2025-12-04T09:22:38.6809298Z Relative difference: inf 2025-12-04T09:22:38.6809419Z 2025-12-04T09:22:38.6809549Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6809900Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6810119Z 2025-12-04T09:22:38.6810281Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6810762Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6811098Z 2025-12-04T09:22:38.6811165Z Expected 0 but got 123. 2025-12-04T09:22:38.6811345Z Absolute difference: 123 2025-12-04T09:22:38.6811517Z Relative difference: inf 2025-12-04T09:22:38.6811646Z 2025-12-04T09:22:38.6811772Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6812117Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6812338Z 2025-12-04T09:22:38.6812502Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6812982Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6813414Z 2025-12-04T09:22:38.6813481Z Expected 0 but got 123. 2025-12-04T09:22:38.6813657Z Absolute difference: 123 2025-12-04T09:22:38.6813836Z Relative difference: inf 2025-12-04T09:22:38.6813945Z 2025-12-04T09:22:38.6814070Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6814415Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6814629Z 2025-12-04T09:22:38.6814789Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6815275Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6815286Z 2025-12-04T09:22:38.6815358Z Expected 0 but got 123. 2025-12-04T09:22:38.6815426Z Absolute difference: 123 2025-12-04T09:22:38.6815497Z Relative difference: inf 2025-12-04T09:22:38.6815503Z 2025-12-04T09:22:38.6815631Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6815771Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6815774Z 2025-12-04T09:22:38.6815947Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6816198Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6816202Z 2025-12-04T09:22:38.6816274Z Expected 0 but got 123. 2025-12-04T09:22:38.6816340Z Absolute difference: 123 2025-12-04T09:22:38.6816408Z Relative difference: inf 2025-12-04T09:22:38.6816411Z 2025-12-04T09:22:38.6816539Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6816678Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6816871Z 2025-12-04T09:22:38.6817093Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6817361Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6817365Z 2025-12-04T09:22:38.6817432Z Expected 0 but got 123. 2025-12-04T09:22:38.6817505Z Absolute difference: 123 2025-12-04T09:22:38.6817572Z Relative difference: inf 2025-12-04T09:22:38.6817575Z 2025-12-04T09:22:38.6817698Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6817844Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6817847Z 2025-12-04T09:22:38.6818000Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6818251Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6818253Z 2025-12-04T09:22:38.6818333Z Expected 0 but got 123. 2025-12-04T09:22:38.6818402Z Absolute difference: 123 2025-12-04T09:22:38.6818480Z Relative difference: inf 2025-12-04T09:22:38.6818483Z 2025-12-04T09:22:38.6818606Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6818742Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6818752Z 2025-12-04T09:22:38.6818906Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6819151Z FAILED [0.0016s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6819154Z 2025-12-04T09:22:38.6819230Z Expected 0 but got 123. 2025-12-04T09:22:38.6819299Z Absolute difference: 123 2025-12-04T09:22:38.6819364Z Relative difference: inf 2025-12-04T09:22:38.6819367Z 2025-12-04T09:22:38.6819496Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6819635Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6819641Z 2025-12-04T09:22:38.6819803Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6820049Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6820052Z 2025-12-04T09:22:38.6820119Z Expected 0 but got 123. 2025-12-04T09:22:38.6820197Z Absolute difference: 123 2025-12-04T09:22:38.6820264Z Relative difference: inf 2025-12-04T09:22:38.6820267Z 2025-12-04T09:22:38.6820407Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6820546Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6820550Z 2025-12-04T09:22:38.6820702Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6820959Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6820965Z 2025-12-04T09:22:38.6821030Z Expected 0 but got 123. 2025-12-04T09:22:38.6821096Z Absolute difference: 123 2025-12-04T09:22:38.6821168Z Relative difference: inf 2025-12-04T09:22:38.6821171Z 2025-12-04T09:22:38.6821293Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6821446Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6821450Z 2025-12-04T09:22:38.6821602Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6821845Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6821849Z 2025-12-04T09:22:38.6821919Z Expected 0 but got 123. 2025-12-04T09:22:38.6821986Z Absolute difference: 123 2025-12-04T09:22:38.6822058Z Relative difference: inf 2025-12-04T09:22:38.6822061Z 2025-12-04T09:22:38.6822182Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6822318Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6822399Z 2025-12-04T09:22:38.6822643Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6822892Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6822895Z 2025-12-04T09:22:38.6822967Z Expected 0 but got 123. 2025-12-04T09:22:38.6823033Z Absolute difference: 123 2025-12-04T09:22:38.6823099Z Relative difference: inf 2025-12-04T09:22:38.6823102Z 2025-12-04T09:22:38.6823229Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6823366Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6823369Z 2025-12-04T09:22:38.6823521Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6823784Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6823791Z 2025-12-04T09:22:38.6823858Z Expected 0 but got 123. 2025-12-04T09:22:38.6823929Z Absolute difference: 123 2025-12-04T09:22:38.6823997Z Relative difference: inf 2025-12-04T09:22:38.6824000Z 2025-12-04T09:22:38.6824121Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6824264Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6824268Z 2025-12-04T09:22:38.6824419Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6824669Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6824673Z 2025-12-04T09:22:38.6824737Z Expected 0 but got 123. 2025-12-04T09:22:38.6824806Z Absolute difference: 123 2025-12-04T09:22:38.6824877Z Relative difference: inf 2025-12-04T09:22:38.6824880Z 2025-12-04T09:22:38.6825001Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6825140Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6825149Z 2025-12-04T09:22:38.6825304Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6825552Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6825555Z 2025-12-04T09:22:38.6825625Z Expected 0 but got 123. 2025-12-04T09:22:38.6825692Z Absolute difference: 123 2025-12-04T09:22:38.6825758Z Relative difference: inf 2025-12-04T09:22:38.6825762Z 2025-12-04T09:22:38.6825890Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6826025Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6826028Z 2025-12-04T09:22:38.6826183Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6826424Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6826430Z 2025-12-04T09:22:38.6826496Z Expected 0 but got 123. 2025-12-04T09:22:38.6826568Z Absolute difference: 123 2025-12-04T09:22:38.6826637Z Relative difference: inf 2025-12-04T09:22:38.6826640Z 2025-12-04T09:22:38.6826764Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6826919Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6826923Z 2025-12-04T09:22:38.6827076Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6827327Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6827330Z 2025-12-04T09:22:38.6827397Z Expected 0 but got 123. 2025-12-04T09:22:38.6827463Z Absolute difference: 123 2025-12-04T09:22:38.6827536Z Relative difference: inf 2025-12-04T09:22:38.6827539Z 2025-12-04T09:22:38.6827661Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6827881Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6827884Z 2025-12-04T09:22:38.6828102Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6828358Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6828361Z 2025-12-04T09:22:38.6828434Z Expected 0 but got 123. 2025-12-04T09:22:38.6828504Z Absolute difference: 123 2025-12-04T09:22:38.6828571Z Relative difference: inf 2025-12-04T09:22:38.6828580Z 2025-12-04T09:22:38.6828703Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6828839Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6828842Z 2025-12-04T09:22:38.6829000Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6829243Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6829249Z 2025-12-04T09:22:38.6829317Z Expected 0 but got 123. 2025-12-04T09:22:38.6829393Z Absolute difference: 123 2025-12-04T09:22:38.6829460Z Relative difference: inf 2025-12-04T09:22:38.6829464Z 2025-12-04T09:22:38.6829588Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6829724Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6829728Z 2025-12-04T09:22:38.6829885Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6830138Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6830141Z 2025-12-04T09:22:38.6830208Z Expected 0 but got 123. 2025-12-04T09:22:38.6830283Z Absolute difference: 123 2025-12-04T09:22:38.6830349Z Relative difference: inf 2025-12-04T09:22:38.6830352Z 2025-12-04T09:22:38.6830475Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6830621Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6830624Z 2025-12-04T09:22:38.6830779Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6831041Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6831051Z 2025-12-04T09:22:38.6831117Z Expected 0 but got 123. 2025-12-04T09:22:38.6831184Z Absolute difference: 123 2025-12-04T09:22:38.6831259Z Relative difference: inf 2025-12-04T09:22:38.6831262Z 2025-12-04T09:22:38.6831384Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6831522Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6831526Z 2025-12-04T09:22:38.6831685Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6831929Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6831935Z 2025-12-04T09:22:38.6832009Z Expected 0 but got 123. 2025-12-04T09:22:38.6832080Z Absolute difference: 123 2025-12-04T09:22:38.6832147Z Relative difference: inf 2025-12-04T09:22:38.6832150Z 2025-12-04T09:22:38.6832281Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6832416Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6832419Z 2025-12-04T09:22:38.6832575Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6832819Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6832822Z 2025-12-04T09:22:38.6832885Z Expected 0 but got 123. 2025-12-04T09:22:38.6832959Z Absolute difference: 123 2025-12-04T09:22:38.6833025Z Relative difference: inf 2025-12-04T09:22:38.6833028Z 2025-12-04T09:22:38.6833147Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6833379Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6833382Z 2025-12-04T09:22:38.6833595Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6833848Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6833851Z 2025-12-04T09:22:38.6833919Z Expected 0 but got 123. 2025-12-04T09:22:38.6833985Z Absolute difference: 123 2025-12-04T09:22:38.6834056Z Relative difference: inf 2025-12-04T09:22:38.6834059Z 2025-12-04T09:22:38.6834180Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6834322Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6834326Z 2025-12-04T09:22:38.6834477Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6834722Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6834729Z 2025-12-04T09:22:38.6834801Z Expected 0 but got 123. 2025-12-04T09:22:38.6834869Z Absolute difference: 123 2025-12-04T09:22:38.6834934Z Relative difference: inf 2025-12-04T09:22:38.6834937Z 2025-12-04T09:22:38.6835063Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6835198Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6835201Z 2025-12-04T09:22:38.6835357Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6835601Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6835604Z 2025-12-04T09:22:38.6835667Z Expected 0 but got 123. 2025-12-04T09:22:38.6835740Z Absolute difference: 123 2025-12-04T09:22:38.6835805Z Relative difference: inf 2025-12-04T09:22:38.6835808Z 2025-12-04T09:22:38.6835936Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6836077Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6836082Z 2025-12-04T09:22:38.6836235Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6836482Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6836485Z 2025-12-04T09:22:38.6836551Z Expected 0 but got 123. 2025-12-04T09:22:38.6836622Z Absolute difference: 123 2025-12-04T09:22:38.6836687Z Relative difference: inf 2025-12-04T09:22:38.6836690Z 2025-12-04T09:22:38.6836811Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6836952Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6836954Z 2025-12-04T09:22:38.6837105Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6837346Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6837358Z 2025-12-04T09:22:38.6837428Z Expected 0 but got 123. 2025-12-04T09:22:38.6837494Z Absolute difference: 123 2025-12-04T09:22:38.6837572Z Relative difference: inf 2025-12-04T09:22:38.6837575Z 2025-12-04T09:22:38.6837694Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6837840Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6837843Z 2025-12-04T09:22:38.6838006Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6838250Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6843249Z 2025-12-04T09:22:38.6843370Z Expected 0 but got 123. 2025-12-04T09:22:38.6843456Z Absolute difference: 123 2025-12-04T09:22:38.6843525Z Relative difference: inf 2025-12-04T09:22:38.6843530Z 2025-12-04T09:22:38.6843677Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6843939Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6844013Z 2025-12-04T09:22:38.6844199Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6844482Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6844486Z 2025-12-04T09:22:38.6844559Z Expected 0 but got 123. 2025-12-04T09:22:38.6844632Z Absolute difference: 123 2025-12-04T09:22:38.6844705Z Relative difference: inf 2025-12-04T09:22:38.6844708Z 2025-12-04T09:22:38.6844844Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6845002Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6845030Z 2025-12-04T09:22:38.6845200Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6845459Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6845464Z 2025-12-04T09:22:38.6845539Z Expected 0 but got 123. 2025-12-04T09:22:38.6845606Z Absolute difference: 123 2025-12-04T09:22:38.6845674Z Relative difference: inf 2025-12-04T09:22:38.6845677Z 2025-12-04T09:22:38.6845810Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6845949Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6845952Z 2025-12-04T09:22:38.6846119Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6846376Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6846382Z 2025-12-04T09:22:38.6846450Z Expected 0 but got 123. 2025-12-04T09:22:38.6846523Z Absolute difference: 123 2025-12-04T09:22:38.6846587Z Relative difference: inf 2025-12-04T09:22:38.6846591Z 2025-12-04T09:22:38.6846723Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6846882Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6846886Z 2025-12-04T09:22:38.6847046Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6847303Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6847306Z 2025-12-04T09:22:38.6847374Z Expected 0 but got 123. 2025-12-04T09:22:38.6847442Z Absolute difference: 123 2025-12-04T09:22:38.6847521Z Relative difference: inf 2025-12-04T09:22:38.6847524Z 2025-12-04T09:22:38.6847652Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6847799Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6847805Z 2025-12-04T09:22:38.6847960Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6848204Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6848209Z 2025-12-04T09:22:38.6848286Z Expected 0 but got 123. 2025-12-04T09:22:38.6848359Z Absolute difference: 123 2025-12-04T09:22:38.6848425Z Relative difference: inf 2025-12-04T09:22:38.6848435Z 2025-12-04T09:22:38.6848557Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6848692Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6848696Z 2025-12-04T09:22:38.6848855Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6849097Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6849181Z 2025-12-04T09:22:38.6849248Z Expected 0 but got 123. 2025-12-04T09:22:38.6849320Z Absolute difference: 123 2025-12-04T09:22:38.6849385Z Relative difference: inf 2025-12-04T09:22:38.6849389Z 2025-12-04T09:22:38.6849577Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6849779Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6849785Z 2025-12-04T09:22:38.6849939Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6850190Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6850193Z 2025-12-04T09:22:38.6850257Z Expected 0 but got 123. 2025-12-04T09:22:38.6850327Z Absolute difference: 123 2025-12-04T09:22:38.6850390Z Relative difference: inf 2025-12-04T09:22:38.6850393Z 2025-12-04T09:22:38.6850514Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6850655Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6850658Z 2025-12-04T09:22:38.6850805Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6851047Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6851059Z 2025-12-04T09:22:38.6851122Z Expected 0 but got 123. 2025-12-04T09:22:38.6851186Z Absolute difference: 123 2025-12-04T09:22:38.6851258Z Relative difference: inf 2025-12-04T09:22:38.6851261Z 2025-12-04T09:22:38.6851379Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6851513Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6851517Z 2025-12-04T09:22:38.6851671Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6851912Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6851917Z 2025-12-04T09:22:38.6851992Z Expected 0 but got 123. 2025-12-04T09:22:38.6852060Z Absolute difference: 123 2025-12-04T09:22:38.6852125Z Relative difference: inf 2025-12-04T09:22:38.6852130Z 2025-12-04T09:22:38.6852255Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6852389Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6852392Z 2025-12-04T09:22:38.6852545Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6852786Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6852789Z 2025-12-04T09:22:38.6852852Z Expected 0 but got 123. 2025-12-04T09:22:38.6852922Z Absolute difference: 123 2025-12-04T09:22:38.6852989Z Relative difference: inf 2025-12-04T09:22:38.6852992Z 2025-12-04T09:22:38.6853109Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6853380Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6853383Z 2025-12-04T09:22:38.6853532Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6853781Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6853787Z 2025-12-04T09:22:38.6853850Z Expected 0 but got 123. 2025-12-04T09:22:38.6853915Z Absolute difference: 123 2025-12-04T09:22:38.6853984Z Relative difference: inf 2025-12-04T09:22:38.6853987Z 2025-12-04T09:22:38.6854106Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6854248Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6854251Z 2025-12-04T09:22:38.6854404Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6854650Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6854722Z 2025-12-04T09:22:38.6854802Z Expected 0 but got 123. 2025-12-04T09:22:38.6854870Z Absolute difference: 123 2025-12-04T09:22:38.6854935Z Relative difference: inf 2025-12-04T09:22:38.6854975Z 2025-12-04T09:22:38.6855108Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6855322Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6855327Z 2025-12-04T09:22:38.6855488Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6855730Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6855733Z 2025-12-04T09:22:38.6855798Z Expected 0 but got 123. 2025-12-04T09:22:38.6855869Z Absolute difference: 123 2025-12-04T09:22:38.6855934Z Relative difference: inf 2025-12-04T09:22:38.6855937Z 2025-12-04T09:22:38.6856066Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6856203Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6856206Z 2025-12-04T09:22:38.6856356Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6856611Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6856615Z 2025-12-04T09:22:38.6856678Z Expected 0 but got 123. 2025-12-04T09:22:38.6856746Z Absolute difference: 123 2025-12-04T09:22:38.6856819Z Relative difference: inf 2025-12-04T09:22:38.6856822Z 2025-12-04T09:22:38.6856942Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6857083Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6857086Z 2025-12-04T09:22:38.6857237Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6857480Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6857491Z 2025-12-04T09:22:38.6857555Z Expected 0 but got 123. 2025-12-04T09:22:38.6857619Z Absolute difference: 123 2025-12-04T09:22:38.6857693Z Relative difference: inf 2025-12-04T09:22:38.6857698Z 2025-12-04T09:22:38.6857822Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6857959Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6857962Z 2025-12-04T09:22:38.6858117Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6858357Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6858360Z 2025-12-04T09:22:38.6858432Z Expected 0 but got 123. 2025-12-04T09:22:38.6858498Z Absolute difference: 123 2025-12-04T09:22:38.6858562Z Relative difference: inf 2025-12-04T09:22:38.6858565Z 2025-12-04T09:22:38.6858692Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6858828Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6858832Z 2025-12-04T09:22:38.6858991Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6859244Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6859247Z 2025-12-04T09:22:38.6859324Z Expected 0 but got 123. 2025-12-04T09:22:38.6859399Z Absolute difference: 123 2025-12-04T09:22:38.6859466Z Relative difference: inf 2025-12-04T09:22:38.6859469Z 2025-12-04T09:22:38.6859589Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6859732Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6859736Z 2025-12-04T09:22:38.6859882Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6860139Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:22:38.6860192Z 2025-12-04T09:22:38.6860269Z Expected 0 but got 123. 2025-12-04T09:22:38.6860339Z Absolute difference: 123 2025-12-04T09:22:38.6860448Z Relative difference: inf 2025-12-04T09:22:38.6860452Z 2025-12-04T09:22:38.6860583Z To execute this test, run the following from the base repo dir: 2025-12-04T09:22:38.6860794Z python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:22:38.6860804Z 2025-12-04T09:22:38.6860965Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:22:38.6861113Z ======== 51 failed, 149 passed, 100 skipped, 1400 deselected in 14.16s ========= 2025-12-04T09:22:38.6861117Z 2025-12-04T09:22:38.6861355Z FINISHED PRINTING LOG FILE of test_cuda 1/1 (test/test-reports/test_cuda_1.1_1d7d8eb7c1053f49_.log) 2025-12-04T09:22:38.6861359Z 2025-12-04T09:22:38.6861521Z Finished test_cuda 1/1 ... [2025-12-04 09:22:38.627779][1794.587323305], took 0.32min 2025-12-04T09:22:38.6861905Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda/test_cuda-8b00a0d732cb0448.xml 2025-12-04T09:22:39.0956159Z Uploading logs for 57116084869 to S3 2025-12-04T09:22:39.1357646Z Uploading artifacts took 0.40 seconds 2025-12-04T09:22:39.1358096Z test_cuda 1/1 failed! 2025-12-04T09:22:42.3042210Z Running higher_order_ops/test_local_map 1/1 ... [2025-12-04 09:22:42.303778][1798.263320783] 2025-12-04T09:22:42.3042720Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:22:42.3044768Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'higher_order_ops/test_local_map.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:22:42.304202] 2025-12-04T09:22:45.9657348Z 2025-12-04T09:22:45.9658370Z higher_order_ops/test_local_map 1/1 was successful, full logs can be found in artifacts with path test/test-reports/higher_order_ops.test_local_map_1.1_cb27141dc638196e_.log 2025-12-04T09:22:45.9659300Z Running 0 items in this shard: 2025-12-04T09:22:45.9659485Z 2025-12-04T09:22:45.9659802Z Finished higher_order_ops/test_local_map 1/1 ... [2025-12-04 09:22:45.965612][1801.925159027], took 0.06min 2025-12-04T09:22:45.9738267Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/higher_order_ops.test_local_map/higher_order_ops.test_local_map-218c693b3a0f40d8.xml 2025-12-04T09:22:49.4731572Z Running test_dataloader 1/1 ... [2025-12-04 09:22:49.472677][1805.432221001] 2025-12-04T09:22:49.4732307Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:22:49.4733952Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_dataloader.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:22:49.473063] 2025-12-04T09:22:53.6007674Z 2025-12-04T09:22:53.6008543Z test_dataloader 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_dataloader_1.1_a3d69a47396d20d9_.log 2025-12-04T09:22:53.6009257Z Running 0 items in this shard: 2025-12-04T09:22:53.6009430Z 2025-12-04T09:22:53.6009713Z Finished test_dataloader 1/1 ... [2025-12-04 09:22:53.600666][1809.560212048], took 0.07min 2025-12-04T09:22:53.6090613Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_dataloader/test_dataloader-ce202fcb501e6304.xml 2025-12-04T09:22:55.9365293Z 2025-12-04T09:22:55.9366532Z test_transformers 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_transformers_1.1_6b85ae180d2e1277_.log 2025-12-04T09:22:55.9367378Z Running 0 items in this shard: 2025-12-04T09:22:55.9367556Z 2025-12-04T09:22:55.9367811Z Finished test_transformers 1/1 ... [2025-12-04 09:22:55.936380][1811.895925654], took 0.39min 2025-12-04T09:22:55.9447005Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_transformers/test_transformers-d83a44de2640a948.xml 2025-12-04T09:22:57.2607279Z Running test_decomp 6/22 ... [2025-12-04 09:22:57.260225][1813.219767816] 2025-12-04T09:22:57.2608040Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:22:57.2609250Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_decomp.py', '-m', 'not serial', '--shard-id=6', '--num-shards=22', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:22:57.260641] 2025-12-04T09:22:59.4805610Z Running test_decomp 8/22 ... [2025-12-04 09:22:59.480057][1815.439599812] 2025-12-04T09:22:59.4806216Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:22:59.4809947Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_decomp.py', '-m', 'not serial', '--shard-id=8', '--num-shards=22', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:22:59.480688] 2025-12-04T09:23:14.7592929Z 2025-12-04T09:23:14.7593810Z test_decomp 6/22 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_6.22_9c63b20df2fb93f2_.log 2025-12-04T09:23:14.7594475Z Running 0 items in this shard: 2025-12-04T09:23:14.7594652Z 2025-12-04T09:23:14.7594880Z Finished test_decomp 6/22 ... [2025-12-04 09:23:14.759179][1830.71872513], took 0.29min 2025-12-04T09:23:14.7685258Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-866538190db3f9f9.xml 2025-12-04T09:23:17.0603191Z 2025-12-04T09:23:17.0603894Z test_decomp 8/22 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_8.22_fb867f77ecfd23a1_.log 2025-12-04T09:23:17.0604848Z Running 0 items in this shard: 2025-12-04T09:23:17.0605040Z 2025-12-04T09:23:17.0605274Z Finished test_decomp 8/22 ... [2025-12-04 09:23:17.060171][1833.01971714], took 0.29min 2025-12-04T09:23:17.0706573Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-5ef04743f41caed7.xml 2025-12-04T09:23:18.4003520Z Running test_decomp 12/22 ... [2025-12-04 09:23:18.399859][1834.359403334] 2025-12-04T09:23:18.4003981Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:23:18.4006743Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_decomp.py', '-m', 'not serial', '--shard-id=12', '--num-shards=22', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:23:18.400323] 2025-12-04T09:23:20.6170478Z Running test_decomp 18/22 ... [2025-12-04 09:23:20.616633][1836.576177202] 2025-12-04T09:23:20.6170971Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:23:20.6173722Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_decomp.py', '-m', 'not serial', '--shard-id=18', '--num-shards=22', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:23:20.617022] 2025-12-04T09:23:35.9078878Z 2025-12-04T09:23:35.9079973Z test_decomp 12/22 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_12.22_5f45f685300755c6_.log 2025-12-04T09:23:35.9080894Z Running 0 items in this shard: 2025-12-04T09:23:35.9081039Z 2025-12-04T09:23:35.9081236Z Finished test_decomp 12/22 ... [2025-12-04 09:23:35.907770][1851.867312128], took 0.29min 2025-12-04T09:23:35.9174097Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-26241f1d9fe6a84a.xml 2025-12-04T09:23:39.4367367Z Running test_decomp 20/22 ... [2025-12-04 09:23:39.436272][1855.395815849] 2025-12-04T09:23:39.4368352Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:23:39.4370897Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_decomp.py', '-m', 'not serial', '--shard-id=20', '--num-shards=22', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:23:39.436748] 2025-12-04T09:23:56.7504214Z 2025-12-04T09:23:56.7505576Z test_decomp 20/22 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_20.22_f3a68dcf10860f63_.log 2025-12-04T09:23:56.7506515Z Running 0 items in this shard: 2025-12-04T09:23:56.7506779Z 2025-12-04T09:23:56.7507113Z Finished test_decomp 20/22 ... [2025-12-04 09:23:56.750233][1872.709776975], took 0.29min 2025-12-04T09:23:56.7600009Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-3733592a1d6e9f3d.xml 2025-12-04T09:24:00.2364717Z Running test_ops 2/9 ... [2025-12-04 09:24:00.235949][1876.195492687] 2025-12-04T09:24:00.2365325Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:24:00.2366952Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops.py', '-m', 'not serial', '--shard-id=2', '--num-shards=9', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:24:00.236358] 2025-12-04T09:24:54.8009482Z 2025-12-04T09:24:54.8010483Z test_ops 2/9 was successful, full logs can be found in artifacts with path test/test-reports/test_ops_2.9_146c8473bb4b76bc_.log 2025-12-04T09:24:54.8011792Z Running 0 items in this shard: 2025-12-04T09:24:54.8012107Z 2025-12-04T09:24:54.8012553Z Finished test_ops 2/9 ... [2025-12-04 09:24:54.800792][1930.760334692], took 0.91min 2025-12-04T09:24:54.8112017Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops/test_ops-5e4cfed161e0a0a1.xml 2025-12-04T09:24:58.3132509Z Running test_ops 8/9 ... [2025-12-04 09:24:58.312786][1934.272330085] 2025-12-04T09:24:58.3132933Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:24:58.3135798Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops.py', '-m', 'not serial', '--shard-id=8', '--num-shards=9', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:24:58.313220] 2025-12-04T09:26:15.7431268Z 2025-12-04T09:26:15.7431910Z PRINTING LOG FILE of test_ops 8/9 (test/test-reports/test_ops_8.9_047a11a4b058dbbd_.log) 2025-12-04T09:26:15.7432607Z Test results will be stored in test-reports/python-pytest/test_ops/test_ops-3bbe4a7adb52bac4.xml 2025-12-04T09:26:15.7433160Z ============================= test session starts ============================== 2025-12-04T09:26:15.7433716Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:26:15.7434223Z cachedir: .pytest_cache 2025-12-04T09:26:15.7434760Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:26:15.7435359Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:26:15.7435649Z configfile: pytest.ini 2025-12-04T09:26:15.7436110Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:26:15.7436619Z collecting ... collected 33666 items 2025-12-04T09:26:15.7436874Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:26:15.7474116Z Running 150 items in this shard: test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 2025-12-04T09:26:15.7511171Z 2025-12-04T09:26:15.7511507Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0003s] (output is non-deterministic) [ 0%] 2025-12-04T09:26:15.7512289Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.1292s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 1%] 2025-12-04T09:26:15.7513047Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 FAILED [0.3288s] [ 2%] 2025-12-04T09:26:15.7513720Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7514412Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7515098Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7515785Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7516458Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7517138Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7517821Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7518502Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7519172Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7519859Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7520576Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7521260Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0003s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7521955Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7522639Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7523446Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7524327Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7525010Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7525684Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7526377Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7527058Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0003s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7527740Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7528418Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7529102Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7529781Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7530464Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7531335Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7532054Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7532744Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0003s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7533537Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7534214Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7534888Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7535565Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7536245Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7536919Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7537599Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7538278Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0003s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7538958Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7539636Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7540310Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7541069Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7541850Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7542531Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7543201Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7543879Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0003s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7544559Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7545252Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7545932Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7546612Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7547291Z test_ops.py::TestCommonCUDA::test_compare_cpu__refs_empty_strided_cuda_float32 SKIPPED [0.0002s] (output is non-deterministic) [ 2%] 2025-12-04T09:26:15.7548063Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0021s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7548900Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0020s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7549748Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7550587Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7551417Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7552250Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0020s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7553073Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7553908Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7554741Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7555576Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0020s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7556403Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7557231Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7558112Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7559007Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0020s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7559901Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7560732Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7561566Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7562397Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0019s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7563236Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7564078Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0017s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7564902Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7565737Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0019s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7566571Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7567407Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7568259Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7569098Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0024s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7569926Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7570758Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0017s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7571592Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7572419Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0019s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7573365Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7574194Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7575026Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7576066Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0019s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7577172Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0017s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7578044Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7578873Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0017s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7579700Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0019s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7580530Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0019s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7581355Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7582190Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7583019Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0019s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7583853Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7584691Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7585525Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7586364Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0019s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7587198Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7588028Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0018s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7588851Z test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 SKIPPED [0.0017s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 2%] 2025-12-04T09:26:15.7589605Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4976s] [ 2%] 2025-12-04T09:26:15.7590268Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4652s] [ 2%] 2025-12-04T09:26:15.7590916Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4659s] [ 2%] 2025-12-04T09:26:15.7591556Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4568s] [ 2%] 2025-12-04T09:26:15.7592200Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4589s] [ 2%] 2025-12-04T09:26:15.7592843Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4631s] [ 2%] 2025-12-04T09:26:15.7594375Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4573s] [ 2%] 2025-12-04T09:26:15.7595013Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4592s] [ 2%] 2025-12-04T09:26:15.7595769Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4825s] [ 2%] 2025-12-04T09:26:15.7596415Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4557s] [ 2%] 2025-12-04T09:26:15.7597057Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4540s] [ 2%] 2025-12-04T09:26:15.7597701Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4502s] [ 2%] 2025-12-04T09:26:15.7598350Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4509s] [ 2%] 2025-12-04T09:26:15.7599014Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4503s] [ 2%] 2025-12-04T09:26:15.7599689Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4489s] [ 2%] 2025-12-04T09:26:15.7600339Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4515s] [ 2%] 2025-12-04T09:26:15.7600987Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4672s] [ 2%] 2025-12-04T09:26:15.7601637Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4806s] [ 2%] 2025-12-04T09:26:15.7602292Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4502s] [ 2%] 2025-12-04T09:26:15.7602933Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4574s] [ 2%] 2025-12-04T09:26:15.7603584Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4521s] [ 2%] 2025-12-04T09:26:15.7604235Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4530s] [ 2%] 2025-12-04T09:26:15.7605183Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4588s] [ 2%] 2025-12-04T09:26:15.7605840Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4521s] [ 2%] 2025-12-04T09:26:15.7606487Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4527s] [ 2%] 2025-12-04T09:26:15.7607133Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4786s] [ 2%] 2025-12-04T09:26:15.7607785Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4498s] [ 2%] 2025-12-04T09:26:15.7608421Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4512s] [ 2%] 2025-12-04T09:26:15.7609069Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4499s] [ 2%] 2025-12-04T09:26:15.7609714Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4504s] [ 2%] 2025-12-04T09:26:15.7610370Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4510s] [ 2%] 2025-12-04T09:26:15.7611006Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4511s] [ 2%] 2025-12-04T09:26:15.7611649Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4546s] [ 2%] 2025-12-04T09:26:15.7612412Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4640s] [ 2%] 2025-12-04T09:26:15.7613317Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4773s] [ 2%] 2025-12-04T09:26:15.7613976Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4473s] [ 2%] 2025-12-04T09:26:15.7614620Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4571s] [ 2%] 2025-12-04T09:26:15.7615262Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4546s] [ 2%] 2025-12-04T09:26:15.7615916Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4512s] [ 2%] 2025-12-04T09:26:15.7616559Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4573s] [ 2%] 2025-12-04T09:26:15.7617197Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4507s] [ 2%] 2025-12-04T09:26:15.7617844Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4524s] [ 2%] 2025-12-04T09:26:15.7618490Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4794s] [ 2%] 2025-12-04T09:26:15.7619138Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4538s] [ 2%] 2025-12-04T09:26:15.7619776Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4499s] [ 2%] 2025-12-04T09:26:15.7620427Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4505s] [ 2%] 2025-12-04T09:26:15.7621077Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4501s] [ 2%] 2025-12-04T09:26:15.7621721Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4510s] [ 2%] 2025-12-04T09:26:15.7622365Z test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 PASSED [0.4507s] [ 2%] 2025-12-04T09:26:15.7622745Z 2025-12-04T09:26:15.7622844Z =================================== FAILURES =================================== 2025-12-04T09:26:15.7623246Z _ TestFakeTensorCUDA.test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 _ 2025-12-04T09:26:15.7623627Z Traceback (most recent call last): 2025-12-04T09:26:15.7624128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_device_type.py", line 1151, in test_wrapper 2025-12-04T09:26:15.7624622Z return test(*args, **kwargs) 2025-12-04T09:26:15.7625093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_device_type.py", line 1435, in only_fn 2025-12-04T09:26:15.7625559Z return fn(slf, *args, **kwargs) 2025-12-04T09:26:15.7625953Z File "/var/lib/jenkins/workspace/test/test_ops.py", line 2983, in test_fake_crossref_backward_amp 2025-12-04T09:26:15.7626446Z self._test_fake_crossref_helper(device, dtype, op, torch.cuda.amp.autocast) 2025-12-04T09:26:15.7626922Z File "/var/lib/jenkins/workspace/test/test_ops.py", line 2957, in _test_fake_crossref_helper 2025-12-04T09:26:15.7627313Z composite_compliance.compute_expected_grads( 2025-12-04T09:26:15.7627864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/composite_compliance.py", line 431, in compute_expected_grads 2025-12-04T09:26:15.7628417Z results = gradcheck_wrapper(op, *args, **kwargs) 2025-12-04T09:26:15.7628901Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/opinfo/core.py", line 831, in 2025-12-04T09:26:15.7629523Z gradcheck_wrapper: Callable = lambda op, *args, **kwargs: op(*args, **kwargs) 2025-12-04T09:26:15.7630089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_utils.py", line 275, in __torch_dispatch__ 2025-12-04T09:26:15.7630671Z _check_alias_info( 2025-12-04T09:26:15.7631076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_utils.py", line 57, in _check_alias_info 2025-12-04T09:26:15.7631535Z raise MetadataMismatchError( 2025-12-04T09:26:15.7632189Z torch._subclasses.fake_tensor.MetadataMismatchError: When comparing the output of aten.view.default on FakeTensor and concrete Tensors, found mismatch in outputs_alias_inputs check False != True 2025-12-04T09:26:15.7632770Z 2025-12-04T09:26:15.7632931Z The above exception was the direct cause of the following exception: 2025-12-04T09:26:15.7633166Z 2025-12-04T09:26:15.7633245Z Traceback (most recent call last): 2025-12-04T09:26:15.7633696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:26:15.7634146Z method(*args, **kwargs) 2025-12-04T09:26:15.7634572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:26:15.7635007Z method(*args, **kwargs) 2025-12-04T09:26:15.7635478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_device_type.py", line 428, in instantiated_test 2025-12-04T09:26:15.7635983Z result = test(self, **param_kwargs) 2025-12-04T09:26:15.7636424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 1736, in wrapper 2025-12-04T09:26:15.7636866Z fn(*args, **kwargs) 2025-12-04T09:26:15.7637312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_device_type.py", line 1163, in test_wrapper 2025-12-04T09:26:15.7637804Z raise e_tracked from e 2025-12-04T09:26:15.7638295Z Exception: When comparing the output of aten.view.default on FakeTensor and concrete Tensors, found mismatch in outputs_alias_inputs check False != True 2025-12-04T09:26:15.7638754Z 2025-12-04T09:26:15.7639540Z Caused by sample input at index 5: SampleInput(input=Tensor[size=(0, 8), device="cuda:0", dtype=torch.float32], args=TensorList[Tensor[size=(0, 8), device="cuda:0", dtype=torch.float32], Tensor[size=(8, 8, 8), device="cuda:0", dtype=torch.float32], Tensor[size=(8,), device="cuda:0", dtype=torch.float32]], kwargs={}, broadcasts_input=False, name='') 2025-12-04T09:26:15.7640410Z 2025-12-04T09:26:15.7640541Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:15.7641120Z PYTORCH_OPINFO_SAMPLE_INPUT_INDEX=5 python test/test_ops.py TestFakeTensorCUDA.test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 2025-12-04T09:26:15.7641566Z 2025-12-04T09:26:15.7641739Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:15.7642119Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:15.7642785Z /var/lib/jenkins/workspace/test/test_ops.py:2954: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. 2025-12-04T09:26:15.7643360Z context(), 2025-12-04T09:26:15.7643891Z /var/lib/jenkins/workspace/test/test_ops.py:2954: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. 2025-12-04T09:26:15.7644458Z context(), 2025-12-04T09:26:15.7644974Z /var/lib/jenkins/workspace/test/test_ops.py:2954: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. 2025-12-04T09:26:15.7645547Z context(), 2025-12-04T09:26:15.7646062Z /var/lib/jenkins/workspace/test/test_ops.py:2954: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. 2025-12-04T09:26:15.7646698Z context(), 2025-12-04T09:26:15.7647222Z /var/lib/jenkins/workspace/test/test_ops.py:2954: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. 2025-12-04T09:26:15.7647889Z context(), 2025-12-04T09:26:15.7648400Z /var/lib/jenkins/workspace/test/test_ops.py:2954: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. 2025-12-04T09:26:15.7648968Z context(), 2025-12-04T09:26:15.7649385Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops/test_ops-3bbe4a7adb52bac4.xml - 2025-12-04T09:26:15.7649903Z =========================== short test summary info ============================ 2025-12-04T09:26:15.7650790Z FAILED [0.3288s] test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 - Exception: When comparing the output of aten.view.default on FakeTensor and concrete Tensors, found mismatch in outputs_alias_inputs check False != True 2025-12-04T09:26:15.7651567Z 2025-12-04T09:26:15.7652361Z Caused by sample input at index 5: SampleInput(input=Tensor[size=(0, 8), device="cuda:0", dtype=torch.float32], args=TensorList[Tensor[size=(0, 8), device="cuda:0", dtype=torch.float32], Tensor[size=(8, 8, 8), device="cuda:0", dtype=torch.float32], Tensor[size=(8,), device="cuda:0", dtype=torch.float32]], kwargs={}, broadcasts_input=False, name='') 2025-12-04T09:26:15.7653320Z 2025-12-04T09:26:15.7653459Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:15.7654039Z PYTORCH_OPINFO_SAMPLE_INPUT_INDEX=5 python test/test_ops.py TestFakeTensorCUDA.test_fake_crossref_backward_amp_nn_functional_bilinear_cuda_float32 2025-12-04T09:26:15.7654486Z 2025-12-04T09:26:15.7654658Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:15.7655026Z ============= 1 failed, 49 passed, 100 skipped in 64.07s (0:01:04) ============= 2025-12-04T09:26:15.7655240Z 2025-12-04T09:26:15.7655457Z FINISHED PRINTING LOG FILE of test_ops 8/9 (test/test-reports/test_ops_8.9_047a11a4b058dbbd_.log) 2025-12-04T09:26:15.7655763Z 2025-12-04T09:26:15.7655919Z Finished test_ops 8/9 ... [2025-12-04 09:26:15.743047][2011.702593099], took 1.29min 2025-12-04T09:26:15.7656548Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops/test_ops-3bbe4a7adb52bac4.xml 2025-12-04T09:26:16.2539097Z Uploading logs for 57116084869 to S3 2025-12-04T09:26:16.2936548Z Uploading artifacts took 0.39 seconds 2025-12-04T09:26:16.2936897Z test_ops 8/9 failed! 2025-12-04T09:26:19.2558196Z Running functorch/test_dims 1/1 ... [2025-12-04 09:26:19.255355][2015.214899305] 2025-12-04T09:26:19.2558683Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:26:19.2560223Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_dims.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:26:19.255752] 2025-12-04T09:26:48.7531635Z 2025-12-04T09:26:48.7532623Z PRINTING LOG FILE of functorch/test_dims 1/1 (test/test-reports/functorch.test_dims_1.1_99eaaa239e58b576_.log) 2025-12-04T09:26:48.7533864Z Test results will be stored in test-reports/python-pytest/functorch.test_dims/functorch.test_dims-09de3759aef74478.xml 2025-12-04T09:26:48.7534913Z ============================= test session starts ============================== 2025-12-04T09:26:48.7535816Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:26:48.7536668Z cachedir: .pytest_cache 2025-12-04T09:26:48.7537465Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:26:48.7539557Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:26:48.7539911Z configfile: pytest.ini 2025-12-04T09:26:48.7540674Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:26:48.7541639Z collecting ... collected 68 items 2025-12-04T09:26:48.7542335Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:26:48.7553408Z Running 50 items in this shard: test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda, test/functorch/test_dims.py::TestMin::test_attn_cuda 2025-12-04T09:26:48.7564847Z 2025-12-04T09:26:48.7565177Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.7465s] [ 2%] 2025-12-04T09:26:48.7565834Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7566437Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5085s] [ 2%] 2025-12-04T09:26:48.7567511Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7568123Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5312s] [ 2%] 2025-12-04T09:26:48.7568713Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7569442Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5263s] [ 2%] 2025-12-04T09:26:48.7570061Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7570864Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5527s] [ 2%] 2025-12-04T09:26:48.7571536Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7572376Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5068s] [ 2%] 2025-12-04T09:26:48.7573351Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7574044Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5153s] [ 2%] 2025-12-04T09:26:48.7574625Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7575105Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5144s] [ 2%] 2025-12-04T09:26:48.7575500Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7575904Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.4989s] [ 2%] 2025-12-04T09:26:48.7576305Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7576702Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5142s] [ 2%] 2025-12-04T09:26:48.7577087Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7577481Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5081s] [ 2%] 2025-12-04T09:26:48.7577866Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7578251Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5169s] [ 2%] 2025-12-04T09:26:48.7578642Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7579025Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5431s] [ 2%] 2025-12-04T09:26:48.7579410Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7579798Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5753s] [ 2%] 2025-12-04T09:26:48.7580177Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7580562Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5083s] [ 2%] 2025-12-04T09:26:48.7580952Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7581335Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.4991s] [ 2%] 2025-12-04T09:26:48.7581722Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7582129Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.4964s] [ 2%] 2025-12-04T09:26:48.7582531Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7582913Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5279s] [ 2%] 2025-12-04T09:26:48.7583301Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7583692Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5092s] [ 2%] 2025-12-04T09:26:48.7584073Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7584461Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5357s] [ 2%] 2025-12-04T09:26:48.7584846Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7585233Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5000s] [ 2%] 2025-12-04T09:26:48.7585612Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7585996Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5007s] [ 2%] 2025-12-04T09:26:48.7586383Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7586768Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.4994s] [ 2%] 2025-12-04T09:26:48.7587147Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7587607Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.4976s] [ 2%] 2025-12-04T09:26:48.7587992Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7588477Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.4939s] [ 2%] 2025-12-04T09:26:48.7588864Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7589250Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.4925s] [ 2%] 2025-12-04T09:26:48.7589637Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7590018Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.4966s] [ 2%] 2025-12-04T09:26:48.7590399Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7590782Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5233s] [ 2%] 2025-12-04T09:26:48.7591161Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7591546Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5393s] [ 2%] 2025-12-04T09:26:48.7591935Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7592326Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.4935s] [ 2%] 2025-12-04T09:26:48.7592708Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7593097Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5174s] [ 2%] 2025-12-04T09:26:48.7593483Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7593868Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.4824s] [ 2%] 2025-12-04T09:26:48.7594254Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7594645Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5143s] [ 2%] 2025-12-04T09:26:48.7595031Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7595433Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5275s] [ 2%] 2025-12-04T09:26:48.7595835Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7596225Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5215s] [ 2%] 2025-12-04T09:26:48.7596634Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7597028Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5562s] [ 2%] 2025-12-04T09:26:48.7597420Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7597813Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.4977s] [ 2%] 2025-12-04T09:26:48.7598198Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7598594Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5006s] [ 2%] 2025-12-04T09:26:48.7598986Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7599380Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.4985s] [ 2%] 2025-12-04T09:26:48.7599774Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7600171Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.4943s] [ 2%] 2025-12-04T09:26:48.7600562Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7600958Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5068s] [ 2%] 2025-12-04T09:26:48.7601354Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7601744Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5132s] [ 2%] 2025-12-04T09:26:48.7602182Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7602565Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5059s] [ 2%] 2025-12-04T09:26:48.7602998Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7603452Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5334s] [ 2%] 2025-12-04T09:26:48.7603843Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7604226Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.4988s] [ 2%] 2025-12-04T09:26:48.7604871Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7605271Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.4973s] [ 2%] 2025-12-04T09:26:48.7605656Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7606041Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5081s] [ 2%] 2025-12-04T09:26:48.7606435Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7606826Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.4997s] [ 2%] 2025-12-04T09:26:48.7607221Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7607616Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5044s] [ 2%] 2025-12-04T09:26:48.7608005Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0001s] [ 2%] 2025-12-04T09:26:48.7608387Z functorch/test_dims.py::TestMin::test_attn_cuda FAILED [0.5144s] [ 2%] 2025-12-04T09:26:48.7608775Z functorch/test_dims.py::TestMin::test_attn_cuda ERROR [0.0005s] [ 2%] 2025-12-04T09:26:48.7609011Z 2025-12-04T09:26:48.7609102Z ==================================== ERRORS ==================================== 2025-12-04T09:26:48.7609424Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7609729Z Traceback (most recent call last): 2025-12-04T09:26:48.7610092Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7610467Z import refcycle 2025-12-04T09:26:48.7610678Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7611003Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7611688Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7612339Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7612973Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7613720Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7614054Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7614362Z Traceback (most recent call last): 2025-12-04T09:26:48.7614703Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7615059Z import refcycle 2025-12-04T09:26:48.7615268Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7615570Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7616227Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7616864Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7617497Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7618236Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7618552Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7619315Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7620019Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7620353Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7620658Z Traceback (most recent call last): 2025-12-04T09:26:48.7621005Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7621351Z import refcycle 2025-12-04T09:26:48.7621558Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7621871Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7622533Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7623161Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7623791Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7624419Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7624720Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7625369Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7625998Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7626323Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7626614Z Traceback (most recent call last): 2025-12-04T09:26:48.7626954Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7627300Z import refcycle 2025-12-04T09:26:48.7627501Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7627799Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7628445Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7629071Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7629690Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7630311Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7630615Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7631260Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7631879Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7632218Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7632512Z Traceback (most recent call last): 2025-12-04T09:26:48.7632850Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7633188Z import refcycle 2025-12-04T09:26:48.7633386Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7633683Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7634326Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7635010Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7635633Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7636382Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7636687Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7637333Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7637961Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7638281Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7638577Z Traceback (most recent call last): 2025-12-04T09:26:48.7638929Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7639283Z import refcycle 2025-12-04T09:26:48.7639476Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7639777Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7640431Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7641063Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7641685Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7642306Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7642610Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7643256Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7643893Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7644229Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7644531Z Traceback (most recent call last): 2025-12-04T09:26:48.7644872Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7645220Z import refcycle 2025-12-04T09:26:48.7645420Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7645713Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7646376Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7647007Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7647640Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7648260Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7648567Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7649217Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7649843Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7650160Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7650452Z Traceback (most recent call last): 2025-12-04T09:26:48.7650786Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7651192Z import refcycle 2025-12-04T09:26:48.7651386Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7651683Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7652440Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7653066Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7653809Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7654432Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7654737Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7655384Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7656012Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7656336Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7656634Z Traceback (most recent call last): 2025-12-04T09:26:48.7656968Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7657312Z import refcycle 2025-12-04T09:26:48.7657509Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7657798Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7658451Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7659089Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7659715Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7660331Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7660636Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7661289Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7661920Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7662236Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7662532Z Traceback (most recent call last): 2025-12-04T09:26:48.7662864Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7663202Z import refcycle 2025-12-04T09:26:48.7663396Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7663696Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7664343Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7664976Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7665599Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7666225Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7666526Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7667169Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7667803Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7668179Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7668461Z Traceback (most recent call last): 2025-12-04T09:26:48.7668831Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7669174Z import refcycle 2025-12-04T09:26:48.7669427Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7669725Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7670374Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7670995Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7671609Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7672232Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7672533Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7673188Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7673812Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7674135Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7674425Z Traceback (most recent call last): 2025-12-04T09:26:48.7674751Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7675093Z import refcycle 2025-12-04T09:26:48.7675288Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7675585Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7676239Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7676869Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7677511Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7678264Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7678586Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7679306Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7680358Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7680679Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7680980Z Traceback (most recent call last): 2025-12-04T09:26:48.7681315Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7681659Z import refcycle 2025-12-04T09:26:48.7681850Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7682153Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7682807Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7683439Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7684056Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7684675Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7685042Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7685690Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7686368Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7686786Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7687102Z Traceback (most recent call last): 2025-12-04T09:26:48.7687438Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7687789Z import refcycle 2025-12-04T09:26:48.7687991Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7688289Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7688948Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7689587Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7690214Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7690841Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7691146Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7691797Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7692439Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7692757Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7693051Z Traceback (most recent call last): 2025-12-04T09:26:48.7693527Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7693866Z import refcycle 2025-12-04T09:26:48.7694069Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7694372Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7695028Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7695653Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7696283Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7696915Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7697218Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7697867Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7698506Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7698831Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7699125Z Traceback (most recent call last): 2025-12-04T09:26:48.7699464Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7699809Z import refcycle 2025-12-04T09:26:48.7700011Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7700306Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7700958Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7701590Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7702290Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7702944Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7703322Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7703980Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7704889Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7705217Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7705511Z Traceback (most recent call last): 2025-12-04T09:26:48.7705854Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7706198Z import refcycle 2025-12-04T09:26:48.7706402Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7706706Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7707386Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7708017Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7708645Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7709268Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7709565Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7710214Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7710843Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7711164Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7711454Z Traceback (most recent call last): 2025-12-04T09:26:48.7711789Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7712134Z import refcycle 2025-12-04T09:26:48.7712321Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7712617Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7713267Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7713899Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7714524Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7715146Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7715450Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7716101Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7716730Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7717050Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7717343Z Traceback (most recent call last): 2025-12-04T09:26:48.7717669Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7718010Z import refcycle 2025-12-04T09:26:48.7718201Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7718595Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7719242Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7720032Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7720673Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7721304Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7721612Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7722265Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7722910Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7723239Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7723543Z Traceback (most recent call last): 2025-12-04T09:26:48.7723889Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7724240Z import refcycle 2025-12-04T09:26:48.7724436Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7724740Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7725401Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7726027Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7726661Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7727290Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7727603Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7728258Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7728887Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7729210Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7729514Z Traceback (most recent call last): 2025-12-04T09:26:48.7729848Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7730191Z import refcycle 2025-12-04T09:26:48.7730390Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7730685Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7731344Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7731973Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7732604Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7746409Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7746866Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7747634Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7748302Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7748645Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7749062Z Traceback (most recent call last): 2025-12-04T09:26:48.7749409Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7750395Z import refcycle 2025-12-04T09:26:48.7750621Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7751009Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7751680Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7752321Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7752952Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7753571Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7753887Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7754557Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7755221Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7755547Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7755855Z Traceback (most recent call last): 2025-12-04T09:26:48.7756210Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7756559Z import refcycle 2025-12-04T09:26:48.7756771Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7757075Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7757746Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7758386Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7759020Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7759654Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7759968Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7760625Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7761270Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7761608Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7761902Z Traceback (most recent call last): 2025-12-04T09:26:48.7762252Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7762598Z import refcycle 2025-12-04T09:26:48.7762803Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7763105Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7763764Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7764403Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7765038Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7765663Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7765973Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7766666Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7767319Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7767709Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7768005Z Traceback (most recent call last): 2025-12-04T09:26:48.7768345Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7768677Z import refcycle 2025-12-04T09:26:48.7768881Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7769179Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7769821Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7770451Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7771075Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7771705Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7772006Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7772657Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7773422Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7773747Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7774032Z Traceback (most recent call last): 2025-12-04T09:26:48.7774369Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7774715Z import refcycle 2025-12-04T09:26:48.7774903Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7775202Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7775858Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7776479Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7777088Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7777703Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7778007Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7778648Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7779273Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7779595Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7779892Z Traceback (most recent call last): 2025-12-04T09:26:48.7780238Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7780580Z import refcycle 2025-12-04T09:26:48.7780776Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7781066Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7781712Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7782337Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7782954Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7783614Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7783953Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7784655Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7785280Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7785588Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7785876Z Traceback (most recent call last): 2025-12-04T09:26:48.7786201Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7786530Z import refcycle 2025-12-04T09:26:48.7786731Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7787031Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7787673Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7788296Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7788915Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7789531Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7789831Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7790472Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7791099Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7791425Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7791711Z Traceback (most recent call last): 2025-12-04T09:26:48.7792049Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7792395Z import refcycle 2025-12-04T09:26:48.7792589Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7792879Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7793520Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7794148Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7794760Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7795383Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7795685Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7796331Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7796956Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7797279Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7797568Z Traceback (most recent call last): 2025-12-04T09:26:48.7797906Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7798242Z import refcycle 2025-12-04T09:26:48.7798434Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7798741Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7799430Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7800090Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7800791Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7801408Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7801703Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7802493Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7803120Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7803443Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7803734Z Traceback (most recent call last): 2025-12-04T09:26:48.7804067Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7804412Z import refcycle 2025-12-04T09:26:48.7804948Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7805263Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7805912Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7806533Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7807148Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7807764Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7808068Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7808707Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7809325Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7809644Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7809934Z Traceback (most recent call last): 2025-12-04T09:26:48.7810260Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7810601Z import refcycle 2025-12-04T09:26:48.7810793Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7811088Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7811727Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7812351Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7812972Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7813692Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7813991Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7814629Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7815251Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7815564Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7815856Z Traceback (most recent call last): 2025-12-04T09:26:48.7816339Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7816679Z import refcycle 2025-12-04T09:26:48.7816865Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7817219Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7817963Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7818584Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7819200Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7819818Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7820140Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7820777Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7821402Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7821723Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7822013Z Traceback (most recent call last): 2025-12-04T09:26:48.7822343Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7822686Z import refcycle 2025-12-04T09:26:48.7822890Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7823182Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7823833Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7824464Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7825087Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7825702Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7826013Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7826656Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7827288Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7827606Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7827894Z Traceback (most recent call last): 2025-12-04T09:26:48.7828224Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7828558Z import refcycle 2025-12-04T09:26:48.7828751Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7829041Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7829692Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7830306Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7830923Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7831538Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7831828Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7832464Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7833140Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7833456Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7833782Z Traceback (most recent call last): 2025-12-04T09:26:48.7834177Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7834519Z import refcycle 2025-12-04T09:26:48.7834704Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7834999Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7835644Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7836268Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7836886Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7837511Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7837814Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7838459Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7839074Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7839392Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7839679Z Traceback (most recent call last): 2025-12-04T09:26:48.7840011Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7840344Z import refcycle 2025-12-04T09:26:48.7840557Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7840850Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7841489Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7842121Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7842738Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7843356Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7843650Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7844291Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7844915Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7845232Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7845517Z Traceback (most recent call last): 2025-12-04T09:26:48.7845852Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7846198Z import refcycle 2025-12-04T09:26:48.7846386Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7846680Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7847331Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7847955Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7848563Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7849228Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7849529Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7850267Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7850887Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7851205Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7851494Z Traceback (most recent call last): 2025-12-04T09:26:48.7851831Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7852172Z import refcycle 2025-12-04T09:26:48.7852361Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7852654Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7853401Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7854030Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7854659Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7855275Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7855569Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7856213Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7856841Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7857153Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7857442Z Traceback (most recent call last): 2025-12-04T09:26:48.7857778Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7858121Z import refcycle 2025-12-04T09:26:48.7858310Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7858602Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7859253Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7859874Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7860484Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7861097Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7861401Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7862033Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7862663Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7862974Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7863264Z Traceback (most recent call last): 2025-12-04T09:26:48.7863600Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7863942Z import refcycle 2025-12-04T09:26:48.7864133Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7864422Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7865065Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7865744Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7866477Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7867102Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7867415Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7868056Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7868677Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7868990Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7869282Z Traceback (most recent call last): 2025-12-04T09:26:48.7869615Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7869947Z import refcycle 2025-12-04T09:26:48.7870144Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7870442Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7871088Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7871703Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7872323Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7872941Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7873245Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7873893Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7874516Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7874839Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7875124Z Traceback (most recent call last): 2025-12-04T09:26:48.7875459Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7875800Z import refcycle 2025-12-04T09:26:48.7875991Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7876281Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7876933Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7877560Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7878175Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7878790Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7879089Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7879737Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7880351Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7880672Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7880958Z Traceback (most recent call last): 2025-12-04T09:26:48.7881290Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7881678Z import refcycle 2025-12-04T09:26:48.7881867Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7882156Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7882897Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7883526Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7884139Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7884753Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7885041Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7885686Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7886322Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7886638Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7886927Z Traceback (most recent call last): 2025-12-04T09:26:48.7887256Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7887593Z import refcycle 2025-12-04T09:26:48.7887777Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7888062Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7888701Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7889317Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7889929Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7890541Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7890849Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7891491Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7892110Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7892427Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7892714Z Traceback (most recent call last): 2025-12-04T09:26:48.7893042Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7893500Z import refcycle 2025-12-04T09:26:48.7893695Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7893984Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7894618Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7895243Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7895863Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7896481Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7896781Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7897418Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7898088Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7898397Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7898727Z Traceback (most recent call last): 2025-12-04T09:26:48.7899123Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7899461Z import refcycle 2025-12-04T09:26:48.7899646Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7899939Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7900585Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7901202Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7901833Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7902469Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7902772Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7903424Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7904039Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7904357Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7904970Z Traceback (most recent call last): 2025-12-04T09:26:48.7905302Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7905645Z import refcycle 2025-12-04T09:26:48.7905836Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7906130Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7906782Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7907408Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7908043Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7908663Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7908955Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7909599Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7910219Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7910532Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7910821Z Traceback (most recent call last): 2025-12-04T09:26:48.7911151Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7911491Z import refcycle 2025-12-04T09:26:48.7911681Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7911974Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7912617Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7913230Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7913846Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7914586Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7914891Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7915527Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7916305Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7916631Z _________________ ERROR at teardown of TestMin.test_attn_cuda __________________ 2025-12-04T09:26:48.7916919Z Traceback (most recent call last): 2025-12-04T09:26:48.7917246Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 109, in tearDown 2025-12-04T09:26:48.7917584Z import refcycle 2025-12-04T09:26:48.7917789Z ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.7918080Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7918725Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7919348Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7919974Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7920587Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7920891Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7921533Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7922153Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7922414Z =================================== FAILURES =================================== 2025-12-04T09:26:48.7922707Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.7922983Z Traceback (most recent call last): 2025-12-04T09:26:48.7923334Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.7923702Z self.attn( 2025-12-04T09:26:48.7923990Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.7924328Z a_out = A(hidden_state) 2025-12-04T09:26:48.7924759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.7925228Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.7925659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.7926087Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.7926449Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.7926886Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.7927368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.7927803Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.7928277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.7928788Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.7929303Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.7929772Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.7932002Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.74 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 286.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 982.64 MiB is allocated by PyTorch, and 19.36 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.7934197Z 2025-12-04T09:26:48.7934334Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.7934671Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.7934873Z 2025-12-04T09:26:48.7935041Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.7935420Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7936088Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7936735Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7937378Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7938007Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7938311Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.7938594Z Traceback (most recent call last): 2025-12-04T09:26:48.7938955Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.7939311Z self.attn( 2025-12-04T09:26:48.7939594Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.7939932Z a_out = A(hidden_state) 2025-12-04T09:26:48.7940363Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.7940819Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.7941247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.7941673Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.7942009Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.7942424Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.7942895Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.7943324Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.7943788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.7944309Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.7944838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.7945306Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.7947429Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.74 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 286.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.7949500Z 2025-12-04T09:26:48.7949707Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.7950050Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.7950256Z 2025-12-04T09:26:48.7950417Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.7950792Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7951457Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7952095Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7952728Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7953354Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7953664Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7954320Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7954943Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7955250Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.7955524Z Traceback (most recent call last): 2025-12-04T09:26:48.7955877Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.7956237Z self.attn( 2025-12-04T09:26:48.7956512Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.7956855Z a_out = A(hidden_state) 2025-12-04T09:26:48.7957281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.7957745Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.7958167Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.7958592Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.7958943Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.7959351Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.7959819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.7960252Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.7960721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.7961217Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.7961738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.7962206Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.7964300Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.74 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 286.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.7966370Z 2025-12-04T09:26:48.7966573Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.7966911Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.7967120Z 2025-12-04T09:26:48.7967283Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.7967651Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7968319Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7968954Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7969589Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7970219Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7970531Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7971175Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7971800Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7972106Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.7972388Z Traceback (most recent call last): 2025-12-04T09:26:48.7972756Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.7973120Z self.attn( 2025-12-04T09:26:48.7973494Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.7973818Z a_out = A(hidden_state) 2025-12-04T09:26:48.7974247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.7974715Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.7975135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.7975558Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.7975895Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.7976310Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.7976770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.7977198Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.7977664Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.7978168Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.7978682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.7979146Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.7981238Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.74 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 286.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.7983329Z 2025-12-04T09:26:48.7983525Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.7983862Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.7984065Z 2025-12-04T09:26:48.7984231Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.7984598Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7985260Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7985895Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7986523Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7987162Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.7987495Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.7988144Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.7988765Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.7989071Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.7989353Z Traceback (most recent call last): 2025-12-04T09:26:48.7989710Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.7990063Z self.attn( 2025-12-04T09:26:48.7990342Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.7990676Z a_out = A(hidden_state) 2025-12-04T09:26:48.7991097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.7991564Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.7991991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.7992417Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.7992752Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.7993170Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.7993643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.7994073Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.7994553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.7995063Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.7995591Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.7996055Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.7998156Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.74 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 286.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8000248Z 2025-12-04T09:26:48.8000447Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8000789Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8000993Z 2025-12-04T09:26:48.8001162Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8001539Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8002201Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8002845Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8003497Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8004128Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8004436Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8005429Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8006067Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8006379Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8006658Z Traceback (most recent call last): 2025-12-04T09:26:48.8007031Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8007404Z self.attn( 2025-12-04T09:26:48.8007679Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8008017Z a_out = A(hidden_state) 2025-12-04T09:26:48.8008471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8008940Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8009372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8009800Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8010138Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8010552Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8011040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8011467Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8011935Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8012433Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8012953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8013522Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8015629Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.74 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 286.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8017800Z 2025-12-04T09:26:48.8018043Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8018377Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8018583Z 2025-12-04T09:26:48.8018744Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8019113Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8019777Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8020414Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8021042Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8021670Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8021982Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8022624Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8023250Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8023559Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8023839Z Traceback (most recent call last): 2025-12-04T09:26:48.8024201Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8024582Z self.attn( 2025-12-04T09:26:48.8024866Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8025194Z a_out = A(hidden_state) 2025-12-04T09:26:48.8025621Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8026087Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8026503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8026935Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8027275Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8027687Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8028147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8028581Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8029057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8029568Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8030088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8030555Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8032651Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8034781Z 2025-12-04T09:26:48.8034921Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8035257Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8035463Z 2025-12-04T09:26:48.8035628Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8036005Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8036680Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8037323Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8037960Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8038592Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8038909Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8039566Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8040193Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8040502Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8040787Z Traceback (most recent call last): 2025-12-04T09:26:48.8041151Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8041511Z self.attn( 2025-12-04T09:26:48.8041792Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8042129Z a_out = A(hidden_state) 2025-12-04T09:26:48.8042555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8043018Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8043444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8043869Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8044212Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8044626Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8045099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8045530Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8046002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8046515Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8047046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8047511Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8049609Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8051768Z 2025-12-04T09:26:48.8051908Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8052243Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8052446Z 2025-12-04T09:26:48.8052617Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8052990Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8053772Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8054419Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8055055Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8055677Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8056000Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8056663Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8057297Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8057604Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8057890Z Traceback (most recent call last): 2025-12-04T09:26:48.8058256Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8058626Z self.attn( 2025-12-04T09:26:48.8058914Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8059258Z a_out = A(hidden_state) 2025-12-04T09:26:48.8059712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8060172Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8060603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8061037Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8061372Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8061788Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8062257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8062687Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8063144Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8063652Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8064178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8064647Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8066820Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8068875Z 2025-12-04T09:26:48.8069013Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8069343Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8069550Z 2025-12-04T09:26:48.8069710Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8070093Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8070765Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8071399Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8072030Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8072664Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8072792Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8073244Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8073355Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8073479Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8073564Z Traceback (most recent call last): 2025-12-04T09:26:48.8073784Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8073847Z self.attn( 2025-12-04T09:26:48.8074042Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8074115Z a_out = A(hidden_state) 2025-12-04T09:26:48.8074431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8074516Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8074788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8074872Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8075061Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8075209Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8075467Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8075572Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8075869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8076013Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8076314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8076413Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8078425Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8078524Z 2025-12-04T09:26:48.8078664Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8078789Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8078792Z 2025-12-04T09:26:48.8078958Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8079089Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8079535Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8079652Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8080101Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8080220Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8080348Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8080788Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8080902Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8081027Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8081110Z Traceback (most recent call last): 2025-12-04T09:26:48.8081334Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8081399Z self.attn( 2025-12-04T09:26:48.8081593Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8081664Z a_out = A(hidden_state) 2025-12-04T09:26:48.8081970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8082059Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8082331Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8082417Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8082607Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8082753Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8083012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8083115Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8083398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8083550Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8083852Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8083947Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8085956Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8086024Z 2025-12-04T09:26:48.8086169Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8086296Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8086299Z 2025-12-04T09:26:48.8086469Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8086601Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8087060Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8087180Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8087618Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8087734Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8087861Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8088303Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8088417Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8088538Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8088614Z Traceback (most recent call last): 2025-12-04T09:26:48.8088840Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8088899Z self.attn( 2025-12-04T09:26:48.8089090Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8089159Z a_out = A(hidden_state) 2025-12-04T09:26:48.8089464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8089551Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8089817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8089893Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8090084Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8090237Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8090485Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8090588Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8090872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8091022Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8091325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8091419Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8093552Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8093642Z 2025-12-04T09:26:48.8093781Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8093904Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8093907Z 2025-12-04T09:26:48.8094066Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8094199Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8094646Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8094764Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8095204Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8095316Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8095448Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8095889Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8096001Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8096122Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8096197Z Traceback (most recent call last): 2025-12-04T09:26:48.8096426Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8096486Z self.attn( 2025-12-04T09:26:48.8096674Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8096749Z a_out = A(hidden_state) 2025-12-04T09:26:48.8097059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8097147Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8097422Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8097499Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8097695Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8097836Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8098085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8098196Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8098482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8098635Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8098942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8099032Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8101053Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8101117Z 2025-12-04T09:26:48.8101255Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8101380Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8101383Z 2025-12-04T09:26:48.8101549Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8101692Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8102149Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8102267Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8102710Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8102819Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8102950Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8103391Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8103499Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8103625Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8103701Z Traceback (most recent call last): 2025-12-04T09:26:48.8103947Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8104008Z self.attn( 2025-12-04T09:26:48.8104195Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8104273Z a_out = A(hidden_state) 2025-12-04T09:26:48.8104895Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8104987Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8105269Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8105346Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8105541Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8105684Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8105934Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8106043Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8106329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8106482Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8106787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8106881Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8108978Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8109083Z 2025-12-04T09:26:48.8109221Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8109354Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8109358Z 2025-12-04T09:26:48.8109517Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8109652Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8110114Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8110229Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8110690Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8110794Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8110926Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8111364Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8111471Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8111602Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8111676Z Traceback (most recent call last): 2025-12-04T09:26:48.8111899Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8111964Z self.attn( 2025-12-04T09:26:48.8112148Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8112223Z a_out = A(hidden_state) 2025-12-04T09:26:48.8112531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8112614Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8112887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8112965Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8113150Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8113297Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8113549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8113654Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8113945Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8114087Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8114393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8114482Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8116496Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8116565Z 2025-12-04T09:26:48.8116698Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8116828Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8116831Z 2025-12-04T09:26:48.8116999Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8117128Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8117583Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8117695Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8118151Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8118257Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8118383Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8118841Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8118948Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8119074Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8119156Z Traceback (most recent call last): 2025-12-04T09:26:48.8119377Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8119444Z self.attn( 2025-12-04T09:26:48.8119629Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8119699Z a_out = A(hidden_state) 2025-12-04T09:26:48.8120010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8120090Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8120364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8120442Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8120628Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8121037Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8121346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8134646Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8135045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8135215Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8135566Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8135675Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8137788Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8137876Z 2025-12-04T09:26:48.8138025Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8138159Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8138163Z 2025-12-04T09:26:48.8138343Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8138487Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8138962Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8139088Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8139547Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8139666Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8139803Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8140258Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8140373Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8140500Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8140593Z Traceback (most recent call last): 2025-12-04T09:26:48.8140823Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8140891Z self.attn( 2025-12-04T09:26:48.8141086Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8141162Z a_out = A(hidden_state) 2025-12-04T09:26:48.8141488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8141574Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8141849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8141935Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8142126Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8142277Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8142547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8142653Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8142957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8143103Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8143407Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8143509Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8145527Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8145604Z 2025-12-04T09:26:48.8145743Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8145873Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8145877Z 2025-12-04T09:26:48.8146052Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8146188Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8146652Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8146767Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8147217Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8147331Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8147463Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8147900Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8148016Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8148152Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8148241Z Traceback (most recent call last): 2025-12-04T09:26:48.8148465Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8148530Z self.attn( 2025-12-04T09:26:48.8148728Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8148802Z a_out = A(hidden_state) 2025-12-04T09:26:48.8149109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8149201Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8149475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8149565Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8149756Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8149905Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8150160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8150264Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8150564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8150708Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8151009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8151107Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8153118Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8153209Z 2025-12-04T09:26:48.8153354Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8153482Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8153485Z 2025-12-04T09:26:48.8153655Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8153784Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8154238Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8154356Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8154801Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8154914Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8155040Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8155480Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8155594Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8155718Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8155805Z Traceback (most recent call last): 2025-12-04T09:26:48.8156027Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8156094Z self.attn( 2025-12-04T09:26:48.8156288Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8156359Z a_out = A(hidden_state) 2025-12-04T09:26:48.8156661Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8156749Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8157021Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8157107Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8157296Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8157441Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8157696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8157801Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8158088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8158237Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8158538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8158635Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8160648Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8160719Z 2025-12-04T09:26:48.8160865Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8160990Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8160995Z 2025-12-04T09:26:48.8161161Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8161294Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8161758Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8161880Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8162325Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8162432Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8162567Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8163012Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8163130Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8163255Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8163336Z Traceback (most recent call last): 2025-12-04T09:26:48.8163568Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8163632Z self.attn( 2025-12-04T09:26:48.8163827Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8163897Z a_out = A(hidden_state) 2025-12-04T09:26:48.8164199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8164286Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8164557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8164637Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8164830Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8164981Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8165237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8165350Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8165641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8165790Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8166089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8166179Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8168210Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8168280Z 2025-12-04T09:26:48.8168419Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8168544Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8168548Z 2025-12-04T09:26:48.8168705Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8168840Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8169293Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8169406Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8169861Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8169964Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8170098Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8170532Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8170638Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8170765Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8170843Z Traceback (most recent call last): 2025-12-04T09:26:48.8171067Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8171130Z self.attn( 2025-12-04T09:26:48.8171317Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8171388Z a_out = A(hidden_state) 2025-12-04T09:26:48.8171689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8171772Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8172055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8172136Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8172327Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8172474Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8172718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8172824Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8173111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8173349Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8173654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8173742Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8175749Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8175837Z 2025-12-04T09:26:48.8175971Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8176103Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8176107Z 2025-12-04T09:26:48.8176266Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8176401Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8176855Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8176965Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8177411Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8177513Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8177645Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8178081Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8178188Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8178317Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8178393Z Traceback (most recent call last): 2025-12-04T09:26:48.8178611Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8178681Z self.attn( 2025-12-04T09:26:48.8178874Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8178950Z a_out = A(hidden_state) 2025-12-04T09:26:48.8179252Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8179350Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8179629Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8179708Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8179898Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8180051Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8180297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8180408Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8180695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8180840Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8181149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8181240Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8183262Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8183349Z 2025-12-04T09:26:48.8183486Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8183620Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8183624Z 2025-12-04T09:26:48.8183785Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8183917Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8184377Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8184495Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8184947Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8185064Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8185197Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8185640Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8185748Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8185877Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8185956Z Traceback (most recent call last): 2025-12-04T09:26:48.8186179Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8186256Z self.attn( 2025-12-04T09:26:48.8186447Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8186517Z a_out = A(hidden_state) 2025-12-04T09:26:48.8186824Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8186908Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8187186Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8187265Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8187453Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8187606Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8187855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8187964Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8188260Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8188408Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8188715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8188806Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8190818Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8190884Z 2025-12-04T09:26:48.8191022Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8191154Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8191157Z 2025-12-04T09:26:48.8191319Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8191453Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8191908Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8192024Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8192474Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8192580Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8192710Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8193154Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8193262Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8193386Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8193467Z Traceback (most recent call last): 2025-12-04T09:26:48.8193689Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8193755Z self.attn( 2025-12-04T09:26:48.8193944Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8194013Z a_out = A(hidden_state) 2025-12-04T09:26:48.8194325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8194407Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8194679Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8194772Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8194966Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8195119Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8195362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8195466Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8195758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8195905Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8196214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8196305Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8198325Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8198401Z 2025-12-04T09:26:48.8198536Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8198660Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8198664Z 2025-12-04T09:26:48.8198831Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8198963Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8199422Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8199539Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8199979Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8200089Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8200215Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8200657Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8200764Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8200888Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8200971Z Traceback (most recent call last): 2025-12-04T09:26:48.8201194Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8201255Z self.attn( 2025-12-04T09:26:48.8201453Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8201524Z a_out = A(hidden_state) 2025-12-04T09:26:48.8201832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8201914Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8202186Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8202284Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8202476Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8202622Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8202874Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8202982Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8203274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8203416Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8203717Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8203812Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8206288Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8206403Z 2025-12-04T09:26:48.8206556Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8206690Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8206693Z 2025-12-04T09:26:48.8206864Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8207001Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8207453Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8207580Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8208025Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8208139Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8208272Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8208712Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8208827Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8208954Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8209045Z Traceback (most recent call last): 2025-12-04T09:26:48.8209276Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8209344Z self.attn( 2025-12-04T09:26:48.8209537Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8209609Z a_out = A(hidden_state) 2025-12-04T09:26:48.8209912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8210003Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8210274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8210359Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8210550Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8210693Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8210950Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8211060Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8211351Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8211493Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8211790Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8211885Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8214010Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8214114Z 2025-12-04T09:26:48.8214256Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8214386Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8214389Z 2025-12-04T09:26:48.8214555Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8214685Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8215135Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8215260Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8215703Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8215812Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8215939Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8216375Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8216495Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8216617Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8216694Z Traceback (most recent call last): 2025-12-04T09:26:48.8216931Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8217010Z self.attn( 2025-12-04T09:26:48.8217204Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8217274Z a_out = A(hidden_state) 2025-12-04T09:26:48.8217575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8217663Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8217935Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8218011Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8218206Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8218349Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8218601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8218708Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8218991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8219138Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8219438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8219532Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8221538Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8221608Z 2025-12-04T09:26:48.8221750Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8221874Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8221877Z 2025-12-04T09:26:48.8222035Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8222175Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8222624Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8222747Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8223189Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8223293Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8223429Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8223879Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8223994Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8224116Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8224192Z Traceback (most recent call last): 2025-12-04T09:26:48.8224418Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8224483Z self.attn( 2025-12-04T09:26:48.8224667Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8224743Z a_out = A(hidden_state) 2025-12-04T09:26:48.8225045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8225131Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8225400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8225478Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8225672Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8225814Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8226070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8226176Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8226460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8226612Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8226911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8227000Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8229012Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8229082Z 2025-12-04T09:26:48.8229231Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8229361Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8229365Z 2025-12-04T09:26:48.8229526Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8229663Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8230108Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8230233Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8230670Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8230774Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8230908Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8231347Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8231463Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8231587Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8231667Z Traceback (most recent call last): 2025-12-04T09:26:48.8231896Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8231962Z self.attn( 2025-12-04T09:26:48.8232146Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8232223Z a_out = A(hidden_state) 2025-12-04T09:26:48.8232519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8232609Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8232877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8232956Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8233153Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8233297Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8233544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8233657Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8233939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8234086Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8234387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8234477Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8236489Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8236555Z 2025-12-04T09:26:48.8236689Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8236837Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8236840Z 2025-12-04T09:26:48.8236999Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8237135Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8237590Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8237701Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8238146Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8238250Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8238380Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8238816Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8238924Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8239051Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8239128Z Traceback (most recent call last): 2025-12-04T09:26:48.8239356Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8239416Z self.attn( 2025-12-04T09:26:48.8239600Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8239675Z a_out = A(hidden_state) 2025-12-04T09:26:48.8239973Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8240053Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8240325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8240404Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8240594Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8240737Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8240991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8241097Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8241381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8241522Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8241832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8241920Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8243969Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8244036Z 2025-12-04T09:26:48.8244177Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8244301Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8244305Z 2025-12-04T09:26:48.8244467Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8244601Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8245056Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8245173Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8245613Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8245716Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8245850Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8246287Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8246403Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8246530Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8246610Z Traceback (most recent call last): 2025-12-04T09:26:48.8246839Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8246903Z self.attn( 2025-12-04T09:26:48.8247089Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8247165Z a_out = A(hidden_state) 2025-12-04T09:26:48.8247464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8247552Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8247819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8247898Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8248089Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8248231Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8248480Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8248588Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8248874Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8249022Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8249317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8249407Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8251417Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8251486Z 2025-12-04T09:26:48.8251627Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8251749Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8251754Z 2025-12-04T09:26:48.8251915Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8252061Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8252522Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8252641Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8253083Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8253274Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8253411Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8253852Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8253970Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8254099Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8254178Z Traceback (most recent call last): 2025-12-04T09:26:48.8254410Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8254471Z self.attn( 2025-12-04T09:26:48.8254656Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8254732Z a_out = A(hidden_state) 2025-12-04T09:26:48.8255034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8255118Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8255393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8255475Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8255667Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8255808Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8256060Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8256168Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8256449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8256602Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8256899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8256986Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8259034Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8259070Z 2025-12-04T09:26:48.8259206Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8259336Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8259341Z 2025-12-04T09:26:48.8259499Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8259630Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8260087Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8260195Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8260645Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8260748Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8260885Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8261331Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8261441Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8261569Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8261646Z Traceback (most recent call last): 2025-12-04T09:26:48.8261867Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8261932Z self.attn( 2025-12-04T09:26:48.8262117Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8262190Z a_out = A(hidden_state) 2025-12-04T09:26:48.8262489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8262570Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8262844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8262923Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8263108Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8263257Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8263505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8263612Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8263896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8264038Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8264357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8264447Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8266485Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8266540Z 2025-12-04T09:26:48.8266672Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8266810Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8266816Z 2025-12-04T09:26:48.8266975Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8267102Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8267564Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8267674Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8268119Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8268221Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8268349Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8268796Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8268909Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8269040Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8269119Z Traceback (most recent call last): 2025-12-04T09:26:48.8269341Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8269409Z self.attn( 2025-12-04T09:26:48.8269597Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8269664Z a_out = A(hidden_state) 2025-12-04T09:26:48.8269973Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8270057Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8270339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8270425Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8270612Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8270761Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8271010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8271116Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8271400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8271541Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8271846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8271937Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8273977Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8274014Z 2025-12-04T09:26:48.8274144Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8274273Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8274279Z 2025-12-04T09:26:48.8274438Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8274567Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8275026Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8275141Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8275587Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8275692Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8275821Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8276266Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8276376Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8276508Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8276588Z Traceback (most recent call last): 2025-12-04T09:26:48.8276807Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8276874Z self.attn( 2025-12-04T09:26:48.8277068Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8277136Z a_out = A(hidden_state) 2025-12-04T09:26:48.8277444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8277528Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8277800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8277885Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8278071Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8278220Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8278468Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8278570Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8278861Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8279001Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8279304Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8279392Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8281445Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8281486Z 2025-12-04T09:26:48.8281627Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8281754Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8281765Z 2025-12-04T09:26:48.8281922Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8282048Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8282509Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8282618Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8283056Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8283165Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8283292Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8283736Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8283845Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8283969Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8284056Z Traceback (most recent call last): 2025-12-04T09:26:48.8284272Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8284336Z self.attn( 2025-12-04T09:26:48.8284526Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8284594Z a_out = A(hidden_state) 2025-12-04T09:26:48.8284898Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8284977Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8285243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8285329Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8285513Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8285662Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8285912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8286015Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8286302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8286444Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8286747Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8286895Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8288886Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8288929Z 2025-12-04T09:26:48.8289058Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8289185Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8289189Z 2025-12-04T09:26:48.8289359Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8289491Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8289950Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8290064Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8290502Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8290613Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8290741Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8291186Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8291297Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8291423Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8291510Z Traceback (most recent call last): 2025-12-04T09:26:48.8291735Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8291802Z self.attn( 2025-12-04T09:26:48.8291994Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8292063Z a_out = A(hidden_state) 2025-12-04T09:26:48.8292373Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8292457Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8292727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8292816Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8293006Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8293245Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8293504Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8293609Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8293900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8294042Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8294339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8294483Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8296500Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8296545Z 2025-12-04T09:26:48.8296690Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8296821Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8296824Z 2025-12-04T09:26:48.8296993Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8297132Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8297587Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8297709Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8298154Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8298265Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8298398Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8298847Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8298971Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8299105Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8299197Z Traceback (most recent call last): 2025-12-04T09:26:48.8299423Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8299486Z self.attn( 2025-12-04T09:26:48.8299680Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8299761Z a_out = A(hidden_state) 2025-12-04T09:26:48.8300076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8300172Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8300449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8300536Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8300740Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8300894Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8301149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8301255Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8301547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8301696Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8301998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8302150Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8304152Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8304195Z 2025-12-04T09:26:48.8304332Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8304470Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8304473Z 2025-12-04T09:26:48.8304958Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8305109Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8305570Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8305698Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8306146Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8306265Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8306397Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8306839Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8306962Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8307094Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8307180Z Traceback (most recent call last): 2025-12-04T09:26:48.8307406Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8307468Z self.attn( 2025-12-04T09:26:48.8307661Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8307735Z a_out = A(hidden_state) 2025-12-04T09:26:48.8308040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8308128Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8308402Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8308486Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8308676Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8308825Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8309077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8309180Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8309464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8309612Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8309910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8310126Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8312168Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8312220Z 2025-12-04T09:26:48.8312372Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8312506Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8312509Z 2025-12-04T09:26:48.8312681Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8312818Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8313283Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8313404Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8313846Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8313958Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8314089Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8314537Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8314655Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8314783Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8314872Z Traceback (most recent call last): 2025-12-04T09:26:48.8315108Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8315170Z self.attn( 2025-12-04T09:26:48.8315364Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8315435Z a_out = A(hidden_state) 2025-12-04T09:26:48.8315743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8315835Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8316111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8316190Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8316388Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8316540Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8316801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8316906Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8317189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8317340Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8317642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8317783Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8319777Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8319813Z 2025-12-04T09:26:48.8319954Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8320083Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8320086Z 2025-12-04T09:26:48.8320248Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8320390Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8320850Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8320966Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8321406Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8321515Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8321651Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8322093Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8322211Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8322337Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8322415Z Traceback (most recent call last): 2025-12-04T09:26:48.8322652Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8322715Z self.attn( 2025-12-04T09:26:48.8322900Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8322975Z a_out = A(hidden_state) 2025-12-04T09:26:48.8323275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8323362Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8323635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8323712Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8323908Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8324057Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8324305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8324412Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8324695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8324846Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8325146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8325290Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8327326Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8327362Z 2025-12-04T09:26:48.8327505Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8327630Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8327633Z 2025-12-04T09:26:48.8327792Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8327936Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8328385Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8328506Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8328951Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8329067Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8329203Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8329644Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8329753Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8329887Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8329964Z Traceback (most recent call last): 2025-12-04T09:26:48.8330190Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8330252Z self.attn( 2025-12-04T09:26:48.8330440Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8330513Z a_out = A(hidden_state) 2025-12-04T09:26:48.8330812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8330897Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8331175Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8331262Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8331458Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8331606Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8331852Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8331959Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8332241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8332387Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8332681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8332827Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8335360Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8335398Z 2025-12-04T09:26:48.8335552Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8335687Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8335690Z 2025-12-04T09:26:48.8335856Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8336001Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8336460Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8336574Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8337032Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8337138Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8337273Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8337712Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8337822Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8337955Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8338033Z Traceback (most recent call last): 2025-12-04T09:26:48.8338254Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8338321Z self.attn( 2025-12-04T09:26:48.8338510Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8338584Z a_out = A(hidden_state) 2025-12-04T09:26:48.8338892Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8338977Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8339255Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8339335Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8339525Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8339678Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8339920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8340029Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8340312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8340452Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8340753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8340889Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8342901Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8342939Z 2025-12-04T09:26:48.8343089Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8343223Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8343229Z 2025-12-04T09:26:48.8343391Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8343522Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8343978Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8344088Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8344536Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8344639Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8344769Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8345213Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8345324Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8345455Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8345532Z Traceback (most recent call last): 2025-12-04T09:26:48.8345749Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8345821Z self.attn( 2025-12-04T09:26:48.8346003Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8346072Z a_out = A(hidden_state) 2025-12-04T09:26:48.8346379Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8346462Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8346748Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8346831Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8347022Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8347170Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8347411Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8347519Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8347801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8347946Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8348252Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8348383Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8350384Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8350421Z 2025-12-04T09:26:48.8350556Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8350688Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8350693Z 2025-12-04T09:26:48.8350852Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8350985Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8351442Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8351551Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8352004Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8352117Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8352248Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8352693Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8352809Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8352945Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8353023Z Traceback (most recent call last): 2025-12-04T09:26:48.8353240Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8353306Z self.attn( 2025-12-04T09:26:48.8353491Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8353560Z a_out = A(hidden_state) 2025-12-04T09:26:48.8353865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8353949Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8354222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8354300Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8354487Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8354635Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8354877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8354979Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8355265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8355405Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8355770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8355860Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8357883Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8357895Z 2025-12-04T09:26:48.8358024Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8358145Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8358155Z 2025-12-04T09:26:48.8358315Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8358448Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8358901Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8359012Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8359451Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8359563Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8359689Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8360136Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8360248Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8360371Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8360453Z Traceback (most recent call last): 2025-12-04T09:26:48.8360668Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8360733Z self.attn( 2025-12-04T09:26:48.8360917Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8360984Z a_out = A(hidden_state) 2025-12-04T09:26:48.8361288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8361370Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8361639Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8361729Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8361919Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8362064Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8362307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8362409Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8362695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8362836Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8363176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8363285Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8365317Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8365330Z 2025-12-04T09:26:48.8365462Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8365584Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8365589Z 2025-12-04T09:26:48.8365760Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8365889Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8366340Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8366450Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8366897Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8367006Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8367136Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8367582Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8367697Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8367820Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8367903Z Traceback (most recent call last): 2025-12-04T09:26:48.8368122Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8368184Z self.attn( 2025-12-04T09:26:48.8368377Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8368448Z a_out = A(hidden_state) 2025-12-04T09:26:48.8368755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8368840Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8369110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8369196Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8369388Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8369534Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8369804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8369909Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8370207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8370351Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8370697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8370805Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8372846Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8372860Z 2025-12-04T09:26:48.8372996Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8373123Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8373128Z 2025-12-04T09:26:48.8373401Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8373537Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8373992Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8374116Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8374558Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8374675Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8374807Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8375247Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8375369Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8375492Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8375578Z Traceback (most recent call last): 2025-12-04T09:26:48.8375802Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8375864Z self.attn( 2025-12-04T09:26:48.8376057Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8376128Z a_out = A(hidden_state) 2025-12-04T09:26:48.8376435Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8376529Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8376803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8376891Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8377084Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8377231Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8377487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8377589Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8377872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8378020Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8378367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8378462Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8380506Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8380515Z 2025-12-04T09:26:48.8380665Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8380797Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8380801Z 2025-12-04T09:26:48.8380970Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8381102Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8381556Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8381674Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8382117Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8382233Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8382363Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8382801Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8382922Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8383043Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8383121Z Traceback (most recent call last): 2025-12-04T09:26:48.8383347Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8383407Z self.attn( 2025-12-04T09:26:48.8383597Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8383665Z a_out = A(hidden_state) 2025-12-04T09:26:48.8383966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8384054Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8384321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8384403Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8384595Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8384739Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8384989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8385094Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8385373Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8385520Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8385862Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8385987Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8387992Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8387998Z 2025-12-04T09:26:48.8388134Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8388261Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8388264Z 2025-12-04T09:26:48.8388425Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8388560Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8389008Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8389122Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8389565Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8389675Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8389808Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8390249Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8390364Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8390486Z ____________________________ TestMin.test_attn_cuda ____________________________ 2025-12-04T09:26:48.8390564Z Traceback (most recent call last): 2025-12-04T09:26:48.8390792Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 291, in test_attn_cuda 2025-12-04T09:26:48.8390852Z self.attn( 2025-12-04T09:26:48.8391041Z File "/var/lib/jenkins/workspace/test/functorch/test_dims.py", line 173, in attn 2025-12-04T09:26:48.8391117Z a_out = A(hidden_state) 2025-12-04T09:26:48.8391416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1778, in _wrapped_call_impl 2025-12-04T09:26:48.8391505Z return self._call_impl(*args, **kwargs) 2025-12-04T09:26:48.8391770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1789, in _call_impl 2025-12-04T09:26:48.8391856Z return forward_call(*args, **kwargs) 2025-12-04T09:26:48.8392048Z File "/var/lib/jenkins/workspace/test/functorch/attn_ft.py", line 100, in forward 2025-12-04T09:26:48.8392191Z attention_scores = (q * k).sum(features) / math.sqrt(features.size) 2025-12-04T09:26:48.8392432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/wrap_type.py", line 30, in impl 2025-12-04T09:26:48.8392553Z return __torch_function__(orig, None, args, kwargs) 2025-12-04T09:26:48.8392834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 537, in __torch_function__ 2025-12-04T09:26:48.8392988Z return _Tensor._torch_function_fallback(func, types, args, kwargs) 2025-12-04T09:26:48.8393329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/functorch/dim/__init__.py", line 576, in _torch_function_fallback 2025-12-04T09:26:48.8393450Z result = func(*unflat_args, **unflat_kwargs) 2025-12-04T09:26:48.8395451Z torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8395459Z 2025-12-04T09:26:48.8395599Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8395725Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8395728Z 2025-12-04T09:26:48.8395889Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8396024Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8396472Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8396589Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8397031Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:98: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8397140Z isinstance(o, (torch.Tensor, Dim, Tensor, DimList)) 2025-12-04T09:26:48.8397272Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:26:48.8397714Z /var/lib/jenkins/workspace/test/functorch/test_dims.py:89: FutureWarning: `torch.distributed.reduce_op` is deprecated, please use `torch.distributed.ReduceOp` instead 2025-12-04T09:26:48.8397826Z if isinstance(o, (torch.Tensor, Dim, Tensor, DimList)): 2025-12-04T09:26:48.8398233Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_dims/functorch.test_dims-09de3759aef74478.xml - 2025-12-04T09:26:48.8398338Z =========================== short test summary info ============================ 2025-12-04T09:26:48.8400466Z FAILED [0.7465s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.74 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 286.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 982.64 MiB is allocated by PyTorch, and 19.36 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8400475Z 2025-12-04T09:26:48.8400606Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8400739Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8400742Z 2025-12-04T09:26:48.8400900Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8403136Z FAILED [0.5085s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.74 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 286.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8403173Z 2025-12-04T09:26:48.8403306Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8403442Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8403445Z 2025-12-04T09:26:48.8403605Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8406060Z FAILED [0.5312s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.74 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 286.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8406068Z 2025-12-04T09:26:48.8406212Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8406358Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8406362Z 2025-12-04T09:26:48.8406530Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8408740Z FAILED [0.5263s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.74 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 286.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8408748Z 2025-12-04T09:26:48.8408881Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8409013Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8409017Z 2025-12-04T09:26:48.8409177Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8411431Z FAILED [0.5527s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.74 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 286.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8411551Z 2025-12-04T09:26:48.8411683Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8411809Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8411813Z 2025-12-04T09:26:48.8411971Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8414179Z FAILED [0.5068s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.74 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 286.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8414188Z 2025-12-04T09:26:48.8414316Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8414442Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8414445Z 2025-12-04T09:26:48.8414602Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8416701Z FAILED [0.5153s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8416708Z 2025-12-04T09:26:48.8416833Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8416957Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8416960Z 2025-12-04T09:26:48.8417116Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8419217Z FAILED [0.5144s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8419265Z 2025-12-04T09:26:48.8419423Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8419609Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8419613Z 2025-12-04T09:26:48.8419767Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8421867Z FAILED [0.4989s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8421878Z 2025-12-04T09:26:48.8422004Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8422130Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8422133Z 2025-12-04T09:26:48.8422287Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8424395Z FAILED [0.5142s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8424402Z 2025-12-04T09:26:48.8424527Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8424651Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8424654Z 2025-12-04T09:26:48.8424807Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8426901Z FAILED [0.5081s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8426909Z 2025-12-04T09:26:48.8427032Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8427157Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8427198Z 2025-12-04T09:26:48.8427354Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8429500Z FAILED [0.5169s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8429538Z 2025-12-04T09:26:48.8429663Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8429786Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8429791Z 2025-12-04T09:26:48.8429947Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8432051Z FAILED [0.5431s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8432058Z 2025-12-04T09:26:48.8432186Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8432314Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8432317Z 2025-12-04T09:26:48.8432469Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8434566Z FAILED [0.5753s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8434572Z 2025-12-04T09:26:48.8434695Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8434819Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8434822Z 2025-12-04T09:26:48.8434976Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8437139Z FAILED [0.5083s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8437203Z 2025-12-04T09:26:48.8437328Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8437456Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8437459Z 2025-12-04T09:26:48.8437613Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8439712Z FAILED [0.4991s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8439725Z 2025-12-04T09:26:48.8439848Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8439968Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8439978Z 2025-12-04T09:26:48.8440132Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8442220Z FAILED [0.4964s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8442230Z 2025-12-04T09:26:48.8442354Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8442475Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8442490Z 2025-12-04T09:26:48.8442642Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8444806Z FAILED [0.5279s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8444882Z 2025-12-04T09:26:48.8445007Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8445128Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8445132Z 2025-12-04T09:26:48.8445292Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8447390Z FAILED [0.5092s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8447405Z 2025-12-04T09:26:48.8447529Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8447648Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8447652Z 2025-12-04T09:26:48.8447814Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8449898Z FAILED [0.5357s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8449913Z 2025-12-04T09:26:48.8450038Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8450157Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8450162Z 2025-12-04T09:26:48.8450324Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8452431Z FAILED [0.5000s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 288.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8452480Z 2025-12-04T09:26:48.8452604Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8452726Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8452764Z 2025-12-04T09:26:48.8452987Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8455155Z FAILED [0.5007s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8455169Z 2025-12-04T09:26:48.8455298Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8455418Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8455421Z 2025-12-04T09:26:48.8455582Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8457671Z FAILED [0.4994s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8457684Z 2025-12-04T09:26:48.8457808Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8457929Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8457932Z 2025-12-04T09:26:48.8458092Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8460177Z FAILED [0.4976s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8460189Z 2025-12-04T09:26:48.8460312Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8460431Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8460434Z 2025-12-04T09:26:48.8460595Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8462802Z FAILED [0.4939s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8462843Z 2025-12-04T09:26:48.8462968Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8463091Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8463095Z 2025-12-04T09:26:48.8463254Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8465347Z FAILED [0.4925s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8465353Z 2025-12-04T09:26:48.8465481Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8465604Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8465610Z 2025-12-04T09:26:48.8465768Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8467845Z FAILED [0.4966s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8467856Z 2025-12-04T09:26:48.8467984Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8468105Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8468108Z 2025-12-04T09:26:48.8468266Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8470418Z FAILED [0.5233s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8470482Z 2025-12-04T09:26:48.8470614Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8470734Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8470737Z 2025-12-04T09:26:48.8470894Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8472982Z FAILED [0.5393s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8472989Z 2025-12-04T09:26:48.8473116Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8473236Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8473240Z 2025-12-04T09:26:48.8473402Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8475494Z FAILED [0.4935s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8475501Z 2025-12-04T09:26:48.8475632Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8475749Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8475754Z 2025-12-04T09:26:48.8475917Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8478028Z FAILED [0.5174s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8478120Z 2025-12-04T09:26:48.8478251Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8478433Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8478437Z 2025-12-04T09:26:48.8478596Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8480699Z FAILED [0.4824s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8480707Z 2025-12-04T09:26:48.8480837Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8480955Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8480958Z 2025-12-04T09:26:48.8481113Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8483216Z FAILED [0.5143s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8483223Z 2025-12-04T09:26:48.8483354Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8483473Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8483476Z 2025-12-04T09:26:48.8483630Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8485730Z FAILED [0.5275s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8485735Z 2025-12-04T09:26:48.8485866Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8486022Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8486025Z 2025-12-04T09:26:48.8486179Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8488392Z FAILED [0.5215s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8488399Z 2025-12-04T09:26:48.8488531Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8488653Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8488656Z 2025-12-04T09:26:48.8488814Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8490904Z FAILED [0.5562s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 290.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8490912Z 2025-12-04T09:26:48.8491045Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8491164Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8491167Z 2025-12-04T09:26:48.8491321Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8493485Z FAILED [0.4977s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8493492Z 2025-12-04T09:26:48.8493624Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8493744Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8493748Z 2025-12-04T09:26:48.8493904Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8496072Z FAILED [0.5006s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8496138Z 2025-12-04T09:26:48.8496268Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8496388Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8496393Z 2025-12-04T09:26:48.8496546Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8498643Z FAILED [0.4985s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8498651Z 2025-12-04T09:26:48.8498779Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8498913Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8498919Z 2025-12-04T09:26:48.8499078Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8501173Z FAILED [0.4943s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8501181Z 2025-12-04T09:26:48.8501305Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8501434Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8501438Z 2025-12-04T09:26:48.8501593Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8503765Z FAILED [0.5068s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8503848Z 2025-12-04T09:26:48.8503974Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8504105Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8504108Z 2025-12-04T09:26:48.8504262Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8506675Z FAILED [0.5132s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8506691Z 2025-12-04T09:26:48.8506822Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8506950Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8506953Z 2025-12-04T09:26:48.8507115Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8509221Z FAILED [0.5059s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8509227Z 2025-12-04T09:26:48.8509353Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8509481Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8509484Z 2025-12-04T09:26:48.8509641Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8511747Z FAILED [0.5334s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8511841Z 2025-12-04T09:26:48.8511970Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8512151Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8512154Z 2025-12-04T09:26:48.8512405Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8514506Z FAILED [0.4988s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8514513Z 2025-12-04T09:26:48.8514641Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8514765Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8514769Z 2025-12-04T09:26:48.8514920Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8517017Z FAILED [0.4973s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8517027Z 2025-12-04T09:26:48.8517150Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8517282Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8517287Z 2025-12-04T09:26:48.8517443Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8519532Z FAILED [0.5081s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8519540Z 2025-12-04T09:26:48.8519662Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8519788Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8519791Z 2025-12-04T09:26:48.8519982Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8522150Z FAILED [0.4997s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8522188Z 2025-12-04T09:26:48.8522314Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8522437Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8522440Z 2025-12-04T09:26:48.8522596Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8524692Z FAILED [0.5044s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8524697Z 2025-12-04T09:26:48.8524833Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8524961Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8524964Z 2025-12-04T09:26:48.8525117Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8527223Z FAILED [0.5144s] functorch/test_dims.py::TestMin::test_attn_cuda - torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 12.00 GiB. GPU 0 has a total capacity of 22.03 GiB of which 19.73 GiB is free. Process 8024 has 186.00 MiB memory in use. Process 8134 has 188.00 MiB memory in use. Process 11952 has 186.00 MiB memory in use. Process 12089 has 292.00 MiB memory in use. Process 16789 has 186.00 MiB memory in use. Including non-PyTorch memory, this process has 1.20 GiB memory in use. 6.39 GiB allowed; Of the allocated memory 983.27 MiB is allocated by PyTorch, and 18.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) 2025-12-04T09:26:48.8527230Z 2025-12-04T09:26:48.8527355Z To execute this test, run the following from the base repo dir: 2025-12-04T09:26:48.8527479Z python test/functorch/test_dims.py TestMin.test_attn_cuda 2025-12-04T09:26:48.8527482Z 2025-12-04T09:26:48.8527636Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:26:48.8527899Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8528155Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8528450Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8528704Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8529051Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8529296Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8529552Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8529797Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8530047Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8530295Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8530538Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8530797Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8531041Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8531291Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8531533Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8531776Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8532040Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8532287Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8532541Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8532787Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8533032Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8533374Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8533620Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8533869Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8534119Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8534373Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8534630Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8534880Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8535132Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8535378Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8535622Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8535922Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8536207Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8536550Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8536809Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8537054Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8537307Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8537552Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8537806Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8538055Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8538308Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8538560Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8538805Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8539056Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8539300Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8539549Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8539803Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8540054Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8540318Z ERROR [0.0001s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8540566Z ERROR [0.0005s] functorch/test_dims.py::TestMin::test_attn_cuda - ModuleNotFoundError: No module named 'refcycle' 2025-12-04T09:26:48.8540675Z ======================== 50 failed, 50 errors in 26.19s ======================== 2025-12-04T09:26:48.8540679Z 2025-12-04T09:26:48.8540976Z FINISHED PRINTING LOG FILE of functorch/test_dims 1/1 (test/test-reports/functorch.test_dims_1.1_99eaaa239e58b576_.log) 2025-12-04T09:26:48.8540982Z 2025-12-04T09:26:48.8541174Z Finished functorch/test_dims 1/1 ... [2025-12-04 09:26:48.757434][2044.716976756], took 0.49min 2025-12-04T09:26:48.8541611Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_dims/functorch.test_dims-09de3759aef74478.xml 2025-12-04T09:26:49.1911030Z Uploading logs for 57116084869 to S3 2025-12-04T09:26:49.2344475Z Uploading artifacts took 0.37 seconds 2025-12-04T09:26:49.2344929Z functorch/test_dims 1/1 failed! 2025-12-04T09:26:52.2270842Z Running test_torchfuzz_repros 1/1 ... [2025-12-04 09:26:52.226597][2048.186140773] 2025-12-04T09:26:52.2271513Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:26:52.2272988Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_torchfuzz_repros.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:26:52.226996] 2025-12-04T09:26:55.4537288Z 2025-12-04T09:26:55.4538451Z test_torchfuzz_repros 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_torchfuzz_repros_1.1_e32a95fbe8787815_.log 2025-12-04T09:26:55.4539673Z Running 0 items in this shard: 2025-12-04T09:26:55.4539817Z 2025-12-04T09:26:55.4540242Z Finished test_torchfuzz_repros 1/1 ... [2025-12-04 09:26:55.453575][2051.413118782], took 0.05min 2025-12-04T09:26:55.4642317Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_torchfuzz_repros/test_torchfuzz_repros-184e97a40463bc74.xml 2025-12-04T09:26:58.9284073Z Running inductor/test_torchinductor 1/2 ... [2025-12-04 09:26:58.927914][2054.887457693] 2025-12-04T09:26:58.9284628Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:26:58.9286594Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor.py', '-m', 'not serial', '--shard-id=1', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:26:58.928328] 2025-12-04T09:27:23.8951461Z 2025-12-04T09:27:23.8954961Z inductor/test_torchinductor 1/2 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_1.2_140eeae6a9700c9e_.log 2025-12-04T09:27:23.8980316Z Running 100 items in this shard: test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mark_unbacked_with_hint_override_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice1_cuda 2025-12-04T09:27:23.9003398Z 2025-12-04T09:27:23.9003634Z Finished inductor/test_torchinductor 1/2 ... [2025-12-04 09:27:23.895099][2079.85464523], took 0.42min 2025-12-04T09:27:23.9058319Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-dbac887b32dd50ab.xml 2025-12-04T09:27:27.4162505Z Running inductor/test_torchinductor 2/2 ... [2025-12-04 09:27:27.415756][2083.37529952] 2025-12-04T09:27:27.4163042Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:27:27.4164625Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor.py', '-m', 'not serial', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:27:27.416151] 2025-12-04T09:28:30.9031213Z 2025-12-04T09:28:30.9035153Z inductor/test_torchinductor 2/2 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_2.2_d10e54041d6e8139_.log 2025-12-04T09:28:30.9143496Z Running 250 items in this shard: test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda 2025-12-04T09:28:30.9250947Z 2025-12-04T09:28:30.9251411Z Finished inductor/test_torchinductor 2/2 ... [2025-12-04 09:28:30.903376][2146.862920098], took 1.06min 2025-12-04T09:28:30.9253011Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-22efee56a99319ab.xml 2025-12-04T09:28:34.4749710Z Running inductor/test_kernel_benchmark 1/1 ... [2025-12-04 09:28:34.474487][2150.434030371] 2025-12-04T09:28:34.4750288Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:28:34.4752453Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_kernel_benchmark.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:28:34.474926] 2025-12-04T09:28:45.7207125Z 2025-12-04T09:28:45.7208408Z PRINTING LOG FILE of inductor/test_kernel_benchmark 1/1 (test/test-reports/inductor.test_kernel_benchmark_1.1_4dd3a5ee490d965b_.log) 2025-12-04T09:28:45.7210268Z W1204 09:28:39.454000 22429 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T09:28:45.7212184Z Test results will be stored in test-reports/python-pytest/inductor.test_kernel_benchmark/inductor.test_kernel_benchmark-8244c39f5fb36d0d.xml 2025-12-04T09:28:45.7213567Z ============================= test session starts ============================== 2025-12-04T09:28:45.7214465Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:28:45.7215227Z cachedir: .pytest_cache 2025-12-04T09:28:45.7216156Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:28:45.7217293Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:28:45.7217660Z configfile: pytest.ini 2025-12-04T09:28:45.7218406Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:28:45.7219194Z collecting ... collected 18 items 2025-12-04T09:28:45.7219658Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:28:45.7263445Z Running 100 items in this shard: test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark, test/inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark 2025-12-04T09:28:45.7300590Z 2025-12-04T09:28:45.7301173Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:40.575000 22429 site-packages/torch/_inductor/utils.py:1703] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-12-04T09:28:45.7302077Z W1204 09:28:40.586000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpk_7rxd7_ 2025-12-04T09:28:45.7302511Z FAILED [1.2294s] [ 1%] 2025-12-04T09:28:45.7303172Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7304332Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:40.851000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpuaaxbr0d 2025-12-04T09:28:45.7305672Z FAILED [0.0606s] [ 2%] 2025-12-04T09:28:45.7306703Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:40.913000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpuffc3rbz 2025-12-04T09:28:45.7307860Z FAILED [0.0606s] [ 2%] 2025-12-04T09:28:45.7308953Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:41.003000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpijb11q7p 2025-12-04T09:28:45.7310081Z FAILED [0.0887s] [ 2%] 2025-12-04T09:28:45.7311133Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:41.092000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp7b91o4wi 2025-12-04T09:28:45.7312358Z FAILED [0.0859s] [ 2%] 2025-12-04T09:28:45.7313561Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:41.155000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp8ene1m5t 2025-12-04T09:28:45.7314705Z FAILED [0.0613s] [ 2%] 2025-12-04T09:28:45.7315736Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:41.220000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmps_ub6ksh 2025-12-04T09:28:45.7316817Z FAILED [0.0625s] [ 2%] 2025-12-04T09:28:45.7317853Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:41.282000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp1elhxu6e 2025-12-04T09:28:45.7318918Z FAILED [0.0604s] [ 2%] 2025-12-04T09:28:45.7319960Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:41.346000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpyxgffvut 2025-12-04T09:28:45.7321040Z FAILED [0.0600s] [ 2%] 2025-12-04T09:28:45.7322037Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:41.409000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpt46jo9jj 2025-12-04T09:28:45.7323093Z FAILED [0.0614s] [ 2%] 2025-12-04T09:28:45.7324084Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:41.473000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpfun1h4ho 2025-12-04T09:28:45.7325121Z FAILED [0.0619s] [ 2%] 2025-12-04T09:28:45.7326119Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:41.534000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpt7uzp97y 2025-12-04T09:28:45.7327206Z FAILED [0.0593s] [ 2%] 2025-12-04T09:28:45.7328249Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:41.595000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpts2kuok2 2025-12-04T09:28:45.7329311Z FAILED [0.0589s] [ 2%] 2025-12-04T09:28:45.7330302Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:41.656000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpl40qsx_t 2025-12-04T09:28:45.7331381Z FAILED [0.0591s] [ 2%] 2025-12-04T09:28:45.7332407Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:41.717000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmprxyaatpq 2025-12-04T09:28:45.7333597Z FAILED [0.0594s] [ 2%] 2025-12-04T09:28:45.7334563Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:41.779000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpe6dvslc0 2025-12-04T09:28:45.7335627Z FAILED [0.0598s] [ 2%] 2025-12-04T09:28:45.7336650Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:41.847000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmprsyieokp 2025-12-04T09:28:45.7337704Z FAILED [0.0663s] [ 2%] 2025-12-04T09:28:45.7338793Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:41.913000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpfi0tlm30 2025-12-04T09:28:45.7339512Z FAILED [0.0628s] [ 2%] 2025-12-04T09:28:45.7340152Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:41.976000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpzg0ugxq3 2025-12-04T09:28:45.7340879Z FAILED [0.0617s] [ 2%] 2025-12-04T09:28:45.7341616Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:42.038000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp2trfo1ik 2025-12-04T09:28:45.7342290Z FAILED [0.0599s] [ 2%] 2025-12-04T09:28:45.7342919Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:42.101000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpkhpka5y8 2025-12-04T09:28:45.7343590Z FAILED [0.0617s] [ 2%] 2025-12-04T09:28:45.7344215Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:42.163000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpztogskix 2025-12-04T09:28:45.7344877Z FAILED [0.0595s] [ 2%] 2025-12-04T09:28:45.7345504Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:42.223000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpbpm4lcoi 2025-12-04T09:28:45.7346176Z FAILED [0.0588s] [ 2%] 2025-12-04T09:28:45.7346800Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:42.284000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpccvitsx8 2025-12-04T09:28:45.7347472Z FAILED [0.0583s] [ 2%] 2025-12-04T09:28:45.7348101Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:42.343000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpg6nmm_oq 2025-12-04T09:28:45.7348758Z FAILED [0.0580s] [ 2%] 2025-12-04T09:28:45.7349385Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:42.404000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpb3vap6pz 2025-12-04T09:28:45.7350052Z FAILED [0.0588s] [ 2%] 2025-12-04T09:28:45.7350685Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:42.467000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp0ipmb4ch 2025-12-04T09:28:45.7351349Z FAILED [0.0596s] [ 2%] 2025-12-04T09:28:45.7351989Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:42.528000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp38a43cuu 2025-12-04T09:28:45.7352651Z FAILED [0.0599s] [ 2%] 2025-12-04T09:28:45.7353277Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:42.593000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp8l3xoxa2 2025-12-04T09:28:45.7353939Z FAILED [0.0627s] [ 2%] 2025-12-04T09:28:45.7354565Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:42.655000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpc91nv2rx 2025-12-04T09:28:45.7355234Z FAILED [0.0598s] [ 2%] 2025-12-04T09:28:45.7355867Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:42.716000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpv1b8umxg 2025-12-04T09:28:45.7356532Z FAILED [0.0598s] [ 2%] 2025-12-04T09:28:45.7357160Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:42.777000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpqweoau28 2025-12-04T09:28:45.7357821Z FAILED [0.0587s] [ 2%] 2025-12-04T09:28:45.7358495Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:42.837000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpv8680i_z 2025-12-04T09:28:45.7359195Z FAILED [0.0580s] [ 2%] 2025-12-04T09:28:45.7359884Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:42.899000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp16mrer_3 2025-12-04T09:28:45.7360548Z FAILED [0.0604s] [ 2%] 2025-12-04T09:28:45.7361172Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:42.960000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpqlkb3z3h 2025-12-04T09:28:45.7361847Z FAILED [0.0589s] [ 2%] 2025-12-04T09:28:45.7362474Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:43.022000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmph8qadxzu 2025-12-04T09:28:45.7363137Z FAILED [0.0587s] [ 2%] 2025-12-04T09:28:45.7363762Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:43.085000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpi8umtwnz 2025-12-04T09:28:45.7364424Z FAILED [0.0624s] [ 2%] 2025-12-04T09:28:45.7365057Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:43.148000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp845kety_ 2025-12-04T09:28:45.7365719Z FAILED [0.0603s] [ 2%] 2025-12-04T09:28:45.7366357Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:43.209000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpcdj11p11 2025-12-04T09:28:45.7367034Z FAILED [0.0586s] [ 2%] 2025-12-04T09:28:45.7367673Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:43.268000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpk22ucx82 2025-12-04T09:28:45.7368340Z FAILED [0.0578s] [ 2%] 2025-12-04T09:28:45.7368982Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:43.329000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp5h7tu44r 2025-12-04T09:28:45.7369647Z FAILED [0.0589s] [ 2%] 2025-12-04T09:28:45.7370275Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:43.391000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp8u4lro4f 2025-12-04T09:28:45.7370939Z FAILED [0.0603s] [ 2%] 2025-12-04T09:28:45.7371571Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:43.453000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpsjcibs2s 2025-12-04T09:28:45.7372235Z FAILED [0.0603s] [ 2%] 2025-12-04T09:28:45.7372873Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:43.515000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpasyj1fq0 2025-12-04T09:28:45.7373692Z FAILED [0.0607s] [ 2%] 2025-12-04T09:28:45.7374328Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:43.579000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmph51clk8k 2025-12-04T09:28:45.7374999Z FAILED [0.0602s] [ 2%] 2025-12-04T09:28:45.7375628Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:43.641000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpexnbci35 2025-12-04T09:28:45.7376344Z FAILED [0.0599s] [ 2%] 2025-12-04T09:28:45.7376965Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:43.703000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp8ao_mw4s 2025-12-04T09:28:45.7377734Z FAILED [0.0606s] [ 2%] 2025-12-04T09:28:45.7378367Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:43.765000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmphvf3r1r_ 2025-12-04T09:28:45.7379026Z FAILED [0.0601s] [ 2%] 2025-12-04T09:28:45.7379649Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:43.827000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp3e2v02s_ 2025-12-04T09:28:45.7380312Z FAILED [0.0604s] [ 2%] 2025-12-04T09:28:45.7380937Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark W1204 09:28:43.889000 22429 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp0_fl5ys3 2025-12-04T09:28:45.7381597Z FAILED [0.0600s] [ 2%] 2025-12-04T09:28:45.7382230Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7383355Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7384469Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7385596Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7386716Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7387829Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7388939Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7390047Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7391163Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7392277Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7393378Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7394483Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7395668Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7396958Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7398278Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7399404Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7400510Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7401643Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7402755Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7403862Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7405373Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7406528Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7407656Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7408781Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7409897Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7411019Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7412152Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7413379Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7414501Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7415622Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7416982Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7418106Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7419523Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7420669Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7421788Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7422907Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7424029Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7425144Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7426255Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7427368Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7428479Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7429591Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7430699Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7431815Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7432925Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7434029Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7435132Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7436291Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7437521Z inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_pw_kernel_benchmark SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:28:45.7438125Z 2025-12-04T09:28:45.7438222Z =================================== FAILURES =================================== 2025-12-04T09:28:45.7438574Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7438899Z Traceback (most recent call last): 2025-12-04T09:28:45.7439364Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7439817Z f(a, b) 2025-12-04T09:28:45.7440205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7440721Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7441264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7441777Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7442270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7442736Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7443216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7443802Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7444385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7444842Z graph.run(*example_inputs) 2025-12-04T09:28:45.7445218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7445609Z return super().run(*args) 2025-12-04T09:28:45.7445966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7446364Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7446775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7447194Z result = super().run_node(n) 2025-12-04T09:28:45.7447578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7448015Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7448466Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7448975Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7449473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7449958Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7450432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7450841Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7451265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7451689Z return autotune_select_algorithm( 2025-12-04T09:28:45.7452177Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7452710Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7453274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7453970Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7454962Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7455875Z target: aten.mm.default 2025-12-04T09:28:45.7456054Z args[0]: TensorBox( 2025-12-04T09:28:45.7456230Z ReinterpretView( 2025-12-04T09:28:45.7456409Z StorageBox( 2025-12-04T09:28:45.7456763Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7457149Z ), 2025-12-04T09:28:45.7457381Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7457686Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7457893Z stack_traces = {, 2025-12-04T09:28:45.7458238Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7458617Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7458816Z , 2025-12-04T09:28:45.7458948Z } 2025-12-04T09:28:45.7459082Z ) 2025-12-04T09:28:45.7459208Z ) 2025-12-04T09:28:45.7459358Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7459734Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7460106Z )) 2025-12-04T09:28:45.7460185Z 2025-12-04T09:28:45.7460610Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7461106Z 2025-12-04T09:28:45.7461110Z 2025-12-04T09:28:45.7461245Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7461706Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7462031Z 2025-12-04T09:28:45.7462192Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7462576Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7462859Z frames [('total', 1)] 2025-12-04T09:28:45.7463037Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7463238Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7463517Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7463793Z graph_break [] 2025-12-04T09:28:45.7463967Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7464294Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7464620Z Traceback (most recent call last): 2025-12-04T09:28:45.7465074Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7465530Z f(a, b) 2025-12-04T09:28:45.7465908Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7466425Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7466959Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7467465Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7467967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7468486Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7468959Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7469642Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7470227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7470677Z graph.run(*example_inputs) 2025-12-04T09:28:45.7471047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7471433Z return super().run(*args) 2025-12-04T09:28:45.7471797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7472185Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7472596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7473002Z result = super().run_node(n) 2025-12-04T09:28:45.7473382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7473825Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7474279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7474787Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7475277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7475755Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7476219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7476630Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7477024Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7477451Z return autotune_select_algorithm( 2025-12-04T09:28:45.7477940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7478427Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7478848Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7479377Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7480360Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7481239Z target: aten.mm.default 2025-12-04T09:28:45.7481421Z args[0]: TensorBox( 2025-12-04T09:28:45.7481592Z ReinterpretView( 2025-12-04T09:28:45.7481758Z StorageBox( 2025-12-04T09:28:45.7482100Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7482474Z ), 2025-12-04T09:28:45.7482707Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7483003Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7483208Z stack_traces = {, 2025-12-04T09:28:45.7483541Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7483915Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7484125Z , 2025-12-04T09:28:45.7484262Z } 2025-12-04T09:28:45.7484389Z ) 2025-12-04T09:28:45.7484523Z ) 2025-12-04T09:28:45.7484723Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7485092Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7485476Z )) 2025-12-04T09:28:45.7485601Z 2025-12-04T09:28:45.7486086Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7486579Z 2025-12-04T09:28:45.7486582Z 2025-12-04T09:28:45.7486721Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7487183Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7487509Z 2025-12-04T09:28:45.7487674Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7488052Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7488333Z frames [('total', 1)] 2025-12-04T09:28:45.7488511Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7488726Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7489009Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7489280Z graph_break [] 2025-12-04T09:28:45.7489461Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7489755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7490033Z frames [('total', 1)] 2025-12-04T09:28:45.7490204Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7490464Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7490752Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7490953Z graph_break [] 2025-12-04T09:28:45.7491130Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7491459Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7491775Z Traceback (most recent call last): 2025-12-04T09:28:45.7492233Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7492677Z f(a, b) 2025-12-04T09:28:45.7493055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7493665Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7494191Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7494685Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7495166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7495630Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7496102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7496687Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7497266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7497724Z graph.run(*example_inputs) 2025-12-04T09:28:45.7498091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7498476Z return super().run(*args) 2025-12-04T09:28:45.7498831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7499225Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7499636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7500040Z result = super().run_node(n) 2025-12-04T09:28:45.7500425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7500906Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7501358Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7502027Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7502520Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7503014Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7503475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7503883Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7504282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7505050Z return autotune_select_algorithm( 2025-12-04T09:28:45.7505538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7506040Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7506468Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7506996Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7507975Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7508866Z target: aten.mm.default 2025-12-04T09:28:45.7509051Z args[0]: TensorBox( 2025-12-04T09:28:45.7509217Z ReinterpretView( 2025-12-04T09:28:45.7509396Z StorageBox( 2025-12-04T09:28:45.7509739Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7510115Z ), 2025-12-04T09:28:45.7510348Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7510662Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7510871Z stack_traces = {, 2025-12-04T09:28:45.7511202Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7511589Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7511795Z , 2025-12-04T09:28:45.7511924Z } 2025-12-04T09:28:45.7512060Z ) 2025-12-04T09:28:45.7512191Z ) 2025-12-04T09:28:45.7512338Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7512707Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7513077Z )) 2025-12-04T09:28:45.7513153Z 2025-12-04T09:28:45.7513575Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7514062Z 2025-12-04T09:28:45.7514065Z 2025-12-04T09:28:45.7514200Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7514656Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7514980Z 2025-12-04T09:28:45.7515140Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7515520Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7515791Z frames [('total', 1)] 2025-12-04T09:28:45.7515967Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7516168Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7516441Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7516912Z graph_break [] 2025-12-04T09:28:45.7517088Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7517378Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7517730Z frames [('total', 1)] 2025-12-04T09:28:45.7517909Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7518292Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7518577Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7518798Z graph_break [] 2025-12-04T09:28:45.7518978Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7519254Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7519530Z frames [('total', 1)] 2025-12-04T09:28:45.7519705Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7519957Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7520247Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7520452Z graph_break [] 2025-12-04T09:28:45.7520625Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7520941Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7521268Z Traceback (most recent call last): 2025-12-04T09:28:45.7521735Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7522199Z f(a, b) 2025-12-04T09:28:45.7522588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7523107Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7523647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7524148Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7524649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7525124Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7525599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7526191Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7526979Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7527449Z graph.run(*example_inputs) 2025-12-04T09:28:45.7527815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7528195Z return super().run(*args) 2025-12-04T09:28:45.7528555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7528968Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7529368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7529777Z result = super().run_node(n) 2025-12-04T09:28:45.7530166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7530593Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7531040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7531533Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7532020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7532488Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7532953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7533522Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7533915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7534379Z return autotune_select_algorithm( 2025-12-04T09:28:45.7534931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7535424Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7535845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7536373Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7537359Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7538245Z target: aten.mm.default 2025-12-04T09:28:45.7538426Z args[0]: TensorBox( 2025-12-04T09:28:45.7538601Z ReinterpretView( 2025-12-04T09:28:45.7538771Z StorageBox( 2025-12-04T09:28:45.7539107Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7539483Z ), 2025-12-04T09:28:45.7539734Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7540041Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7540241Z stack_traces = {, 2025-12-04T09:28:45.7540577Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7540957Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7541160Z , 2025-12-04T09:28:45.7541297Z } 2025-12-04T09:28:45.7541435Z ) 2025-12-04T09:28:45.7541562Z ) 2025-12-04T09:28:45.7541724Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7542102Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7542469Z )) 2025-12-04T09:28:45.7542553Z 2025-12-04T09:28:45.7542979Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7543476Z 2025-12-04T09:28:45.7543479Z 2025-12-04T09:28:45.7543609Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7544068Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7544390Z 2025-12-04T09:28:45.7544558Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7544927Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7545210Z frames [('total', 1)] 2025-12-04T09:28:45.7545395Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7545593Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7545876Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7546148Z graph_break [] 2025-12-04T09:28:45.7546321Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7546606Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7546877Z frames [('total', 1)] 2025-12-04T09:28:45.7547051Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7547321Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7547603Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7547802Z graph_break [] 2025-12-04T09:28:45.7547967Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7548245Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7548578Z frames [('total', 1)] 2025-12-04T09:28:45.7548743Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7548991Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7549313Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7549504Z graph_break [] 2025-12-04T09:28:45.7549737Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7550015Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7550286Z frames [('total', 1)] 2025-12-04T09:28:45.7550451Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7550703Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7550984Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7551176Z graph_break [] 2025-12-04T09:28:45.7551358Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7551675Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7551994Z Traceback (most recent call last): 2025-12-04T09:28:45.7552446Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7552894Z f(a, b) 2025-12-04T09:28:45.7553274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7553781Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7554313Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7554812Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7555298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7555770Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7556253Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7556838Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7557423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7557883Z graph.run(*example_inputs) 2025-12-04T09:28:45.7558258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7558643Z return super().run(*args) 2025-12-04T09:28:45.7559001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7559398Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7559812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7560211Z result = super().run_node(n) 2025-12-04T09:28:45.7560599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7561048Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7561516Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7562011Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7562503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7562986Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7563441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7563849Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7564244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7564707Z return autotune_select_algorithm( 2025-12-04T09:28:45.7565186Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7565712Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7566202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7566731Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7567703Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7568584Z target: aten.mm.default 2025-12-04T09:28:45.7568782Z args[0]: TensorBox( 2025-12-04T09:28:45.7568953Z ReinterpretView( 2025-12-04T09:28:45.7569121Z StorageBox( 2025-12-04T09:28:45.7569459Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7569841Z ), 2025-12-04T09:28:45.7570073Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7570379Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7570583Z stack_traces = {, 2025-12-04T09:28:45.7570916Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7571299Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7571510Z , 2025-12-04T09:28:45.7571647Z } 2025-12-04T09:28:45.7571780Z ) 2025-12-04T09:28:45.7571913Z ) 2025-12-04T09:28:45.7572056Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7572427Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7572804Z )) 2025-12-04T09:28:45.7572893Z 2025-12-04T09:28:45.7573441Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7573937Z 2025-12-04T09:28:45.7573945Z 2025-12-04T09:28:45.7574080Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7574543Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7574875Z 2025-12-04T09:28:45.7575039Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7575416Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7575691Z frames [('total', 1)] 2025-12-04T09:28:45.7575876Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7576086Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7576360Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7576643Z graph_break [] 2025-12-04T09:28:45.7576823Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7577113Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7577382Z frames [('total', 1)] 2025-12-04T09:28:45.7577562Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7577826Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7578104Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7578312Z graph_break [] 2025-12-04T09:28:45.7578489Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7578772Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7579045Z frames [('total', 1)] 2025-12-04T09:28:45.7579217Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7579464Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7579806Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7580001Z graph_break [] 2025-12-04T09:28:45.7580168Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7580496Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7580771Z frames [('total', 1)] 2025-12-04T09:28:45.7581007Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7581254Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7581538Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7581736Z graph_break [] 2025-12-04T09:28:45.7581901Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7582173Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7582442Z frames [('total', 1)] 2025-12-04T09:28:45.7582608Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7582860Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7583145Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7583360Z graph_break [] 2025-12-04T09:28:45.7583522Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7583841Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7584160Z Traceback (most recent call last): 2025-12-04T09:28:45.7584617Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7585065Z f(a, b) 2025-12-04T09:28:45.7585439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7585952Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7586481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7586979Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7587481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7587944Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7588435Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7589014Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7589597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7590048Z graph.run(*example_inputs) 2025-12-04T09:28:45.7590443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7590836Z return super().run(*args) 2025-12-04T09:28:45.7591201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7591592Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7617670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7618273Z result = super().run_node(n) 2025-12-04T09:28:45.7618727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7619199Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7619688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7620227Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7620744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7621242Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7621960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7622393Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7624167Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7624737Z return autotune_select_algorithm( 2025-12-04T09:28:45.7625254Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7625744Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7626169Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7626693Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7627684Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7628591Z target: aten.mm.default 2025-12-04T09:28:45.7628781Z args[0]: TensorBox( 2025-12-04T09:28:45.7628956Z ReinterpretView( 2025-12-04T09:28:45.7629136Z StorageBox( 2025-12-04T09:28:45.7629489Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7629867Z ), 2025-12-04T09:28:45.7630106Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7630416Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7630629Z stack_traces = {, 2025-12-04T09:28:45.7630970Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7631357Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7631572Z , 2025-12-04T09:28:45.7631708Z } 2025-12-04T09:28:45.7631848Z ) 2025-12-04T09:28:45.7631987Z ) 2025-12-04T09:28:45.7632138Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7632512Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7632891Z )) 2025-12-04T09:28:45.7632974Z 2025-12-04T09:28:45.7633398Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7633897Z 2025-12-04T09:28:45.7633900Z 2025-12-04T09:28:45.7634035Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7634506Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7634827Z 2025-12-04T09:28:45.7635001Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7635403Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7635678Z frames [('total', 1)] 2025-12-04T09:28:45.7635866Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7636082Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7636364Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7636639Z graph_break [] 2025-12-04T09:28:45.7636820Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7637112Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7637395Z frames [('total', 1)] 2025-12-04T09:28:45.7637567Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7637820Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7638106Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7638313Z graph_break [] 2025-12-04T09:28:45.7638489Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7638852Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7639117Z frames [('total', 1)] 2025-12-04T09:28:45.7639286Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7639578Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7639863Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7640131Z graph_break [] 2025-12-04T09:28:45.7640313Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7640599Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7640869Z frames [('total', 1)] 2025-12-04T09:28:45.7641040Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7641291Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7641570Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7641764Z graph_break [] 2025-12-04T09:28:45.7641935Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7642210Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7642477Z frames [('total', 1)] 2025-12-04T09:28:45.7642641Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7642899Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7643177Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7643374Z graph_break [] 2025-12-04T09:28:45.7643537Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7643809Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7644071Z frames [('total', 1)] 2025-12-04T09:28:45.7644241Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7644486Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7644762Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7644958Z graph_break [] 2025-12-04T09:28:45.7645120Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7645431Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7645750Z Traceback (most recent call last): 2025-12-04T09:28:45.7646204Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7646650Z f(a, b) 2025-12-04T09:28:45.7647034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7647543Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7648081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7648584Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7649067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7649533Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7650010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7650582Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7651156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7651602Z graph.run(*example_inputs) 2025-12-04T09:28:45.7651972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7652354Z return super().run(*args) 2025-12-04T09:28:45.7652712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7653195Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7653605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7654052Z result = super().run_node(n) 2025-12-04T09:28:45.7654431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7654898Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7655399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7655894Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7656380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7656861Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7657325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7657728Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7658135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7658557Z return autotune_select_algorithm( 2025-12-04T09:28:45.7659046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7659562Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7660001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7660526Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7661513Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7662396Z target: aten.mm.default 2025-12-04T09:28:45.7662581Z args[0]: TensorBox( 2025-12-04T09:28:45.7662751Z ReinterpretView( 2025-12-04T09:28:45.7662931Z StorageBox( 2025-12-04T09:28:45.7663277Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7663650Z ), 2025-12-04T09:28:45.7663887Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7664192Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7664396Z stack_traces = {, 2025-12-04T09:28:45.7664728Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7665104Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7665308Z , 2025-12-04T09:28:45.7665436Z } 2025-12-04T09:28:45.7665570Z ) 2025-12-04T09:28:45.7665700Z ) 2025-12-04T09:28:45.7665846Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7666214Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7666588Z )) 2025-12-04T09:28:45.7666664Z 2025-12-04T09:28:45.7667096Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7667596Z 2025-12-04T09:28:45.7667599Z 2025-12-04T09:28:45.7667734Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7668198Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7668522Z 2025-12-04T09:28:45.7668691Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7669065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7669337Z frames [('total', 1)] 2025-12-04T09:28:45.7669516Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7669774Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7670049Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7670324Z graph_break [] 2025-12-04T09:28:45.7670501Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7670827Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7671166Z frames [('total', 1)] 2025-12-04T09:28:45.7671348Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7671597Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7671880Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7672087Z graph_break [] 2025-12-04T09:28:45.7672259Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7672546Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7672823Z frames [('total', 1)] 2025-12-04T09:28:45.7672996Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7673247Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7673532Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7673734Z graph_break [] 2025-12-04T09:28:45.7673901Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7674184Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7674469Z frames [('total', 1)] 2025-12-04T09:28:45.7674645Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7674906Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7675193Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7675394Z graph_break [] 2025-12-04T09:28:45.7675559Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7675835Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7676105Z frames [('total', 1)] 2025-12-04T09:28:45.7676286Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7676548Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7676833Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7677029Z graph_break [] 2025-12-04T09:28:45.7677197Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7677475Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7677765Z frames [('total', 1)] 2025-12-04T09:28:45.7677930Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7678183Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7678466Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7678659Z graph_break [] 2025-12-04T09:28:45.7678824Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7679097Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7679363Z frames [('total', 1)] 2025-12-04T09:28:45.7679533Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7679801Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7680078Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7680274Z graph_break [] 2025-12-04T09:28:45.7680447Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7680763Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7681083Z Traceback (most recent call last): 2025-12-04T09:28:45.7681539Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7681990Z f(a, b) 2025-12-04T09:28:45.7682366Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7682881Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7683416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7683914Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7684454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7684956Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7685579Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7686166Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7686742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7687210Z graph.run(*example_inputs) 2025-12-04T09:28:45.7687586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7687971Z return super().run(*args) 2025-12-04T09:28:45.7688333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7688744Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7689157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7689565Z result = super().run_node(n) 2025-12-04T09:28:45.7689953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7690390Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7690834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7691333Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7691827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7692309Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7692779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7693316Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7693734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7694154Z return autotune_select_algorithm( 2025-12-04T09:28:45.7694642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7695130Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7695552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7696072Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7697075Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7697969Z target: aten.mm.default 2025-12-04T09:28:45.7698155Z args[0]: TensorBox( 2025-12-04T09:28:45.7698328Z ReinterpretView( 2025-12-04T09:28:45.7698499Z StorageBox( 2025-12-04T09:28:45.7698846Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7699216Z ), 2025-12-04T09:28:45.7699454Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7699772Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7699980Z stack_traces = {, 2025-12-04T09:28:45.7700323Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7700703Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7700961Z , 2025-12-04T09:28:45.7701091Z } 2025-12-04T09:28:45.7701225Z ) 2025-12-04T09:28:45.7701359Z ) 2025-12-04T09:28:45.7701505Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7701923Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7702373Z )) 2025-12-04T09:28:45.7702452Z 2025-12-04T09:28:45.7702883Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7703381Z 2025-12-04T09:28:45.7703385Z 2025-12-04T09:28:45.7703519Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7703974Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7704297Z 2025-12-04T09:28:45.7704468Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7705571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7705875Z frames [('total', 1)] 2025-12-04T09:28:45.7706069Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7706297Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7706603Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7706946Z graph_break [] 2025-12-04T09:28:45.7707177Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7707527Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7707868Z frames [('total', 1)] 2025-12-04T09:28:45.7708084Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7708396Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7708728Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7708947Z graph_break [] 2025-12-04T09:28:45.7709126Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7709412Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7709683Z frames [('total', 1)] 2025-12-04T09:28:45.7709858Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7710108Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7710410Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7710613Z graph_break [] 2025-12-04T09:28:45.7710780Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7711060Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7711334Z frames [('total', 1)] 2025-12-04T09:28:45.7711512Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7711760Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7712041Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7712245Z graph_break [] 2025-12-04T09:28:45.7712410Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7712692Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7712962Z frames [('total', 1)] 2025-12-04T09:28:45.7713127Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7713386Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7713689Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7713881Z graph_break [] 2025-12-04T09:28:45.7714046Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7714320Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7714592Z frames [('total', 1)] 2025-12-04T09:28:45.7714755Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7715002Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7715278Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7715468Z graph_break [] 2025-12-04T09:28:45.7715634Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7716113Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7716375Z frames [('total', 1)] 2025-12-04T09:28:45.7716563Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7716954Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7717293Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7717674Z graph_break [] 2025-12-04T09:28:45.7717880Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7718206Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7718505Z frames [('total', 1)] 2025-12-04T09:28:45.7718679Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7718930Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7719206Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7719407Z graph_break [] 2025-12-04T09:28:45.7719575Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7719883Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7720209Z Traceback (most recent call last): 2025-12-04T09:28:45.7720673Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7721128Z f(a, b) 2025-12-04T09:28:45.7721518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7722043Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7722576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7723066Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7723555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7724017Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7724496Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7725075Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7725665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7726146Z graph.run(*example_inputs) 2025-12-04T09:28:45.7726519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7726905Z return super().run(*args) 2025-12-04T09:28:45.7727276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7727671Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7728075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7728486Z result = super().run_node(n) 2025-12-04T09:28:45.7728869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7729308Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7729781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7730294Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7730794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7731269Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7731726Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7732133Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7732586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7733005Z return autotune_select_algorithm( 2025-12-04T09:28:45.7733652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7734214Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7734635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7735152Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7736137Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7737043Z target: aten.mm.default 2025-12-04T09:28:45.7737226Z args[0]: TensorBox( 2025-12-04T09:28:45.7737393Z ReinterpretView( 2025-12-04T09:28:45.7737561Z StorageBox( 2025-12-04T09:28:45.7737900Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7738269Z ), 2025-12-04T09:28:45.7738507Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7738810Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7739011Z stack_traces = {, 2025-12-04T09:28:45.7739363Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7739741Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7739946Z , 2025-12-04T09:28:45.7740075Z } 2025-12-04T09:28:45.7740208Z ) 2025-12-04T09:28:45.7740344Z ) 2025-12-04T09:28:45.7740488Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7740860Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7741230Z )) 2025-12-04T09:28:45.7741307Z 2025-12-04T09:28:45.7741732Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7742234Z 2025-12-04T09:28:45.7742237Z 2025-12-04T09:28:45.7742373Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7742826Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7743151Z 2025-12-04T09:28:45.7743320Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7743694Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7743976Z frames [('total', 1)] 2025-12-04T09:28:45.7744160Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7744365Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7744644Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7744930Z graph_break [] 2025-12-04T09:28:45.7745114Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7745397Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7745669Z frames [('total', 1)] 2025-12-04T09:28:45.7745844Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7746097Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7746386Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7746595Z graph_break [] 2025-12-04T09:28:45.7746785Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7747059Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7747330Z frames [('total', 1)] 2025-12-04T09:28:45.7747510Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7747805Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7748087Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7748288Z graph_break [] 2025-12-04T09:28:45.7748453Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7748771Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7749154Z frames [('total', 1)] 2025-12-04T09:28:45.7749338Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7749583Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7749866Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7750065Z graph_break [] 2025-12-04T09:28:45.7750226Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7750503Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7750770Z frames [('total', 1)] 2025-12-04T09:28:45.7750933Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7751187Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7751471Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7751664Z graph_break [] 2025-12-04T09:28:45.7751831Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7752109Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7752384Z frames [('total', 1)] 2025-12-04T09:28:45.7752559Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7752811Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7753087Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7753277Z graph_break [] 2025-12-04T09:28:45.7753444Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7753718Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7753984Z frames [('total', 1)] 2025-12-04T09:28:45.7754156Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7754405Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7754680Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7754882Z graph_break [] 2025-12-04T09:28:45.7755047Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7755326Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7755595Z frames [('total', 1)] 2025-12-04T09:28:45.7755765Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7756024Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7756305Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7756506Z graph_break [] 2025-12-04T09:28:45.7756676Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7756944Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7757224Z frames [('total', 1)] 2025-12-04T09:28:45.7757398Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7757659Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7757943Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7758141Z graph_break [] 2025-12-04T09:28:45.7758309Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7758624Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7758948Z Traceback (most recent call last): 2025-12-04T09:28:45.7759411Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7759849Z f(a, b) 2025-12-04T09:28:45.7760243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7760760Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7761294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7761791Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7762334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7762842Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7763382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7763969Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7764544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7764998Z graph.run(*example_inputs) 2025-12-04T09:28:45.7765361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7765746Z return super().run(*args) 2025-12-04T09:28:45.7766102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7766493Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7766897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7767309Z result = super().run_node(n) 2025-12-04T09:28:45.7767700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7768135Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7768585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7769080Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7769576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7770054Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7770534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7770952Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7771361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7771779Z return autotune_select_algorithm( 2025-12-04T09:28:45.7772263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7772758Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7773043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7773302Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7774056Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7774135Z target: aten.mm.default 2025-12-04T09:28:45.7774204Z args[0]: TensorBox( 2025-12-04T09:28:45.7774274Z ReinterpretView( 2025-12-04T09:28:45.7774336Z StorageBox( 2025-12-04T09:28:45.7774586Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7774643Z ), 2025-12-04T09:28:45.7774794Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7774875Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7774939Z stack_traces = {, 2025-12-04T09:28:45.7775175Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7775251Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7775359Z , 2025-12-04T09:28:45.7775420Z } 2025-12-04T09:28:45.7775474Z ) 2025-12-04T09:28:45.7775528Z ) 2025-12-04T09:28:45.7775607Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7775881Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7776016Z )) 2025-12-04T09:28:45.7776021Z 2025-12-04T09:28:45.7776454Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7776457Z 2025-12-04T09:28:45.7776461Z 2025-12-04T09:28:45.7776598Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7776849Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7776853Z 2025-12-04T09:28:45.7777018Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7777174Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7777240Z frames [('total', 1)] 2025-12-04T09:28:45.7777312Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7777399Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7777537Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7777600Z graph_break [] 2025-12-04T09:28:45.7777692Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7777820Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7777888Z frames [('total', 1)] 2025-12-04T09:28:45.7777968Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7778095Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7778177Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7778237Z graph_break [] 2025-12-04T09:28:45.7778318Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7778461Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7778528Z frames [('total', 1)] 2025-12-04T09:28:45.7778598Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7778737Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7778814Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7778874Z graph_break [] 2025-12-04T09:28:45.7778964Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7779089Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7779160Z frames [('total', 1)] 2025-12-04T09:28:45.7779233Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7779358Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7779436Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7779496Z graph_break [] 2025-12-04T09:28:45.7779571Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7779704Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7779765Z frames [('total', 1)] 2025-12-04T09:28:45.7779835Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7779965Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7780041Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7780101Z graph_break [] 2025-12-04T09:28:45.7780182Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7780304Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7780372Z frames [('total', 1)] 2025-12-04T09:28:45.7780439Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7780563Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7780638Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7780697Z graph_break [] 2025-12-04T09:28:45.7780773Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7780950Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7781012Z frames [('total', 1)] 2025-12-04T09:28:45.7781079Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7781246Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7781318Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7781455Z graph_break [] 2025-12-04T09:28:45.7781531Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7781657Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7781724Z frames [('total', 1)] 2025-12-04T09:28:45.7781791Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7781914Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7781993Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7782052Z graph_break [] 2025-12-04T09:28:45.7782127Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7782255Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7782323Z frames [('total', 1)] 2025-12-04T09:28:45.7782397Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7782517Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7782594Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7782664Z graph_break [] 2025-12-04T09:28:45.7782742Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7782865Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7782936Z frames [('total', 1)] 2025-12-04T09:28:45.7783005Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7783130Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7783211Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7783272Z graph_break [] 2025-12-04T09:28:45.7783357Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7783521Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7783601Z Traceback (most recent call last): 2025-12-04T09:28:45.7783922Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7783983Z f(a, b) 2025-12-04T09:28:45.7784277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7784449Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7784756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7784882Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7785180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7785273Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7785595Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7785791Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7786105Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7786181Z graph.run(*example_inputs) 2025-12-04T09:28:45.7786416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7786491Z return super().run(*args) 2025-12-04T09:28:45.7786722Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7786800Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7787069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7787188Z result = super().run_node(n) 2025-12-04T09:28:45.7787444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7787606Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7787941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7788097Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7788360Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7788492Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7788756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7788829Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7789103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7789182Z return autotune_select_algorithm( 2025-12-04T09:28:45.7789522Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7789611Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7789900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7790068Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7790814Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7790886Z target: aten.mm.default 2025-12-04T09:28:45.7790959Z args[0]: TensorBox( 2025-12-04T09:28:45.7791024Z ReinterpretView( 2025-12-04T09:28:45.7791085Z StorageBox( 2025-12-04T09:28:45.7791337Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7791398Z ), 2025-12-04T09:28:45.7791557Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7791635Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7791698Z stack_traces = {, 2025-12-04T09:28:45.7791929Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7792004Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7792061Z , 2025-12-04T09:28:45.7792122Z } 2025-12-04T09:28:45.7792177Z ) 2025-12-04T09:28:45.7792231Z ) 2025-12-04T09:28:45.7792309Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7792543Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7792599Z )) 2025-12-04T09:28:45.7792609Z 2025-12-04T09:28:45.7793033Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7793042Z 2025-12-04T09:28:45.7793046Z 2025-12-04T09:28:45.7793182Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7793452Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7793455Z 2025-12-04T09:28:45.7793624Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7793761Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7793826Z frames [('total', 1)] 2025-12-04T09:28:45.7793897Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7794026Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7794161Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7794221Z graph_break [] 2025-12-04T09:28:45.7794310Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7794476Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7794609Z frames [('total', 1)] 2025-12-04T09:28:45.7794679Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7794807Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7794884Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7794943Z graph_break [] 2025-12-04T09:28:45.7795021Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7795152Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7795214Z frames [('total', 1)] 2025-12-04T09:28:45.7795281Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7795423Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7795502Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7795561Z graph_break [] 2025-12-04T09:28:45.7795643Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7795769Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7795842Z frames [('total', 1)] 2025-12-04T09:28:45.7795910Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7796033Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7796112Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7796171Z graph_break [] 2025-12-04T09:28:45.7796248Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7796377Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7796437Z frames [('total', 1)] 2025-12-04T09:28:45.7796506Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7796635Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7796713Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7796778Z graph_break [] 2025-12-04T09:28:45.7796855Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7796982Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7797053Z frames [('total', 1)] 2025-12-04T09:28:45.7797121Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7797243Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7797323Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7797382Z graph_break [] 2025-12-04T09:28:45.7797459Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7797604Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7797668Z frames [('total', 1)] 2025-12-04T09:28:45.7797741Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7797864Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7797939Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7798004Z graph_break [] 2025-12-04T09:28:45.7798081Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7798207Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7798276Z frames [('total', 1)] 2025-12-04T09:28:45.7798349Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7798473Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7798552Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7798611Z graph_break [] 2025-12-04T09:28:45.7798692Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7798816Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7798878Z frames [('total', 1)] 2025-12-04T09:28:45.7798950Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7799084Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7799214Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7799282Z graph_break [] 2025-12-04T09:28:45.7799360Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7799522Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7799591Z frames [('total', 1)] 2025-12-04T09:28:45.7799727Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7799857Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7799930Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7799990Z graph_break [] 2025-12-04T09:28:45.7800075Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7800200Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7800266Z frames [('total', 1)] 2025-12-04T09:28:45.7800342Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7800465Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7800543Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7800610Z graph_break [] 2025-12-04T09:28:45.7800685Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7800855Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7800935Z Traceback (most recent call last): 2025-12-04T09:28:45.7801256Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7801321Z f(a, b) 2025-12-04T09:28:45.7801609Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7801760Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7802067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7802184Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7802488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7802579Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7802905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7803109Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7803408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7803486Z graph.run(*example_inputs) 2025-12-04T09:28:45.7803720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7803788Z return super().run(*args) 2025-12-04T09:28:45.7804026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7804106Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7804355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7804432Z result = super().run_node(n) 2025-12-04T09:28:45.7805013Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7805127Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7805397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7805544Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7805812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7805942Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7806350Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7806429Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7806752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7806939Z return autotune_select_algorithm( 2025-12-04T09:28:45.7807292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7807365Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7807673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7807835Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7808584Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7808657Z target: aten.mm.default 2025-12-04T09:28:45.7808723Z args[0]: TensorBox( 2025-12-04T09:28:45.7808794Z ReinterpretView( 2025-12-04T09:28:45.7808858Z StorageBox( 2025-12-04T09:28:45.7809098Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7809160Z ), 2025-12-04T09:28:45.7809313Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7809390Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7809456Z stack_traces = {, 2025-12-04T09:28:45.7809687Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7809768Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7809827Z , 2025-12-04T09:28:45.7809881Z } 2025-12-04T09:28:45.7809942Z ) 2025-12-04T09:28:45.7809996Z ) 2025-12-04T09:28:45.7810069Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7810307Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7810365Z )) 2025-12-04T09:28:45.7810376Z 2025-12-04T09:28:45.7810810Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7810813Z 2025-12-04T09:28:45.7810816Z 2025-12-04T09:28:45.7810951Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7811195Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7811199Z 2025-12-04T09:28:45.7811369Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7811504Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7811575Z frames [('total', 1)] 2025-12-04T09:28:45.7811649Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7811727Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7811867Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7811940Z graph_break [] 2025-12-04T09:28:45.7812032Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7812167Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7812232Z frames [('total', 1)] 2025-12-04T09:28:45.7812301Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7812436Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7812510Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7812576Z graph_break [] 2025-12-04T09:28:45.7812652Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7812847Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7812920Z frames [('total', 1)] 2025-12-04T09:28:45.7812989Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7813206Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7813344Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7813475Z graph_break [] 2025-12-04T09:28:45.7813561Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7813704Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7813768Z frames [('total', 1)] 2025-12-04T09:28:45.7813843Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7813975Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7814052Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7814119Z graph_break [] 2025-12-04T09:28:45.7814198Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7814325Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7814401Z frames [('total', 1)] 2025-12-04T09:28:45.7814471Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7814596Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7814679Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7814743Z graph_break [] 2025-12-04T09:28:45.7814828Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7814955Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7815018Z frames [('total', 1)] 2025-12-04T09:28:45.7815092Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7815218Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7815302Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7815369Z graph_break [] 2025-12-04T09:28:45.7815451Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7815577Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7815649Z frames [('total', 1)] 2025-12-04T09:28:45.7815716Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7815845Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7815920Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7815980Z graph_break [] 2025-12-04T09:28:45.7816064Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7816187Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7816249Z frames [('total', 1)] 2025-12-04T09:28:45.7816323Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7816448Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7816518Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7816586Z graph_break [] 2025-12-04T09:28:45.7816661Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7816790Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7816853Z frames [('total', 1)] 2025-12-04T09:28:45.7816930Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7817057Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7817130Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7817188Z graph_break [] 2025-12-04T09:28:45.7817275Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7817399Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7817462Z frames [('total', 1)] 2025-12-04T09:28:45.7817537Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7817660Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7817736Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7817799Z graph_break [] 2025-12-04T09:28:45.7817874Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7818004Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7818114Z frames [('total', 1)] 2025-12-04T09:28:45.7818182Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7818310Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7818418Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7818478Z graph_break [] 2025-12-04T09:28:45.7818639Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7818767Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7818829Z frames [('total', 1)] 2025-12-04T09:28:45.7818904Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7819028Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7819107Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7819166Z graph_break [] 2025-12-04T09:28:45.7819242Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7819410Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7819489Z Traceback (most recent call last): 2025-12-04T09:28:45.7819808Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7819876Z f(a, b) 2025-12-04T09:28:45.7820181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7820347Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7820649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7820768Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7821074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7821164Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7821476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7821684Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7821996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7822082Z graph.run(*example_inputs) 2025-12-04T09:28:45.7822322Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7822393Z return super().run(*args) 2025-12-04T09:28:45.7822638Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7822722Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7822983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7823057Z result = super().run_node(n) 2025-12-04T09:28:45.7823309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7823419Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7823694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7823844Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7824120Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7824267Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7824544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7824619Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7824887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7825026Z return autotune_select_algorithm( 2025-12-04T09:28:45.7825370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7825481Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7825843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7826009Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7826760Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7826831Z target: aten.mm.default 2025-12-04T09:28:45.7826896Z args[0]: TensorBox( 2025-12-04T09:28:45.7826980Z ReinterpretView( 2025-12-04T09:28:45.7827039Z StorageBox( 2025-12-04T09:28:45.7827288Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7827352Z ), 2025-12-04T09:28:45.7827508Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7827590Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7827655Z stack_traces = {, 2025-12-04T09:28:45.7827884Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7827973Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7828032Z , 2025-12-04T09:28:45.7828088Z } 2025-12-04T09:28:45.7828151Z ) 2025-12-04T09:28:45.7828206Z ) 2025-12-04T09:28:45.7828279Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7828531Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7828589Z )) 2025-12-04T09:28:45.7828593Z 2025-12-04T09:28:45.7829025Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7829031Z 2025-12-04T09:28:45.7829034Z 2025-12-04T09:28:45.7829177Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7829434Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7829437Z 2025-12-04T09:28:45.7829604Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7829738Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7829809Z frames [('total', 1)] 2025-12-04T09:28:45.7829880Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7829956Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7830104Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7830169Z graph_break [] 2025-12-04T09:28:45.7830268Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7830399Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7830463Z frames [('total', 1)] 2025-12-04T09:28:45.7830545Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7830676Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7830750Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7830819Z graph_break [] 2025-12-04T09:28:45.7830896Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7831021Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7831088Z frames [('total', 1)] 2025-12-04T09:28:45.7831156Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7831289Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7831406Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7831466Z graph_break [] 2025-12-04T09:28:45.7831548Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7831671Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7831773Z frames [('total', 1)] 2025-12-04T09:28:45.7831941Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7832074Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7832146Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7832211Z graph_break [] 2025-12-04T09:28:45.7832286Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7832411Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7832479Z frames [('total', 1)] 2025-12-04T09:28:45.7832547Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7832679Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7832756Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7832815Z graph_break [] 2025-12-04T09:28:45.7832898Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7833022Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7833087Z frames [('total', 1)] 2025-12-04T09:28:45.7833167Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7833292Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7833365Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7833435Z graph_break [] 2025-12-04T09:28:45.7833512Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7833646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7833709Z frames [('total', 1)] 2025-12-04T09:28:45.7833777Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7833917Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7833995Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7834054Z graph_break [] 2025-12-04T09:28:45.7834140Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7834266Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7834331Z frames [('total', 1)] 2025-12-04T09:28:45.7834409Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7834534Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7834612Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7834671Z graph_break [] 2025-12-04T09:28:45.7834748Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7834881Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7834942Z frames [('total', 1)] 2025-12-04T09:28:45.7835013Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7835143Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7835215Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7835278Z graph_break [] 2025-12-04T09:28:45.7835362Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7835485Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7835569Z frames [('total', 1)] 2025-12-04T09:28:45.7835641Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7835771Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7835853Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7835913Z graph_break [] 2025-12-04T09:28:45.7835989Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7836122Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7836184Z frames [('total', 1)] 2025-12-04T09:28:45.7836255Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7836389Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7836461Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7836572Z graph_break [] 2025-12-04T09:28:45.7836650Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7836776Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7836883Z frames [('total', 1)] 2025-12-04T09:28:45.7836952Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7837229Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7837312Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7837372Z graph_break [] 2025-12-04T09:28:45.7837449Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7837580Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7837643Z frames [('total', 1)] 2025-12-04T09:28:45.7837717Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7837845Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7837917Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7837988Z graph_break [] 2025-12-04T09:28:45.7838064Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7838225Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7838311Z Traceback (most recent call last): 2025-12-04T09:28:45.7838642Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7838698Z f(a, b) 2025-12-04T09:28:45.7838999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7839153Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7839464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7839584Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7839883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7839985Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7840299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7840505Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7840813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7840886Z graph.run(*example_inputs) 2025-12-04T09:28:45.7841132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7841202Z return super().run(*args) 2025-12-04T09:28:45.7841438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7841537Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7841792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7841869Z result = super().run_node(n) 2025-12-04T09:28:45.7842126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7842231Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7842511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7842661Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7842937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7843070Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7843331Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7843455Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7843717Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7843831Z return autotune_select_algorithm( 2025-12-04T09:28:45.7844246Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7844320Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7844619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7844781Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7845528Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7845607Z target: aten.mm.default 2025-12-04T09:28:45.7845670Z args[0]: TensorBox( 2025-12-04T09:28:45.7845745Z ReinterpretView( 2025-12-04T09:28:45.7845806Z StorageBox( 2025-12-04T09:28:45.7846049Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7846112Z ), 2025-12-04T09:28:45.7846264Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7846338Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7846409Z stack_traces = {, 2025-12-04T09:28:45.7846634Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7846710Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7846784Z , 2025-12-04T09:28:45.7846843Z } 2025-12-04T09:28:45.7846901Z ) 2025-12-04T09:28:45.7846962Z ) 2025-12-04T09:28:45.7847036Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7847287Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7847343Z )) 2025-12-04T09:28:45.7847348Z 2025-12-04T09:28:45.7847780Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7847783Z 2025-12-04T09:28:45.7847786Z 2025-12-04T09:28:45.7847929Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7848176Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7848179Z 2025-12-04T09:28:45.7848351Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7848484Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7848551Z frames [('total', 1)] 2025-12-04T09:28:45.7848630Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7848705Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7848842Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7848915Z graph_break [] 2025-12-04T09:28:45.7848999Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7849133Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7849196Z frames [('total', 1)] 2025-12-04T09:28:45.7849274Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7860293Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7860430Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7860502Z graph_break [] 2025-12-04T09:28:45.7860606Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7860762Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7860946Z frames [('total', 1)] 2025-12-04T09:28:45.7861031Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7861182Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7861321Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7861386Z graph_break [] 2025-12-04T09:28:45.7861550Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7861707Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7861777Z frames [('total', 1)] 2025-12-04T09:28:45.7861852Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7861999Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7862078Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7862142Z graph_break [] 2025-12-04T09:28:45.7862238Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7862374Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7862451Z frames [('total', 1)] 2025-12-04T09:28:45.7862525Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7862657Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7862743Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7862805Z graph_break [] 2025-12-04T09:28:45.7862889Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7863029Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7863093Z frames [('total', 1)] 2025-12-04T09:28:45.7863161Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7863291Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7863365Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7863427Z graph_break [] 2025-12-04T09:28:45.7863513Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7863644Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7863716Z frames [('total', 1)] 2025-12-04T09:28:45.7863786Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7863916Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7863997Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7864062Z graph_break [] 2025-12-04T09:28:45.7864144Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7864288Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7864354Z frames [('total', 1)] 2025-12-04T09:28:45.7864425Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7864565Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7864639Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7864706Z graph_break [] 2025-12-04T09:28:45.7864785Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7864912Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7864984Z frames [('total', 1)] 2025-12-04T09:28:45.7865055Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7865184Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7865274Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7865343Z graph_break [] 2025-12-04T09:28:45.7865421Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7865557Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7865621Z frames [('total', 1)] 2025-12-04T09:28:45.7865699Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7865827Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7865900Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7865966Z graph_break [] 2025-12-04T09:28:45.7866044Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7866170Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7866241Z frames [('total', 1)] 2025-12-04T09:28:45.7866366Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7866492Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7866572Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7866671Z graph_break [] 2025-12-04T09:28:45.7866767Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7866989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7867062Z frames [('total', 1)] 2025-12-04T09:28:45.7867136Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7867264Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7867338Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7867405Z graph_break [] 2025-12-04T09:28:45.7867481Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7867605Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7867674Z frames [('total', 1)] 2025-12-04T09:28:45.7867746Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7867887Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7867963Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7868030Z graph_break [] 2025-12-04T09:28:45.7868112Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7868244Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7868308Z frames [('total', 1)] 2025-12-04T09:28:45.7868383Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7868508Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7868580Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7868648Z graph_break [] 2025-12-04T09:28:45.7868724Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7868900Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7868981Z Traceback (most recent call last): 2025-12-04T09:28:45.7869321Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7869387Z f(a, b) 2025-12-04T09:28:45.7869691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7869858Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7870174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7870302Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7870608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7870704Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7871027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7871237Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7871549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7871634Z graph.run(*example_inputs) 2025-12-04T09:28:45.7871880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7871955Z return super().run(*args) 2025-12-04T09:28:45.7872196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7872281Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7872540Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7872625Z result = super().run_node(n) 2025-12-04T09:28:45.7872880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7873040Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7873316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7873573Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7873848Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7873983Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7874240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7874321Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7874583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7874672Z return autotune_select_algorithm( 2025-12-04T09:28:45.7875034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7875121Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7875428Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7875591Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7876350Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7876423Z target: aten.mm.default 2025-12-04T09:28:45.7876488Z args[0]: TensorBox( 2025-12-04T09:28:45.7876564Z ReinterpretView( 2025-12-04T09:28:45.7876627Z StorageBox( 2025-12-04T09:28:45.7876869Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7876934Z ), 2025-12-04T09:28:45.7877095Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7877185Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7877251Z stack_traces = {, 2025-12-04T09:28:45.7877484Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7877564Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7877622Z , 2025-12-04T09:28:45.7877682Z } 2025-12-04T09:28:45.7877743Z ) 2025-12-04T09:28:45.7877797Z ) 2025-12-04T09:28:45.7877872Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7878116Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7878172Z )) 2025-12-04T09:28:45.7878177Z 2025-12-04T09:28:45.7878602Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7878616Z 2025-12-04T09:28:45.7878619Z 2025-12-04T09:28:45.7878758Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7879008Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7879012Z 2025-12-04T09:28:45.7879195Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7879335Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7879400Z frames [('total', 1)] 2025-12-04T09:28:45.7879479Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7879555Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7879697Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7879810Z graph_break [] 2025-12-04T09:28:45.7879892Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7880025Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7880125Z frames [('total', 1)] 2025-12-04T09:28:45.7880195Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7880404Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7880482Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7880542Z graph_break [] 2025-12-04T09:28:45.7880627Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7880754Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7880822Z frames [('total', 1)] 2025-12-04T09:28:45.7880893Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7881019Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7881098Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7881163Z graph_break [] 2025-12-04T09:28:45.7881242Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7881375Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7881444Z frames [('total', 1)] 2025-12-04T09:28:45.7881511Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7881647Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7881723Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7881790Z graph_break [] 2025-12-04T09:28:45.7881869Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7881994Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7882064Z frames [('total', 1)] 2025-12-04T09:28:45.7882134Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7882259Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7882339Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7882401Z graph_break [] 2025-12-04T09:28:45.7882478Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7882610Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7882675Z frames [('total', 1)] 2025-12-04T09:28:45.7882754Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7882896Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7882974Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7883041Z graph_break [] 2025-12-04T09:28:45.7883118Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7883242Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7883312Z frames [('total', 1)] 2025-12-04T09:28:45.7883382Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7883506Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7883585Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7883645Z graph_break [] 2025-12-04T09:28:45.7883722Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7883857Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7883919Z frames [('total', 1)] 2025-12-04T09:28:45.7883996Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7884124Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7884198Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7884266Z graph_break [] 2025-12-04T09:28:45.7884343Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7884468Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7884536Z frames [('total', 1)] 2025-12-04T09:28:45.7884608Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7884739Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7884817Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7884878Z graph_break [] 2025-12-04T09:28:45.7885004Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7885128Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7885193Z frames [('total', 1)] 2025-12-04T09:28:45.7885302Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7885492Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7885568Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7885633Z graph_break [] 2025-12-04T09:28:45.7885708Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7885832Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7885900Z frames [('total', 1)] 2025-12-04T09:28:45.7885971Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7886100Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7886173Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7886233Z graph_break [] 2025-12-04T09:28:45.7886316Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7886438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7886500Z frames [('total', 1)] 2025-12-04T09:28:45.7886586Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7886713Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7886790Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7886858Z graph_break [] 2025-12-04T09:28:45.7886933Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7887063Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7887125Z frames [('total', 1)] 2025-12-04T09:28:45.7887193Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7887325Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7887399Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7887458Z graph_break [] 2025-12-04T09:28:45.7887539Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7887666Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7887728Z frames [('total', 1)] 2025-12-04T09:28:45.7887805Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7887931Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7888013Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7888074Z graph_break [] 2025-12-04T09:28:45.7888151Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7888282Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7888344Z frames [('total', 1)] 2025-12-04T09:28:45.7888412Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7888542Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7888615Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7888675Z graph_break [] 2025-12-04T09:28:45.7888758Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7888923Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7889005Z Traceback (most recent call last): 2025-12-04T09:28:45.7889319Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7889379Z f(a, b) 2025-12-04T09:28:45.7889679Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7889832Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7890131Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7890255Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7890549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7890693Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7891007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7891254Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7891631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7891718Z graph.run(*example_inputs) 2025-12-04T09:28:45.7891966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7892038Z return super().run(*args) 2025-12-04T09:28:45.7892269Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7892355Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7892606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7892681Z result = super().run_node(n) 2025-12-04T09:28:45.7892933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7893039Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7893420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7893569Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7893832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7893969Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7894227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7894301Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7894570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7894651Z return autotune_select_algorithm( 2025-12-04T09:28:45.7895005Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7895078Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7895361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7895531Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7896279Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7896363Z target: aten.mm.default 2025-12-04T09:28:45.7896425Z args[0]: TensorBox( 2025-12-04T09:28:45.7896490Z ReinterpretView( 2025-12-04T09:28:45.7896559Z StorageBox( 2025-12-04T09:28:45.7896825Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7896897Z ), 2025-12-04T09:28:45.7897079Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7897167Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7897251Z stack_traces = {, 2025-12-04T09:28:45.7897525Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7897617Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7897694Z , 2025-12-04T09:28:45.7897759Z } 2025-12-04T09:28:45.7897826Z ) 2025-12-04T09:28:45.7897898Z ) 2025-12-04T09:28:45.7897985Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7898327Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7898398Z )) 2025-12-04T09:28:45.7898403Z 2025-12-04T09:28:45.7898972Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7898977Z 2025-12-04T09:28:45.7898980Z 2025-12-04T09:28:45.7899124Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7899367Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7899371Z 2025-12-04T09:28:45.7899531Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7899666Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7899729Z frames [('total', 1)] 2025-12-04T09:28:45.7899808Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7899884Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7900017Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7900088Z graph_break [] 2025-12-04T09:28:45.7900170Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7900299Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7900369Z frames [('total', 1)] 2025-12-04T09:28:45.7900436Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7900564Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7900642Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7900703Z graph_break [] 2025-12-04T09:28:45.7900786Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7900914Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7900977Z frames [('total', 1)] 2025-12-04T09:28:45.7901050Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7901179Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7901252Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7901320Z graph_break [] 2025-12-04T09:28:45.7901400Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7901527Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7901609Z frames [('total', 1)] 2025-12-04T09:28:45.7901678Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7901808Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7901881Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7901941Z graph_break [] 2025-12-04T09:28:45.7902025Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7902150Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7902213Z frames [('total', 1)] 2025-12-04T09:28:45.7902287Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7902415Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7902488Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7902554Z graph_break [] 2025-12-04T09:28:45.7902631Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7902767Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7902829Z frames [('total', 1)] 2025-12-04T09:28:45.7902896Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7903027Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7903100Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7903160Z graph_break [] 2025-12-04T09:28:45.7903245Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7903372Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7903436Z frames [('total', 1)] 2025-12-04T09:28:45.7903510Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7903693Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7903773Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7903834Z graph_break [] 2025-12-04T09:28:45.7903950Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7904148Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7904212Z frames [('total', 1)] 2025-12-04T09:28:45.7904282Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7904414Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7904489Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7904856Z graph_break [] 2025-12-04T09:28:45.7904980Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7905111Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7905180Z frames [('total', 1)] 2025-12-04T09:28:45.7905249Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7905378Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7905455Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7905515Z graph_break [] 2025-12-04T09:28:45.7905592Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7905726Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7905798Z frames [('total', 1)] 2025-12-04T09:28:45.7905866Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7906000Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7906075Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7906149Z graph_break [] 2025-12-04T09:28:45.7906229Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7906356Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7906426Z frames [('total', 1)] 2025-12-04T09:28:45.7906496Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7906648Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7906750Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7906824Z graph_break [] 2025-12-04T09:28:45.7906915Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7907077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7907160Z frames [('total', 1)] 2025-12-04T09:28:45.7907244Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7907404Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7907494Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7907571Z graph_break [] 2025-12-04T09:28:45.7907661Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7907812Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7907910Z frames [('total', 1)] 2025-12-04T09:28:45.7907996Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7908145Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7908241Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7908310Z graph_break [] 2025-12-04T09:28:45.7908399Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7908556Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7908636Z frames [('total', 1)] 2025-12-04T09:28:45.7908721Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7908861Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7908932Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7908998Z graph_break [] 2025-12-04T09:28:45.7909074Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7909198Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7909267Z frames [('total', 1)] 2025-12-04T09:28:45.7909336Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7909459Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7909698Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7909759Z graph_break [] 2025-12-04T09:28:45.7909843Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7910030Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7910090Z frames [('total', 1)] 2025-12-04T09:28:45.7910274Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7910401Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7910474Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7910551Z graph_break [] 2025-12-04T09:28:45.7910629Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7910794Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7910874Z Traceback (most recent call last): 2025-12-04T09:28:45.7911194Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7911259Z f(a, b) 2025-12-04T09:28:45.7911550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7911704Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7912015Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7912131Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7912430Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7912521Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7912834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7913034Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7913338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7913411Z graph.run(*example_inputs) 2025-12-04T09:28:45.7913660Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7913729Z return super().run(*args) 2025-12-04T09:28:45.7913966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7914044Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7914295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7914372Z result = super().run_node(n) 2025-12-04T09:28:45.7914619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7914724Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7915001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7915148Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7915428Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7915562Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7915819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7915910Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7916173Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7916259Z return autotune_select_algorithm( 2025-12-04T09:28:45.7916597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7916767Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7917109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7917430Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7918311Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7918402Z target: aten.mm.default 2025-12-04T09:28:45.7918476Z args[0]: TensorBox( 2025-12-04T09:28:45.7918553Z ReinterpretView( 2025-12-04T09:28:45.7918613Z StorageBox( 2025-12-04T09:28:45.7918853Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7918919Z ), 2025-12-04T09:28:45.7919066Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7919142Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7919214Z stack_traces = {, 2025-12-04T09:28:45.7919443Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7919527Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7919585Z , 2025-12-04T09:28:45.7919642Z } 2025-12-04T09:28:45.7919704Z ) 2025-12-04T09:28:45.7919759Z ) 2025-12-04T09:28:45.7919832Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7920074Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7920130Z )) 2025-12-04T09:28:45.7920133Z 2025-12-04T09:28:45.7920557Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7920570Z 2025-12-04T09:28:45.7920573Z 2025-12-04T09:28:45.7920713Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7920965Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7920969Z 2025-12-04T09:28:45.7921140Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7921271Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7921357Z frames [('total', 1)] 2025-12-04T09:28:45.7921431Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7921505Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7921648Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7921710Z graph_break [] 2025-12-04T09:28:45.7921792Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7921927Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7921990Z frames [('total', 1)] 2025-12-04T09:28:45.7922059Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7922195Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7922276Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7922344Z graph_break [] 2025-12-04T09:28:45.7922428Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7922554Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7922621Z frames [('total', 1)] 2025-12-04T09:28:45.7922688Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7922812Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7922891Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7922951Z graph_break [] 2025-12-04T09:28:45.7923027Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7923206Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7923268Z frames [('total', 1)] 2025-12-04T09:28:45.7923338Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7923508Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7923581Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7923714Z graph_break [] 2025-12-04T09:28:45.7923791Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7923915Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7923984Z frames [('total', 1)] 2025-12-04T09:28:45.7924054Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7924187Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7924259Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7924320Z graph_break [] 2025-12-04T09:28:45.7924407Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7924534Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7924598Z frames [('total', 1)] 2025-12-04T09:28:45.7924673Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7924799Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7924876Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7924941Z graph_break [] 2025-12-04T09:28:45.7925017Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7925149Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7925210Z frames [('total', 1)] 2025-12-04T09:28:45.7925280Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7925412Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7925484Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7925554Z graph_break [] 2025-12-04T09:28:45.7925642Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7925765Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7925840Z frames [('total', 1)] 2025-12-04T09:28:45.7925906Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7926029Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7926110Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7926172Z graph_break [] 2025-12-04T09:28:45.7926249Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7926379Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7926444Z frames [('total', 1)] 2025-12-04T09:28:45.7926513Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7926645Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7926718Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7926780Z graph_break [] 2025-12-04T09:28:45.7926864Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7926989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7927061Z frames [('total', 1)] 2025-12-04T09:28:45.7927135Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7927258Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7927339Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7927400Z graph_break [] 2025-12-04T09:28:45.7927483Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7927625Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7927690Z frames [('total', 1)] 2025-12-04T09:28:45.7927759Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7927894Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7927966Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7928033Z graph_break [] 2025-12-04T09:28:45.7928109Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7928233Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7928353Z frames [('total', 1)] 2025-12-04T09:28:45.7928422Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7928547Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7928660Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7928720Z graph_break [] 2025-12-04T09:28:45.7928862Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7928991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7929053Z frames [('total', 1)] 2025-12-04T09:28:45.7929127Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7929251Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7929323Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7929391Z graph_break [] 2025-12-04T09:28:45.7929466Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7929590Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7929662Z frames [('total', 1)] 2025-12-04T09:28:45.7929729Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7929853Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7929933Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7929993Z graph_break [] 2025-12-04T09:28:45.7930079Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7930204Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7930265Z frames [('total', 1)] 2025-12-04T09:28:45.7930337Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7930463Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7930534Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7930599Z graph_break [] 2025-12-04T09:28:45.7930675Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7930810Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7930884Z frames [('total', 1)] 2025-12-04T09:28:45.7930953Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7931082Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7931158Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7931217Z graph_break [] 2025-12-04T09:28:45.7931303Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7931426Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7931487Z frames [('total', 1)] 2025-12-04T09:28:45.7931562Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7931686Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7931760Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7931826Z graph_break [] 2025-12-04T09:28:45.7931902Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7932072Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7932150Z Traceback (most recent call last): 2025-12-04T09:28:45.7932466Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7932533Z f(a, b) 2025-12-04T09:28:45.7932821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7932971Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7933379Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7933498Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7933805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7933896Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7934207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7934458Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7934797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7934960Z graph.run(*example_inputs) 2025-12-04T09:28:45.7935197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7935269Z return super().run(*args) 2025-12-04T09:28:45.7935510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7935588Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7935839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7935918Z result = super().run_node(n) 2025-12-04T09:28:45.7936167Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7936275Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7936551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7936698Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7936971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7937100Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7937364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7937444Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7937709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7937797Z return autotune_select_algorithm( 2025-12-04T09:28:45.7938140Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7938216Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7938517Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7938682Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7939429Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7939502Z target: aten.mm.default 2025-12-04T09:28:45.7939565Z args[0]: TensorBox( 2025-12-04T09:28:45.7939641Z ReinterpretView( 2025-12-04T09:28:45.7939701Z StorageBox( 2025-12-04T09:28:45.7939943Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7940010Z ), 2025-12-04T09:28:45.7940166Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7940250Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7940319Z stack_traces = {, 2025-12-04T09:28:45.7940546Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7940632Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7940689Z , 2025-12-04T09:28:45.7940746Z } 2025-12-04T09:28:45.7940817Z ) 2025-12-04T09:28:45.7940875Z ) 2025-12-04T09:28:45.7940950Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7941190Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7941289Z )) 2025-12-04T09:28:45.7941293Z 2025-12-04T09:28:45.7941718Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7941756Z 2025-12-04T09:28:45.7941759Z 2025-12-04T09:28:45.7941963Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7942210Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7942214Z 2025-12-04T09:28:45.7942383Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7942518Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7942586Z frames [('total', 1)] 2025-12-04T09:28:45.7942657Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7942731Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7942873Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7942935Z graph_break [] 2025-12-04T09:28:45.7943020Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7943153Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7943219Z frames [('total', 1)] 2025-12-04T09:28:45.7943292Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7943427Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7943499Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7943564Z graph_break [] 2025-12-04T09:28:45.7943643Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7943769Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7943835Z frames [('total', 1)] 2025-12-04T09:28:45.7943904Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7944030Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7944111Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7944170Z graph_break [] 2025-12-04T09:28:45.7944247Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7944376Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7944442Z frames [('total', 1)] 2025-12-04T09:28:45.7944522Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7944648Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7944721Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7944787Z graph_break [] 2025-12-04T09:28:45.7944863Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7944987Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7945054Z frames [('total', 1)] 2025-12-04T09:28:45.7945123Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7945246Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7945331Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7945391Z graph_break [] 2025-12-04T09:28:45.7945472Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7945597Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7945663Z frames [('total', 1)] 2025-12-04T09:28:45.7945753Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7945882Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7945955Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7946022Z graph_break [] 2025-12-04T09:28:45.7946099Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7946223Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7946294Z frames [('total', 1)] 2025-12-04T09:28:45.7946362Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7946494Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7946569Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7946676Z graph_break [] 2025-12-04T09:28:45.7946760Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7946882Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7946980Z frames [('total', 1)] 2025-12-04T09:28:45.7947058Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7947328Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7947405Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7947471Z graph_break [] 2025-12-04T09:28:45.7947549Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7947678Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7947741Z frames [('total', 1)] 2025-12-04T09:28:45.7947808Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7947940Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7948013Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7948077Z graph_break [] 2025-12-04T09:28:45.7948159Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7948282Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7948348Z frames [('total', 1)] 2025-12-04T09:28:45.7948433Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7948566Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7948647Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7948708Z graph_break [] 2025-12-04T09:28:45.7948786Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7948918Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7948980Z frames [('total', 1)] 2025-12-04T09:28:45.7949050Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7949182Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7949254Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7949317Z graph_break [] 2025-12-04T09:28:45.7949404Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7949527Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7949601Z frames [('total', 1)] 2025-12-04T09:28:45.7949671Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7949800Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7949878Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7949938Z graph_break [] 2025-12-04T09:28:45.7950025Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7950156Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7950219Z frames [('total', 1)] 2025-12-04T09:28:45.7950287Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7950421Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7950495Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7950557Z graph_break [] 2025-12-04T09:28:45.7950646Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7950772Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7950844Z frames [('total', 1)] 2025-12-04T09:28:45.7950913Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7951043Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7951122Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7951184Z graph_break [] 2025-12-04T09:28:45.7951269Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7951394Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7951456Z frames [('total', 1)] 2025-12-04T09:28:45.7951532Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7951658Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7951731Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7951797Z graph_break [] 2025-12-04T09:28:45.7951923Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7952050Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7952120Z frames [('total', 1)] 2025-12-04T09:28:45.7952226Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7952432Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7952512Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7952575Z graph_break [] 2025-12-04T09:28:45.7952659Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7952784Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7952849Z frames [('total', 1)] 2025-12-04T09:28:45.7952922Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7953047Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7953119Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7953185Z graph_break [] 2025-12-04T09:28:45.7953263Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7953392Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7953458Z frames [('total', 1)] 2025-12-04T09:28:45.7953528Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7953666Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7953738Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7953798Z graph_break [] 2025-12-04T09:28:45.7953885Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7954047Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7954121Z Traceback (most recent call last): 2025-12-04T09:28:45.7954444Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7954501Z f(a, b) 2025-12-04T09:28:45.7954799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7954952Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7955254Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7955392Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7955697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7955788Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7956101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7956296Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7956605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7956682Z graph.run(*example_inputs) 2025-12-04T09:28:45.7956918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7956995Z return super().run(*args) 2025-12-04T09:28:45.7957237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7957323Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7957577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7957650Z result = super().run_node(n) 2025-12-04T09:28:45.7957910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7958014Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7958284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7958488Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7958748Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7959003Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7959259Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7959332Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7959599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7959678Z return autotune_select_algorithm( 2025-12-04T09:28:45.7960016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7960095Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7960384Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7960554Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7961305Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7961386Z target: aten.mm.default 2025-12-04T09:28:45.7961458Z args[0]: TensorBox( 2025-12-04T09:28:45.7961524Z ReinterpretView( 2025-12-04T09:28:45.7961590Z StorageBox( 2025-12-04T09:28:45.7961832Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7961889Z ), 2025-12-04T09:28:45.7962045Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7962122Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7962185Z stack_traces = {, 2025-12-04T09:28:45.7962416Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7962496Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7962558Z , 2025-12-04T09:28:45.7962625Z } 2025-12-04T09:28:45.7962681Z ) 2025-12-04T09:28:45.7962741Z ) 2025-12-04T09:28:45.7962818Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7963050Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7963109Z )) 2025-12-04T09:28:45.7963113Z 2025-12-04T09:28:45.7963536Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7963540Z 2025-12-04T09:28:45.7963544Z 2025-12-04T09:28:45.7963684Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7963928Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7963933Z 2025-12-04T09:28:45.7964099Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7964238Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7964304Z frames [('total', 1)] 2025-12-04T09:28:45.7964375Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7964458Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7964594Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7964662Z graph_break [] 2025-12-04T09:28:45.7964745Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7964882Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7965899Z frames [('total', 1)] 2025-12-04T09:28:45.7965969Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7966103Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7966186Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7966287Z graph_break [] 2025-12-04T09:28:45.7966368Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7966578Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7966647Z frames [('total', 1)] 2025-12-04T09:28:45.7966724Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7966852Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7966925Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7966991Z graph_break [] 2025-12-04T09:28:45.7967072Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7967198Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7967267Z frames [('total', 1)] 2025-12-04T09:28:45.7967339Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7967465Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7967542Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7967604Z graph_break [] 2025-12-04T09:28:45.7967686Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7967816Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7967879Z frames [('total', 1)] 2025-12-04T09:28:45.7967958Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7968083Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7968154Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7968220Z graph_break [] 2025-12-04T09:28:45.7968294Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7968419Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7968487Z frames [('total', 1)] 2025-12-04T09:28:45.7968559Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7968700Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7968774Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7968837Z graph_break [] 2025-12-04T09:28:45.7968917Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7969046Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7969108Z frames [('total', 1)] 2025-12-04T09:28:45.7969182Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7969306Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7969378Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7969445Z graph_break [] 2025-12-04T09:28:45.7969519Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7969643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7969709Z frames [('total', 1)] 2025-12-04T09:28:45.7969781Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7969911Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7969984Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7970046Z graph_break [] 2025-12-04T09:28:45.7970127Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7970255Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7970319Z frames [('total', 1)] 2025-12-04T09:28:45.7970393Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7970518Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7970590Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7970655Z graph_break [] 2025-12-04T09:28:45.7970731Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7970872Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7970937Z frames [('total', 1)] 2025-12-04T09:28:45.7971006Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7971183Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7971256Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7971317Z graph_break [] 2025-12-04T09:28:45.7971436Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7971634Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7971701Z frames [('total', 1)] 2025-12-04T09:28:45.7971776Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7971905Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7971982Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7972041Z graph_break [] 2025-12-04T09:28:45.7972117Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7972247Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7972309Z frames [('total', 1)] 2025-12-04T09:28:45.7972377Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7972511Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7972584Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7972644Z graph_break [] 2025-12-04T09:28:45.7972732Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7972861Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7972930Z frames [('total', 1)] 2025-12-04T09:28:45.7973000Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7973236Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7973318Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7973378Z graph_break [] 2025-12-04T09:28:45.7973455Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7973586Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7973648Z frames [('total', 1)] 2025-12-04T09:28:45.7973716Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7973854Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7973928Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7973997Z graph_break [] 2025-12-04T09:28:45.7974074Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7974201Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7974278Z frames [('total', 1)] 2025-12-04T09:28:45.7974347Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7974474Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7974554Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7974616Z graph_break [] 2025-12-04T09:28:45.7974692Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7974829Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7974892Z frames [('total', 1)] 2025-12-04T09:28:45.7974972Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7975099Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7975173Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7975243Z graph_break [] 2025-12-04T09:28:45.7975320Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7975449Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7975526Z frames [('total', 1)] 2025-12-04T09:28:45.7975595Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7975722Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7975805Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7975865Z graph_break [] 2025-12-04T09:28:45.7975949Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7976078Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7976140Z frames [('total', 1)] 2025-12-04T09:28:45.7976212Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7976336Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7976461Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7976526Z graph_break [] 2025-12-04T09:28:45.7976613Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7976780Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7976914Z frames [('total', 1)] 2025-12-04T09:28:45.7976984Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7977118Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7977197Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7977259Z graph_break [] 2025-12-04T09:28:45.7977339Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7977503Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.7977579Z Traceback (most recent call last): 2025-12-04T09:28:45.7977898Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7977958Z f(a, b) 2025-12-04T09:28:45.7978243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.7978403Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.7978707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.7978844Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.7979140Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.7979231Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.7979549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.7979744Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.7980053Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.7980127Z graph.run(*example_inputs) 2025-12-04T09:28:45.7980368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.7980444Z return super().run(*args) 2025-12-04T09:28:45.7980675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.7980753Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.7981012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.7981092Z result = super().run_node(n) 2025-12-04T09:28:45.7981348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.7981455Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.7981721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.7981876Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.7982142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.7982274Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.7982537Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.7982609Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.7982875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.7982954Z return autotune_select_algorithm( 2025-12-04T09:28:45.7983296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.7983421Z return cache(*args, **kwargs) 2025-12-04T09:28:45.7983707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.7983993Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.7984738Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.7984808Z target: aten.mm.default 2025-12-04T09:28:45.7984881Z args[0]: TensorBox( 2025-12-04T09:28:45.7984948Z ReinterpretView( 2025-12-04T09:28:45.7985008Z StorageBox( 2025-12-04T09:28:45.7985257Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.7985320Z ), 2025-12-04T09:28:45.7985484Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.7985561Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.7985625Z stack_traces = {, 2025-12-04T09:28:45.7985861Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.7985937Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.7985996Z , 2025-12-04T09:28:45.7986059Z } 2025-12-04T09:28:45.7986113Z ) 2025-12-04T09:28:45.7986168Z ) 2025-12-04T09:28:45.7986247Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.7986479Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.7986540Z )) 2025-12-04T09:28:45.7986544Z 2025-12-04T09:28:45.7986966Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.7986972Z 2025-12-04T09:28:45.7986975Z 2025-12-04T09:28:45.7987112Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.7987374Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.7987377Z 2025-12-04T09:28:45.7987537Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.7987675Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7987741Z frames [('total', 1)] 2025-12-04T09:28:45.7987814Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7987895Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7988030Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7988094Z graph_break [] 2025-12-04T09:28:45.7988182Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7988309Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7988378Z frames [('total', 1)] 2025-12-04T09:28:45.7988446Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7988577Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7988660Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7988732Z graph_break [] 2025-12-04T09:28:45.7988814Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7988951Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7989013Z frames [('total', 1)] 2025-12-04T09:28:45.7989082Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7989218Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7989289Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7989355Z graph_break [] 2025-12-04T09:28:45.7989432Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7989603Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7989669Z frames [('total', 1)] 2025-12-04T09:28:45.7989737Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7989900Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7990040Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7990101Z graph_break [] 2025-12-04T09:28:45.7990178Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7990309Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7990372Z frames [('total', 1)] 2025-12-04T09:28:45.7990444Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7990568Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7990641Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7990704Z graph_break [] 2025-12-04T09:28:45.7990779Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7990905Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7990971Z frames [('total', 1)] 2025-12-04T09:28:45.7991039Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7991166Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7991250Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7991310Z graph_break [] 2025-12-04T09:28:45.7991386Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7991514Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7991587Z frames [('total', 1)] 2025-12-04T09:28:45.7991664Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7991789Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7991861Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7991924Z graph_break [] 2025-12-04T09:28:45.7992000Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7992126Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7992194Z frames [('total', 1)] 2025-12-04T09:28:45.7992262Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7992388Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7992468Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7992531Z graph_break [] 2025-12-04T09:28:45.7992612Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7992736Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7992797Z frames [('total', 1)] 2025-12-04T09:28:45.7992869Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7992994Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7993065Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7993132Z graph_break [] 2025-12-04T09:28:45.7993209Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7993334Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7993403Z frames [('total', 1)] 2025-12-04T09:28:45.7993471Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7993601Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7993676Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7993741Z graph_break [] 2025-12-04T09:28:45.7993830Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7993959Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7994024Z frames [('total', 1)] 2025-12-04T09:28:45.7994099Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7994223Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7994295Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7994365Z graph_break [] 2025-12-04T09:28:45.7994442Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7994573Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7994683Z frames [('total', 1)] 2025-12-04T09:28:45.7994751Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7994881Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7994994Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7995120Z graph_break [] 2025-12-04T09:28:45.7995205Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7995330Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7995392Z frames [('total', 1)] 2025-12-04T09:28:45.7995468Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7995593Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7995673Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7995732Z graph_break [] 2025-12-04T09:28:45.7995815Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7995957Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7996026Z frames [('total', 1)] 2025-12-04T09:28:45.7996096Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7996231Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7996308Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7996369Z graph_break [] 2025-12-04T09:28:45.7996460Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7996585Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7996655Z frames [('total', 1)] 2025-12-04T09:28:45.7996724Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7996852Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7996932Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7996991Z graph_break [] 2025-12-04T09:28:45.7997070Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7997200Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7997266Z frames [('total', 1)] 2025-12-04T09:28:45.7997336Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7997472Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7997550Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7997616Z graph_break [] 2025-12-04T09:28:45.7997697Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7997826Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7997898Z frames [('total', 1)] 2025-12-04T09:28:45.7997967Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7998090Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7998169Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7998231Z graph_break [] 2025-12-04T09:28:45.7998307Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7998440Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7998506Z frames [('total', 1)] 2025-12-04T09:28:45.7998578Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7998703Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7998777Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7998843Z graph_break [] 2025-12-04T09:28:45.7998923Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7999057Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7999125Z frames [('total', 1)] 2025-12-04T09:28:45.7999193Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7999317Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.7999395Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.7999454Z graph_break [] 2025-12-04T09:28:45.7999532Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.7999659Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.7999769Z frames [('total', 1)] 2025-12-04T09:28:45.7999841Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.7999965Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8000074Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8000137Z graph_break [] 2025-12-04T09:28:45.8000283Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8000447Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8000529Z Traceback (most recent call last): 2025-12-04T09:28:45.8000842Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8000907Z f(a, b) 2025-12-04T09:28:45.8001191Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8001341Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8001651Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8001768Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8002069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8002169Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8002480Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8002678Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8002981Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8003065Z graph.run(*example_inputs) 2025-12-04T09:28:45.8003310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8003381Z return super().run(*args) 2025-12-04T09:28:45.8003618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8003702Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8003954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8004031Z result = super().run_node(n) 2025-12-04T09:28:45.8004276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8004379Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8004994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8005151Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8005421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8005557Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8005816Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8005899Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8006159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8006238Z return autotune_select_algorithm( 2025-12-04T09:28:45.8006582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8006654Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8006943Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8007235Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8008094Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8008225Z target: aten.mm.default 2025-12-04T09:28:45.8008289Z args[0]: TensorBox( 2025-12-04T09:28:45.8008361Z ReinterpretView( 2025-12-04T09:28:45.8008421Z StorageBox( 2025-12-04T09:28:45.8008664Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8008727Z ), 2025-12-04T09:28:45.8008875Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8008949Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8009024Z stack_traces = {, 2025-12-04T09:28:45.8009264Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8009340Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8009404Z , 2025-12-04T09:28:45.8009463Z } 2025-12-04T09:28:45.8009526Z ) 2025-12-04T09:28:45.8009580Z ) 2025-12-04T09:28:45.8009658Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8009898Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8009954Z )) 2025-12-04T09:28:45.8009957Z 2025-12-04T09:28:45.8010376Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8010381Z 2025-12-04T09:28:45.8010391Z 2025-12-04T09:28:45.8010528Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8010771Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8010776Z 2025-12-04T09:28:45.8010947Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8011082Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8011149Z frames [('total', 1)] 2025-12-04T09:28:45.8011234Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8011316Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8011456Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8011520Z graph_break [] 2025-12-04T09:28:45.8011603Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8011737Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8011798Z frames [('total', 1)] 2025-12-04T09:28:45.8011868Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8012001Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8012081Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8012143Z graph_break [] 2025-12-04T09:28:45.8012230Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8012353Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8012422Z frames [('total', 1)] 2025-12-04T09:28:45.8012495Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8012619Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8012698Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8012758Z graph_break [] 2025-12-04T09:28:45.8012837Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8012970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8013031Z frames [('total', 1)] 2025-12-04T09:28:45.8013183Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8013327Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8013452Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8013520Z graph_break [] 2025-12-04T09:28:45.8013597Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8013723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8013829Z frames [('total', 1)] 2025-12-04T09:28:45.8013961Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8014089Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8014171Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8014230Z graph_break [] 2025-12-04T09:28:45.8014305Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8014434Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8014496Z frames [('total', 1)] 2025-12-04T09:28:45.8014563Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8014693Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8014770Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8014836Z graph_break [] 2025-12-04T09:28:45.8014912Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8015034Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8015105Z frames [('total', 1)] 2025-12-04T09:28:45.8015173Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8015302Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8015382Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8015444Z graph_break [] 2025-12-04T09:28:45.8015523Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8015652Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8015714Z frames [('total', 1)] 2025-12-04T09:28:45.8015788Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8015911Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8015984Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8016052Z graph_break [] 2025-12-04T09:28:45.8016131Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8016255Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8016327Z frames [('total', 1)] 2025-12-04T09:28:45.8016394Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8016522Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8016602Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8016662Z graph_break [] 2025-12-04T09:28:45.8016743Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8016870Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8016932Z frames [('total', 1)] 2025-12-04T09:28:45.8017007Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8017130Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8017202Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8017271Z graph_break [] 2025-12-04T09:28:45.8017347Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8017489Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8017565Z frames [('total', 1)] 2025-12-04T09:28:45.8017634Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8017769Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8017841Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8017900Z graph_break [] 2025-12-04T09:28:45.8017981Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8018105Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8018167Z frames [('total', 1)] 2025-12-04T09:28:45.8018245Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8018368Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8018442Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8018557Z graph_break [] 2025-12-04T09:28:45.8018634Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8018768Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8018866Z frames [('total', 1)] 2025-12-04T09:28:45.8018935Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8019133Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8019209Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8019280Z graph_break [] 2025-12-04T09:28:45.8019366Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8019493Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8019558Z frames [('total', 1)] 2025-12-04T09:28:45.8019635Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8019757Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8019834Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8019900Z graph_break [] 2025-12-04T09:28:45.8019977Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8020109Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8020176Z frames [('total', 1)] 2025-12-04T09:28:45.8020244Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8020381Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8020459Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8020522Z graph_break [] 2025-12-04T09:28:45.8020608Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8020732Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8020800Z frames [('total', 1)] 2025-12-04T09:28:45.8020867Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8020993Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8021070Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8021130Z graph_break [] 2025-12-04T09:28:45.8021210Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8021339Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8021400Z frames [('total', 1)] 2025-12-04T09:28:45.8021471Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8021603Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8021675Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8021740Z graph_break [] 2025-12-04T09:28:45.8021816Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8021940Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8022008Z frames [('total', 1)] 2025-12-04T09:28:45.8022076Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8022201Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8022280Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8022339Z graph_break [] 2025-12-04T09:28:45.8022420Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8022549Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8022611Z frames [('total', 1)] 2025-12-04T09:28:45.8022686Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8022810Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8022887Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8022951Z graph_break [] 2025-12-04T09:28:45.8023026Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8023151Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8023220Z frames [('total', 1)] 2025-12-04T09:28:45.8023287Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8023411Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8023492Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8023551Z graph_break [] 2025-12-04T09:28:45.8023626Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8023809Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8023873Z frames [('total', 1)] 2025-12-04T09:28:45.8023984Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8024109Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8024276Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8024345Z graph_break [] 2025-12-04T09:28:45.8024425Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8024588Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8024671Z Traceback (most recent call last): 2025-12-04T09:28:45.8024984Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8025049Z f(a, b) 2025-12-04T09:28:45.8025336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8025490Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8025792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8025913Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8026210Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8026308Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8026619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8026819Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8027130Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8027206Z graph.run(*example_inputs) 2025-12-04T09:28:45.8027449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8027525Z return super().run(*args) 2025-12-04T09:28:45.8027764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8027843Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8028096Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8028173Z result = super().run_node(n) 2025-12-04T09:28:45.8028421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8028522Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8028803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8028956Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8029231Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8029364Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8029630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8029711Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8029972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8030051Z return autotune_select_algorithm( 2025-12-04T09:28:45.8030402Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8030476Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8030814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8030975Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8031815Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8031895Z target: aten.mm.default 2025-12-04T09:28:45.8031963Z args[0]: TensorBox( 2025-12-04T09:28:45.8032036Z ReinterpretView( 2025-12-04T09:28:45.8032096Z StorageBox( 2025-12-04T09:28:45.8032336Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8032400Z ), 2025-12-04T09:28:45.8032552Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8032627Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8032699Z stack_traces = {, 2025-12-04T09:28:45.8032924Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8033002Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8033065Z , 2025-12-04T09:28:45.8033124Z } 2025-12-04T09:28:45.8033186Z ) 2025-12-04T09:28:45.8033242Z ) 2025-12-04T09:28:45.8033316Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8033553Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8033607Z )) 2025-12-04T09:28:45.8033610Z 2025-12-04T09:28:45.8034032Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8034035Z 2025-12-04T09:28:45.8034045Z 2025-12-04T09:28:45.8034184Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8034426Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8034431Z 2025-12-04T09:28:45.8034596Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8034741Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8034808Z frames [('total', 1)] 2025-12-04T09:28:45.8034884Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8034958Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8035096Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8035155Z graph_break [] 2025-12-04T09:28:45.8035235Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8035365Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8035431Z frames [('total', 1)] 2025-12-04T09:28:45.8035503Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8035637Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8035712Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8035776Z graph_break [] 2025-12-04T09:28:45.8035860Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8035989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8036062Z frames [('total', 1)] 2025-12-04T09:28:45.8036130Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8036259Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8036337Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8036397Z graph_break [] 2025-12-04T09:28:45.8036473Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8036603Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8036666Z frames [('total', 1)] 2025-12-04T09:28:45.8036782Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8036912Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8036985Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8037095Z graph_break [] 2025-12-04T09:28:45.8037171Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8037365Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8037434Z frames [('total', 1)] 2025-12-04T09:28:45.8037512Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8037640Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8037722Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8037782Z graph_break [] 2025-12-04T09:28:45.8037861Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8037990Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8038052Z frames [('total', 1)] 2025-12-04T09:28:45.8038120Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8038253Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8038325Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8038389Z graph_break [] 2025-12-04T09:28:45.8038469Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8038600Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8038670Z frames [('total', 1)] 2025-12-04T09:28:45.8038739Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8038865Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8038947Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8039007Z graph_break [] 2025-12-04T09:28:45.8039087Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8039218Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8039283Z frames [('total', 1)] 2025-12-04T09:28:45.8039358Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8039488Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8039562Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8039628Z graph_break [] 2025-12-04T09:28:45.8039707Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8039836Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8039905Z frames [('total', 1)] 2025-12-04T09:28:45.8039973Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8040099Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8040178Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8040248Z graph_break [] 2025-12-04T09:28:45.8040340Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8040466Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8040528Z frames [('total', 1)] 2025-12-04T09:28:45.8064975Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8065216Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8065306Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8065372Z graph_break [] 2025-12-04T09:28:45.8065474Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8065629Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8065706Z frames [('total', 1)] 2025-12-04T09:28:45.8065782Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8065926Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8066013Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8066081Z graph_break [] 2025-12-04T09:28:45.8066169Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8066312Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8066380Z frames [('total', 1)] 2025-12-04T09:28:45.8066450Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8066679Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8066758Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8066826Z graph_break [] 2025-12-04T09:28:45.8066906Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8067095Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8067245Z frames [('total', 1)] 2025-12-04T09:28:45.8067317Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8067446Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8067528Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8067591Z graph_break [] 2025-12-04T09:28:45.8067668Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8067799Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8067862Z frames [('total', 1)] 2025-12-04T09:28:45.8067938Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8068067Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8068146Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8068215Z graph_break [] 2025-12-04T09:28:45.8068293Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8068426Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8068503Z frames [('total', 1)] 2025-12-04T09:28:45.8068578Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8068711Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8068794Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8068856Z graph_break [] 2025-12-04T09:28:45.8068944Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8069076Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8069141Z frames [('total', 1)] 2025-12-04T09:28:45.8069218Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8069349Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8069427Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8069496Z graph_break [] 2025-12-04T09:28:45.8069575Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8069706Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8069781Z frames [('total', 1)] 2025-12-04T09:28:45.8069852Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8069985Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8070060Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8070120Z graph_break [] 2025-12-04T09:28:45.8070203Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8070332Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8070395Z frames [('total', 1)] 2025-12-04T09:28:45.8070470Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8070596Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8070671Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8070739Z graph_break [] 2025-12-04T09:28:45.8070817Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8070950Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8071015Z frames [('total', 1)] 2025-12-04T09:28:45.8071096Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8071231Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8071305Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8071365Z graph_break [] 2025-12-04T09:28:45.8071451Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8071577Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8071640Z frames [('total', 1)] 2025-12-04T09:28:45.8071717Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8071843Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8071962Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8072028Z graph_break [] 2025-12-04T09:28:45.8072104Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8072239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8072338Z frames [('total', 1)] 2025-12-04T09:28:45.8072481Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8072622Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8072695Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8072756Z graph_break [] 2025-12-04T09:28:45.8072842Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8072968Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8073030Z frames [('total', 1)] 2025-12-04T09:28:45.8073106Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8073231Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8073314Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8073374Z graph_break [] 2025-12-04T09:28:45.8073456Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8073639Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8073722Z Traceback (most recent call last): 2025-12-04T09:28:45.8074058Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8074127Z f(a, b) 2025-12-04T09:28:45.8074439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8074610Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8074928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8075054Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8075365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8075461Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8075784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8075991Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8076296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8076381Z graph.run(*example_inputs) 2025-12-04T09:28:45.8076618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8076691Z return super().run(*args) 2025-12-04T09:28:45.8076930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8077015Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8077274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8077350Z result = super().run_node(n) 2025-12-04T09:28:45.8077602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8077717Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8077984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8078134Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8078421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8078556Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8078890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8078966Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8079283Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8079440Z return autotune_select_algorithm( 2025-12-04T09:28:45.8079782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8079856Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8080151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8080315Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8081069Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8081144Z target: aten.mm.default 2025-12-04T09:28:45.8081211Z args[0]: TensorBox( 2025-12-04T09:28:45.8081283Z ReinterpretView( 2025-12-04T09:28:45.8081348Z StorageBox( 2025-12-04T09:28:45.8081604Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8081662Z ), 2025-12-04T09:28:45.8081816Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8081897Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8081963Z stack_traces = {, 2025-12-04T09:28:45.8082201Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8082293Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8082354Z , 2025-12-04T09:28:45.8082409Z } 2025-12-04T09:28:45.8082471Z ) 2025-12-04T09:28:45.8082527Z ) 2025-12-04T09:28:45.8082603Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8082844Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8082902Z )) 2025-12-04T09:28:45.8082912Z 2025-12-04T09:28:45.8083341Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8083345Z 2025-12-04T09:28:45.8083349Z 2025-12-04T09:28:45.8083485Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8083731Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8083735Z 2025-12-04T09:28:45.8083907Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8084048Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8084122Z frames [('total', 1)] 2025-12-04T09:28:45.8084195Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8084273Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8084416Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8084482Z graph_break [] 2025-12-04T09:28:45.8084569Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8084705Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8084770Z frames [('total', 1)] 2025-12-04T09:28:45.8084839Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8084976Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8085051Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8085115Z graph_break [] 2025-12-04T09:28:45.8085194Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8085368Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8085438Z frames [('total', 1)] 2025-12-04T09:28:45.8085508Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8085638Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8085755Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8085881Z graph_break [] 2025-12-04T09:28:45.8085961Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8086093Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8086158Z frames [('total', 1)] 2025-12-04T09:28:45.8086243Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8086372Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8086445Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8086510Z graph_break [] 2025-12-04T09:28:45.8086589Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8086715Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8086789Z frames [('total', 1)] 2025-12-04T09:28:45.8086859Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8086984Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8087067Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8087140Z graph_break [] 2025-12-04T09:28:45.8087222Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8087346Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8087410Z frames [('total', 1)] 2025-12-04T09:28:45.8087487Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8087611Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8087693Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8087759Z graph_break [] 2025-12-04T09:28:45.8087835Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8087960Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8088029Z frames [('total', 1)] 2025-12-04T09:28:45.8088097Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8088225Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8088304Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8088363Z graph_break [] 2025-12-04T09:28:45.8088448Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8088571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8088635Z frames [('total', 1)] 2025-12-04T09:28:45.8088708Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8088831Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8088907Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8088973Z graph_break [] 2025-12-04T09:28:45.8089049Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8089177Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8089242Z frames [('total', 1)] 2025-12-04T09:28:45.8089310Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8089439Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8089515Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8089575Z graph_break [] 2025-12-04T09:28:45.8089665Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8089792Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8089855Z frames [('total', 1)] 2025-12-04T09:28:45.8089928Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8090053Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8090130Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8090189Z graph_break [] 2025-12-04T09:28:45.8090265Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8090392Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8090503Z frames [('total', 1)] 2025-12-04T09:28:45.8090572Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8090701Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8090809Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8090867Z graph_break [] 2025-12-04T09:28:45.8091015Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8091142Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8091213Z frames [('total', 1)] 2025-12-04T09:28:45.8091281Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8091407Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8091486Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8091546Z graph_break [] 2025-12-04T09:28:45.8091633Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8091766Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8091833Z frames [('total', 1)] 2025-12-04T09:28:45.8091905Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8092037Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8092111Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8092177Z graph_break [] 2025-12-04T09:28:45.8092258Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8092381Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8092459Z frames [('total', 1)] 2025-12-04T09:28:45.8092530Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8092656Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8092735Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8092794Z graph_break [] 2025-12-04T09:28:45.8092870Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8093001Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8093068Z frames [('total', 1)] 2025-12-04T09:28:45.8093241Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8093378Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8093454Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8093520Z graph_break [] 2025-12-04T09:28:45.8093602Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8093726Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8093794Z frames [('total', 1)] 2025-12-04T09:28:45.8093862Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8093987Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8094067Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8094127Z graph_break [] 2025-12-04T09:28:45.8094204Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8094338Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8094405Z frames [('total', 1)] 2025-12-04T09:28:45.8094480Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8094605Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8094677Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8094748Z graph_break [] 2025-12-04T09:28:45.8094824Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8094953Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8095022Z frames [('total', 1)] 2025-12-04T09:28:45.8095095Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8095225Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8095304Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8095363Z graph_break [] 2025-12-04T09:28:45.8095445Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8095569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8095632Z frames [('total', 1)] 2025-12-04T09:28:45.8095759Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8095884Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8095955Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8096058Z graph_break [] 2025-12-04T09:28:45.8096133Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8096336Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8096405Z frames [('total', 1)] 2025-12-04T09:28:45.8096474Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8096603Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8096675Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8096736Z graph_break [] 2025-12-04T09:28:45.8096816Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8096940Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8097012Z frames [('total', 1)] 2025-12-04T09:28:45.8097087Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8097213Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8097288Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8097355Z graph_break [] 2025-12-04T09:28:45.8097431Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8097564Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8097626Z frames [('total', 1)] 2025-12-04T09:28:45.8097693Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8097822Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8097897Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8097956Z graph_break [] 2025-12-04T09:28:45.8098038Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8098173Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8098236Z frames [('total', 1)] 2025-12-04T09:28:45.8098315Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8098442Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8098519Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8098582Z graph_break [] 2025-12-04T09:28:45.8098659Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8098829Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8098906Z Traceback (most recent call last): 2025-12-04T09:28:45.8099223Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8099287Z f(a, b) 2025-12-04T09:28:45.8099578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8099734Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8100039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8100163Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8100465Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8100565Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8100891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8101101Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8101418Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8101500Z graph.run(*example_inputs) 2025-12-04T09:28:45.8101741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8101864Z return super().run(*args) 2025-12-04T09:28:45.8102105Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8102187Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8102610Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8102693Z result = super().run_node(n) 2025-12-04T09:28:45.8102943Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8103055Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8103335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8103486Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8103761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8103899Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8104166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8104240Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8104784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8104932Z return autotune_select_algorithm( 2025-12-04T09:28:45.8105281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8105354Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8105648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8105809Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8106568Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8106643Z target: aten.mm.default 2025-12-04T09:28:45.8106710Z args[0]: TensorBox( 2025-12-04T09:28:45.8106784Z ReinterpretView( 2025-12-04T09:28:45.8106843Z StorageBox( 2025-12-04T09:28:45.8107093Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8107159Z ), 2025-12-04T09:28:45.8107309Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8107388Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8107452Z stack_traces = {, 2025-12-04T09:28:45.8107681Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8107768Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8107825Z , 2025-12-04T09:28:45.8107880Z } 2025-12-04T09:28:45.8107943Z ) 2025-12-04T09:28:45.8108003Z ) 2025-12-04T09:28:45.8108077Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8108325Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8108389Z )) 2025-12-04T09:28:45.8108393Z 2025-12-04T09:28:45.8108824Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8108828Z 2025-12-04T09:28:45.8108831Z 2025-12-04T09:28:45.8108967Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8109212Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8109343Z 2025-12-04T09:28:45.8109511Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8109646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8109779Z frames [('total', 1)] 2025-12-04T09:28:45.8109852Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8110023Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8110173Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8110236Z graph_break [] 2025-12-04T09:28:45.8110319Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8110458Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8110523Z frames [('total', 1)] 2025-12-04T09:28:45.8110599Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8110730Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8110803Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8110878Z graph_break [] 2025-12-04T09:28:45.8110956Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8111087Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8111154Z frames [('total', 1)] 2025-12-04T09:28:45.8111223Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8111360Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8111435Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8111497Z graph_break [] 2025-12-04T09:28:45.8111594Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8111720Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8111782Z frames [('total', 1)] 2025-12-04T09:28:45.8111859Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8111983Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8112062Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8112127Z graph_break [] 2025-12-04T09:28:45.8112203Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8112335Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8112400Z frames [('total', 1)] 2025-12-04T09:28:45.8112470Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8112601Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8112677Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8112737Z graph_break [] 2025-12-04T09:28:45.8112819Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8112942Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8113006Z frames [('total', 1)] 2025-12-04T09:28:45.8113079Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8113205Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8113283Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8113343Z graph_break [] 2025-12-04T09:28:45.8113422Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8113553Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8113614Z frames [('total', 1)] 2025-12-04T09:28:45.8113682Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8113815Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8113891Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8113950Z graph_break [] 2025-12-04T09:28:45.8114032Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8114165Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8114234Z frames [('total', 1)] 2025-12-04T09:28:45.8114303Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8114428Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8114505Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8114565Z graph_break [] 2025-12-04T09:28:45.8114688Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8114819Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8114880Z frames [('total', 1)] 2025-12-04T09:28:45.8114985Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8115179Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8115253Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8115319Z graph_break [] 2025-12-04T09:28:45.8115396Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8115520Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8115590Z frames [('total', 1)] 2025-12-04T09:28:45.8115661Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8115795Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8115876Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8115935Z graph_break [] 2025-12-04T09:28:45.8116017Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8116152Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8116213Z frames [('total', 1)] 2025-12-04T09:28:45.8116291Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8116417Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8116495Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8116564Z graph_break [] 2025-12-04T09:28:45.8116641Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8116772Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8116842Z frames [('total', 1)] 2025-12-04T09:28:45.8116910Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8117041Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8117122Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8117181Z graph_break [] 2025-12-04T09:28:45.8117268Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8117400Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8117463Z frames [('total', 1)] 2025-12-04T09:28:45.8117541Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8117673Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8117749Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8117817Z graph_break [] 2025-12-04T09:28:45.8117906Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8118034Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8118106Z frames [('total', 1)] 2025-12-04T09:28:45.8118175Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8118307Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8118381Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8118441Z graph_break [] 2025-12-04T09:28:45.8118522Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8118651Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8118714Z frames [('total', 1)] 2025-12-04T09:28:45.8118790Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8118918Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8118999Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8119076Z graph_break [] 2025-12-04T09:28:45.8119156Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8119292Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8119354Z frames [('total', 1)] 2025-12-04T09:28:45.8119421Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8119555Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8119627Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8119687Z graph_break [] 2025-12-04T09:28:45.8119768Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8119939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8120002Z frames [('total', 1)] 2025-12-04T09:28:45.8120079Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8120239Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8120400Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8120461Z graph_break [] 2025-12-04T09:28:45.8120537Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8120668Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8120731Z frames [('total', 1)] 2025-12-04T09:28:45.8120798Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8120934Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8121007Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8121066Z graph_break [] 2025-12-04T09:28:45.8121148Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8121276Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8121338Z frames [('total', 1)] 2025-12-04T09:28:45.8121424Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8121553Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8121633Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8121697Z graph_break [] 2025-12-04T09:28:45.8121774Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8121904Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8121966Z frames [('total', 1)] 2025-12-04T09:28:45.8122032Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8122162Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8122233Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8122292Z graph_break [] 2025-12-04T09:28:45.8122373Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8122496Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8122569Z frames [('total', 1)] 2025-12-04T09:28:45.8122636Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8122760Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8122844Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8122908Z graph_break [] 2025-12-04T09:28:45.8122985Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8123116Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8123177Z frames [('total', 1)] 2025-12-04T09:28:45.8123244Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8123374Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8123446Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8123510Z graph_break [] 2025-12-04T09:28:45.8123587Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8123709Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8123780Z frames [('total', 1)] 2025-12-04T09:28:45.8123847Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8123969Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8124047Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8124110Z graph_break [] 2025-12-04T09:28:45.8124188Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8124320Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8124383Z frames [('total', 1)] 2025-12-04T09:28:45.8124457Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8124582Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8124653Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8124717Z graph_break [] 2025-12-04T09:28:45.8124792Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8124956Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8125085Z Traceback (most recent call last): 2025-12-04T09:28:45.8125402Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8125499Z f(a, b) 2025-12-04T09:28:45.8125863Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8126016Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8126321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8126438Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8126731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8126842Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8127166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8127367Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8127677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8127751Z graph.run(*example_inputs) 2025-12-04T09:28:45.8127993Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8128062Z return super().run(*args) 2025-12-04T09:28:45.8128294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8128380Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8128631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8128711Z result = super().run_node(n) 2025-12-04T09:28:45.8128958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8129064Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8129344Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8129494Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8129761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8129893Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8130161Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8130240Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8130508Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8130590Z return autotune_select_algorithm( 2025-12-04T09:28:45.8130935Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8131013Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8131307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8131470Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8132213Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8132291Z target: aten.mm.default 2025-12-04T09:28:45.8132399Z args[0]: TensorBox( 2025-12-04T09:28:45.8132472Z ReinterpretView( 2025-12-04T09:28:45.8132533Z StorageBox( 2025-12-04T09:28:45.8132777Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8132875Z ), 2025-12-04T09:28:45.8133086Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8133266Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8133340Z stack_traces = {, 2025-12-04T09:28:45.8133579Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8133658Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8133722Z , 2025-12-04T09:28:45.8133778Z } 2025-12-04T09:28:45.8133835Z ) 2025-12-04T09:28:45.8133898Z ) 2025-12-04T09:28:45.8133976Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8134226Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8134283Z )) 2025-12-04T09:28:45.8134286Z 2025-12-04T09:28:45.8134710Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8134721Z 2025-12-04T09:28:45.8134724Z 2025-12-04T09:28:45.8134867Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8135113Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8135116Z 2025-12-04T09:28:45.8135284Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8135418Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8135481Z frames [('total', 1)] 2025-12-04T09:28:45.8135562Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8135636Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8135775Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8135844Z graph_break [] 2025-12-04T09:28:45.8135926Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8136064Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8136131Z frames [('total', 1)] 2025-12-04T09:28:45.8136201Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8136336Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8136409Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8136468Z graph_break [] 2025-12-04T09:28:45.8136552Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8136678Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8136741Z frames [('total', 1)] 2025-12-04T09:28:45.8136815Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8136941Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8137035Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8137101Z graph_break [] 2025-12-04T09:28:45.8137183Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8137317Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8137382Z frames [('total', 1)] 2025-12-04T09:28:45.8137452Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8137586Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8137660Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8137719Z graph_break [] 2025-12-04T09:28:45.8137805Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8137930Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8137997Z frames [('total', 1)] 2025-12-04T09:28:45.8138066Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8138194Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8138326Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8138385Z graph_break [] 2025-12-04T09:28:45.8138462Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8138632Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8138695Z frames [('total', 1)] 2025-12-04T09:28:45.8138827Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8138961Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8139035Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8139101Z graph_break [] 2025-12-04T09:28:45.8139179Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8139306Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8139375Z frames [('total', 1)] 2025-12-04T09:28:45.8139443Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8139572Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8139652Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8139713Z graph_break [] 2025-12-04T09:28:45.8139792Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8139924Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8139987Z frames [('total', 1)] 2025-12-04T09:28:45.8140065Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8140193Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8140267Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8140332Z graph_break [] 2025-12-04T09:28:45.8140409Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8140536Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8140603Z frames [('total', 1)] 2025-12-04T09:28:45.8140672Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8140809Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8140894Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8140953Z graph_break [] 2025-12-04T09:28:45.8141036Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8141162Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8141227Z frames [('total', 1)] 2025-12-04T09:28:45.8141306Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8141433Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8141505Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8141573Z graph_break [] 2025-12-04T09:28:45.8141651Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8141777Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8141848Z frames [('total', 1)] 2025-12-04T09:28:45.8141916Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8142047Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8142121Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8142182Z graph_break [] 2025-12-04T09:28:45.8142264Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8142391Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8142459Z frames [('total', 1)] 2025-12-04T09:28:45.8142540Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8142667Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8142741Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8142813Z graph_break [] 2025-12-04T09:28:45.8142890Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8143012Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8143081Z frames [('total', 1)] 2025-12-04T09:28:45.8143149Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8143278Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8143397Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8143455Z graph_break [] 2025-12-04T09:28:45.8143538Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8143660Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8143775Z frames [('total', 1)] 2025-12-04T09:28:45.8143849Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8144037Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8144111Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8144179Z graph_break [] 2025-12-04T09:28:45.8144256Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8144387Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8144449Z frames [('total', 1)] 2025-12-04T09:28:45.8144515Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8144645Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8144717Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8144781Z graph_break [] 2025-12-04T09:28:45.8144862Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8144985Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8145049Z frames [('total', 1)] 2025-12-04T09:28:45.8145120Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8145261Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8145339Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8145398Z graph_break [] 2025-12-04T09:28:45.8145475Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8145603Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8145667Z frames [('total', 1)] 2025-12-04T09:28:45.8145733Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8145862Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8145934Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8145996Z graph_break [] 2025-12-04T09:28:45.8146076Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8146198Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8146268Z frames [('total', 1)] 2025-12-04T09:28:45.8146335Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8146469Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8146546Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8146605Z graph_break [] 2025-12-04T09:28:45.8146682Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8146812Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8146879Z frames [('total', 1)] 2025-12-04T09:28:45.8146947Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8147075Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8147147Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8147214Z graph_break [] 2025-12-04T09:28:45.8147291Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8147414Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8147489Z frames [('total', 1)] 2025-12-04T09:28:45.8147557Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8147684Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8147777Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8147839Z graph_break [] 2025-12-04T09:28:45.8147917Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8148047Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8148109Z frames [('total', 1)] 2025-12-04T09:28:45.8148183Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8148309Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8148381Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8148499Z graph_break [] 2025-12-04T09:28:45.8148576Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8148702Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8148808Z frames [('total', 1)] 2025-12-04T09:28:45.8148877Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8149067Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8149150Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8149208Z graph_break [] 2025-12-04T09:28:45.8149291Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8149415Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8149477Z frames [('total', 1)] 2025-12-04T09:28:45.8149557Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8149685Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8149760Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8149829Z graph_break [] 2025-12-04T09:28:45.8149908Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8150033Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8150105Z frames [('total', 1)] 2025-12-04T09:28:45.8150175Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8150311Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8150384Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8150444Z graph_break [] 2025-12-04T09:28:45.8150529Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8150663Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8150728Z frames [('total', 1)] 2025-12-04T09:28:45.8150804Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8150930Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8151002Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8151069Z graph_break [] 2025-12-04T09:28:45.8151149Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8151317Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8151402Z Traceback (most recent call last): 2025-12-04T09:28:45.8151727Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8151789Z f(a, b) 2025-12-04T09:28:45.8152078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8152228Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8152545Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8152665Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8152970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8153065Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8153380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8153587Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8153891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8153965Z graph.run(*example_inputs) 2025-12-04T09:28:45.8154208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8154277Z return super().run(*args) 2025-12-04T09:28:45.8154517Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8154597Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8154896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8154973Z result = super().run_node(n) 2025-12-04T09:28:45.8155218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8155425Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8155700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8155847Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8156114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8156245Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8156502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8156582Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8156842Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8156930Z return autotune_select_algorithm( 2025-12-04T09:28:45.8157283Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8157357Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8157652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8157810Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8158562Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8158634Z target: aten.mm.default 2025-12-04T09:28:45.8158697Z args[0]: TensorBox( 2025-12-04T09:28:45.8158767Z ReinterpretView( 2025-12-04T09:28:45.8158830Z StorageBox( 2025-12-04T09:28:45.8159071Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8159134Z ), 2025-12-04T09:28:45.8159283Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8159356Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8159427Z stack_traces = {, 2025-12-04T09:28:45.8159653Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8159734Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8159790Z , 2025-12-04T09:28:45.8159846Z } 2025-12-04T09:28:45.8159905Z ) 2025-12-04T09:28:45.8159959Z ) 2025-12-04T09:28:45.8160035Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8160277Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8160334Z )) 2025-12-04T09:28:45.8160337Z 2025-12-04T09:28:45.8160771Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8160775Z 2025-12-04T09:28:45.8160778Z 2025-12-04T09:28:45.8160920Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8161164Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8161168Z 2025-12-04T09:28:45.8161340Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8161474Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8161595Z frames [('total', 1)] 2025-12-04T09:28:45.8161672Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8161749Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8161892Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8161990Z graph_break [] 2025-12-04T09:28:45.8162073Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8162288Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8162354Z frames [('total', 1)] 2025-12-04T09:28:45.8162424Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8162559Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8162636Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8162703Z graph_break [] 2025-12-04T09:28:45.8162784Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8162911Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8162980Z frames [('total', 1)] 2025-12-04T09:28:45.8163052Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8163179Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8163260Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8163324Z graph_break [] 2025-12-04T09:28:45.8163401Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8163537Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8163599Z frames [('total', 1)] 2025-12-04T09:28:45.8163675Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8163801Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8163875Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8163943Z graph_break [] 2025-12-04T09:28:45.8164020Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8164147Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8164217Z frames [('total', 1)] 2025-12-04T09:28:45.8164288Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8164416Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8164496Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8164557Z graph_break [] 2025-12-04T09:28:45.8164634Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8164782Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8164846Z frames [('total', 1)] 2025-12-04T09:28:45.8164923Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8165051Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8165124Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8165192Z graph_break [] 2025-12-04T09:28:45.8165268Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8165393Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8165461Z frames [('total', 1)] 2025-12-04T09:28:45.8165531Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8165656Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8165735Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8165797Z graph_break [] 2025-12-04T09:28:45.8165878Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8166004Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8166069Z frames [('total', 1)] 2025-12-04T09:28:45.8166145Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8166268Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8166340Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8166411Z graph_break [] 2025-12-04T09:28:45.8166486Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8166610Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8166682Z frames [('total', 1)] 2025-12-04T09:28:45.8166750Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8166949Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8167022Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8167083Z graph_break [] 2025-12-04T09:28:45.8167212Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8167402Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8167466Z frames [('total', 1)] 2025-12-04T09:28:45.8167538Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8167665Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8167737Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8167803Z graph_break [] 2025-12-04T09:28:45.8167880Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8168008Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8168069Z frames [('total', 1)] 2025-12-04T09:28:45.8168137Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8168271Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8168343Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8168402Z graph_break [] 2025-12-04T09:28:45.8168489Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8168619Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8168685Z frames [('total', 1)] 2025-12-04T09:28:45.8168762Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8168885Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8168963Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8169023Z graph_break [] 2025-12-04T09:28:45.8169099Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8169240Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8169306Z frames [('total', 1)] 2025-12-04T09:28:45.8169375Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8169510Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8169582Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8169642Z graph_break [] 2025-12-04T09:28:45.8169729Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8169857Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8169925Z frames [('total', 1)] 2025-12-04T09:28:45.8169996Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8170120Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8170198Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8170259Z graph_break [] 2025-12-04T09:28:45.8170337Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8170470Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8170533Z frames [('total', 1)] 2025-12-04T09:28:45.8170601Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8170736Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8170808Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8170874Z graph_break [] 2025-12-04T09:28:45.8170950Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8171087Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8171161Z frames [('total', 1)] 2025-12-04T09:28:45.8171231Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8171357Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8171436Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8171498Z graph_break [] 2025-12-04T09:28:45.8171575Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8171710Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8171773Z frames [('total', 1)] 2025-12-04T09:28:45.8171849Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8171976Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8172096Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8172169Z graph_break [] 2025-12-04T09:28:45.8172247Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8172409Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8172542Z frames [('total', 1)] 2025-12-04T09:28:45.8172615Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8172741Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8172821Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8172882Z graph_break [] 2025-12-04T09:28:45.8172960Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8173089Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8173243Z frames [('total', 1)] 2025-12-04T09:28:45.8173320Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8173447Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8173523Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8173589Z graph_break [] 2025-12-04T09:28:45.8173668Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8173799Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8173871Z frames [('total', 1)] 2025-12-04T09:28:45.8173941Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8174064Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8174144Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8174204Z graph_break [] 2025-12-04T09:28:45.8174286Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8174409Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8174472Z frames [('total', 1)] 2025-12-04T09:28:45.8174552Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8174680Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8174753Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8174830Z graph_break [] 2025-12-04T09:28:45.8174909Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8175035Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8175102Z frames [('total', 1)] 2025-12-04T09:28:45.8175172Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8175300Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8175371Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8175429Z graph_break [] 2025-12-04T09:28:45.8175510Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8175634Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8175696Z frames [('total', 1)] 2025-12-04T09:28:45.8175771Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8175897Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8175970Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8176037Z graph_break [] 2025-12-04T09:28:45.8176114Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8176244Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8176308Z frames [('total', 1)] 2025-12-04T09:28:45.8176379Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8176509Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8176581Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8176642Z graph_break [] 2025-12-04T09:28:45.8176724Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8176848Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8176921Z frames [('total', 1)] 2025-12-04T09:28:45.8176996Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8177121Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8177246Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8177305Z graph_break [] 2025-12-04T09:28:45.8177382Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8177511Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8177610Z frames [('total', 1)] 2025-12-04T09:28:45.8177742Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8177874Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8177946Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8178007Z graph_break [] 2025-12-04T09:28:45.8178092Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8178254Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8178337Z Traceback (most recent call last): 2025-12-04T09:28:45.8178648Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8178709Z f(a, b) 2025-12-04T09:28:45.8179004Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8179155Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8179462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8179591Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8179886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8179985Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8180295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8180492Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8180803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8180877Z graph.run(*example_inputs) 2025-12-04T09:28:45.8181118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8181193Z return super().run(*args) 2025-12-04T09:28:45.8181425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8181511Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8181762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8181835Z result = super().run_node(n) 2025-12-04T09:28:45.8182089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8182204Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8182484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8182637Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8182907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8183046Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8183304Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8183376Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8183647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8183727Z return autotune_select_algorithm( 2025-12-04T09:28:45.8184074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8184190Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8184478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8184763Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8185506Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8185588Z target: aten.mm.default 2025-12-04T09:28:45.8185651Z args[0]: TensorBox( 2025-12-04T09:28:45.8185721Z ReinterpretView( 2025-12-04T09:28:45.8185791Z StorageBox( 2025-12-04T09:28:45.8186034Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8186093Z ), 2025-12-04T09:28:45.8186251Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8186324Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8186400Z stack_traces = {, 2025-12-04T09:28:45.8186633Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8186710Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8186770Z , 2025-12-04T09:28:45.8186824Z } 2025-12-04T09:28:45.8186878Z ) 2025-12-04T09:28:45.8186937Z ) 2025-12-04T09:28:45.8187009Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8187241Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8187304Z )) 2025-12-04T09:28:45.8187307Z 2025-12-04T09:28:45.8187728Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8187733Z 2025-12-04T09:28:45.8187737Z 2025-12-04T09:28:45.8187877Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8188122Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8188130Z 2025-12-04T09:28:45.8188292Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8188426Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8188489Z frames [('total', 1)] 2025-12-04T09:28:45.8188565Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8188638Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8188773Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8188837Z graph_break [] 2025-12-04T09:28:45.8188917Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8189045Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8189125Z frames [('total', 1)] 2025-12-04T09:28:45.8189194Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8189323Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8189404Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8189466Z graph_break [] 2025-12-04T09:28:45.8189550Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8189676Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8189739Z frames [('total', 1)] 2025-12-04T09:28:45.8189814Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8189941Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8190013Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8190079Z graph_break [] 2025-12-04T09:28:45.8190158Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8190283Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8190402Z frames [('total', 1)] 2025-12-04T09:28:45.8190470Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8190600Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8190708Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8190829Z graph_break [] 2025-12-04T09:28:45.8190914Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8191040Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8191104Z frames [('total', 1)] 2025-12-04T09:28:45.8191179Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8191305Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8191377Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8191446Z graph_break [] 2025-12-04T09:28:45.8191522Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8191653Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8191717Z frames [('total', 1)] 2025-12-04T09:28:45.8191786Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8191918Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8191992Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8192060Z graph_break [] 2025-12-04T09:28:45.8192150Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8192276Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8192338Z frames [('total', 1)] 2025-12-04T09:28:45.8192416Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8192543Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8192621Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8192681Z graph_break [] 2025-12-04T09:28:45.8192758Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8192888Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8192954Z frames [('total', 1)] 2025-12-04T09:28:45.8193023Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8193155Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8193231Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8193293Z graph_break [] 2025-12-04T09:28:45.8193381Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8193506Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8193579Z frames [('total', 1)] 2025-12-04T09:28:45.8193649Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8193774Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8193854Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8193915Z graph_break [] 2025-12-04T09:28:45.8193992Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8194126Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8194192Z frames [('total', 1)] 2025-12-04T09:28:45.8194260Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8194395Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8194470Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8194536Z graph_break [] 2025-12-04T09:28:45.8194616Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8194741Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8194813Z frames [('total', 1)] 2025-12-04T09:28:45.8194882Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8195010Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8195098Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8195162Z graph_break [] 2025-12-04T09:28:45.8195240Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8195371Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8195483Z frames [('total', 1)] 2025-12-04T09:28:45.8195561Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8195687Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8195793Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8195855Z graph_break [] 2025-12-04T09:28:45.8196050Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8196180Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8196247Z frames [('total', 1)] 2025-12-04T09:28:45.8196315Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8196441Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8196523Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8196582Z graph_break [] 2025-12-04T09:28:45.8196659Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8196789Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8196855Z frames [('total', 1)] 2025-12-04T09:28:45.8196929Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8197055Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8197129Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8197192Z graph_break [] 2025-12-04T09:28:45.8197275Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8197402Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8197471Z frames [('total', 1)] 2025-12-04T09:28:45.8197539Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8197677Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8197756Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8197815Z graph_break [] 2025-12-04T09:28:45.8197897Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8198023Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8198088Z frames [('total', 1)] 2025-12-04T09:28:45.8198161Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8198286Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8198356Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8198424Z graph_break [] 2025-12-04T09:28:45.8198502Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8198630Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8198700Z frames [('total', 1)] 2025-12-04T09:28:45.8198768Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8198899Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8198972Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8199030Z graph_break [] 2025-12-04T09:28:45.8199116Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8199240Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8199302Z frames [('total', 1)] 2025-12-04T09:28:45.8199379Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8199505Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8199578Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8199646Z graph_break [] 2025-12-04T09:28:45.8199722Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8199858Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8199921Z frames [('total', 1)] 2025-12-04T09:28:45.8199990Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8200131Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8200206Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8200267Z graph_break [] 2025-12-04T09:28:45.8200396Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8200553Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8200780Z frames [('total', 1)] 2025-12-04T09:28:45.8200998Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8201172Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8201278Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8201438Z graph_break [] 2025-12-04T09:28:45.8201534Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8201894Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8201995Z frames [('total', 1)] 2025-12-04T09:28:45.8202096Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8202293Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8202412Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8202485Z graph_break [] 2025-12-04T09:28:45.8202696Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8202853Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8202979Z frames [('total', 1)] 2025-12-04T09:28:45.8203095Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8203251Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8203419Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8203525Z graph_break [] 2025-12-04T09:28:45.8203684Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8203890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8204002Z frames [('total', 1)] 2025-12-04T09:28:45.8204100Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8204323Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8204444Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8204891Z graph_break [] 2025-12-04T09:28:45.8205040Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8205206Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8205319Z frames [('total', 1)] 2025-12-04T09:28:45.8205509Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8205702Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8205843Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8205954Z graph_break [] 2025-12-04T09:28:45.8206069Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8206255Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8206394Z frames [('total', 1)] 2025-12-04T09:28:45.8206540Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8206718Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8206820Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8206944Z graph_break [] 2025-12-04T09:28:45.8207036Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8207243Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8207437Z frames [('total', 1)] 2025-12-04T09:28:45.8207549Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8207738Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8207838Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8207911Z graph_break [] 2025-12-04T09:28:45.8208112Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8208286Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8208377Z frames [('total', 1)] 2025-12-04T09:28:45.8208506Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8208658Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8208821Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8208934Z graph_break [] 2025-12-04T09:28:45.8209059Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8209286Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8209390Z Traceback (most recent call last): 2025-12-04T09:28:45.8209746Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8210013Z f(a, b) 2025-12-04T09:28:45.8210379Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8210753Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8211096Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8211244Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8211604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8211825Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8212181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8212447Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8212778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8212932Z graph.run(*example_inputs) 2025-12-04T09:28:45.8213298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8213454Z return super().run(*args) 2025-12-04T09:28:45.8213759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8213868Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8214188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8214307Z result = super().run_node(n) 2025-12-04T09:28:45.8214572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8214803Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8215103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8215292Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8215659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8215823Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8216182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8216298Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8216589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8216732Z return autotune_select_algorithm( 2025-12-04T09:28:45.8217121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8217286Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8217648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8217861Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8218677Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8218794Z target: aten.mm.default 2025-12-04T09:28:45.8218920Z args[0]: TensorBox( 2025-12-04T09:28:45.8218999Z ReinterpretView( 2025-12-04T09:28:45.8219135Z StorageBox( 2025-12-04T09:28:45.8219501Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8219603Z ), 2025-12-04T09:28:45.8219780Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8219954Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8220098Z stack_traces = {, 2025-12-04T09:28:45.8220492Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8220608Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8220693Z , 2025-12-04T09:28:45.8220808Z } 2025-12-04T09:28:45.8220891Z ) 2025-12-04T09:28:45.8220957Z ) 2025-12-04T09:28:45.8221168Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8221431Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8221513Z )) 2025-12-04T09:28:45.8221601Z 2025-12-04T09:28:45.8222054Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8222058Z 2025-12-04T09:28:45.8222064Z 2025-12-04T09:28:45.8222230Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8222587Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8222591Z 2025-12-04T09:28:45.8222819Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8223014Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8223108Z frames [('total', 1)] 2025-12-04T09:28:45.8223210Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8223333Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8223571Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8223677Z graph_break [] 2025-12-04T09:28:45.8223826Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8223984Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8224108Z frames [('total', 1)] 2025-12-04T09:28:45.8224194Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8224414Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8224565Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8224654Z graph_break [] 2025-12-04T09:28:45.8224762Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8224951Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8225045Z frames [('total', 1)] 2025-12-04T09:28:45.8225184Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8225389Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8225492Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8225657Z graph_break [] 2025-12-04T09:28:45.8225784Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8225925Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8226103Z frames [('total', 1)] 2025-12-04T09:28:45.8226202Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8226358Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8226492Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8226597Z graph_break [] 2025-12-04T09:28:45.8226762Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8226929Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8227023Z frames [('total', 1)] 2025-12-04T09:28:45.8227151Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8227322Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8227421Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8227573Z graph_break [] 2025-12-04T09:28:45.8227738Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8227892Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8228033Z frames [('total', 1)] 2025-12-04T09:28:45.8228167Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8228341Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8228559Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8228724Z graph_break [] 2025-12-04T09:28:45.8228881Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8229035Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8229123Z frames [('total', 1)] 2025-12-04T09:28:45.8229238Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8229444Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8229608Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8229698Z graph_break [] 2025-12-04T09:28:45.8229803Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8229992Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8230069Z frames [('total', 1)] 2025-12-04T09:28:45.8230213Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8230437Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8230543Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8230631Z graph_break [] 2025-12-04T09:28:45.8230769Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8230907Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8231104Z frames [('total', 1)] 2025-12-04T09:28:45.8231201Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8231355Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8231486Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8231576Z graph_break [] 2025-12-04T09:28:45.8231666Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8231978Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8232070Z frames [('total', 1)] 2025-12-04T09:28:45.8232198Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8232352Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8232457Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8232645Z graph_break [] 2025-12-04T09:28:45.8232765Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8232919Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8233042Z frames [('total', 1)] 2025-12-04T09:28:45.8233147Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8233304Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8233484Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8233587Z graph_break [] 2025-12-04T09:28:45.8233723Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8233878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8233970Z frames [('total', 1)] 2025-12-04T09:28:45.8234098Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8234294Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8234413Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8234534Z graph_break [] 2025-12-04T09:28:45.8234640Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8234792Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8234917Z frames [('total', 1)] 2025-12-04T09:28:45.8235061Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8235313Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8235413Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8235499Z graph_break [] 2025-12-04T09:28:45.8235650Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8235833Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8235969Z frames [('total', 1)] 2025-12-04T09:28:45.8236126Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8236320Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8236468Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8236621Z graph_break [] 2025-12-04T09:28:45.8236713Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8236959Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8237050Z frames [('total', 1)] 2025-12-04T09:28:45.8237168Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8237354Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8237454Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8237602Z graph_break [] 2025-12-04T09:28:45.8237722Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8237892Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8238018Z frames [('total', 1)] 2025-12-04T09:28:45.8238123Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8238327Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8238498Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8238605Z graph_break [] 2025-12-04T09:28:45.8238729Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8238914Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8239004Z frames [('total', 1)] 2025-12-04T09:28:45.8239118Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8239319Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8239453Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8239574Z graph_break [] 2025-12-04T09:28:45.8239678Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8239829Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8239950Z frames [('total', 1)] 2025-12-04T09:28:45.8240098Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8240280Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8240414Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8240504Z graph_break [] 2025-12-04T09:28:45.8240640Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8240783Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8240920Z frames [('total', 1)] 2025-12-04T09:28:45.8241080Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8241233Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8241331Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8241502Z graph_break [] 2025-12-04T09:28:45.8241593Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8241858Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8241951Z frames [('total', 1)] 2025-12-04T09:28:45.8242051Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8242235Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8242339Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8242413Z graph_break [] 2025-12-04T09:28:45.8242661Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8242818Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8242908Z frames [('total', 1)] 2025-12-04T09:28:45.8243038Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8243196Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8243393Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8243499Z graph_break [] 2025-12-04T09:28:45.8243605Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8243793Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8243928Z frames [('total', 1)] 2025-12-04T09:28:45.8244044Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8244294Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8244458Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8244578Z graph_break [] 2025-12-04T09:28:45.8244814Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8244972Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8245081Z frames [('total', 1)] 2025-12-04T09:28:45.8245227Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8245397Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8245547Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8245644Z graph_break [] 2025-12-04T09:28:45.8245753Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8245927Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8246067Z frames [('total', 1)] 2025-12-04T09:28:45.8246212Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8246380Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8246482Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8246602Z graph_break [] 2025-12-04T09:28:45.8246698Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8246897Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8247054Z frames [('total', 1)] 2025-12-04T09:28:45.8247151Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8247345Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8247445Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8247518Z graph_break [] 2025-12-04T09:28:45.8247716Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8247937Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8248030Z frames [('total', 1)] 2025-12-04T09:28:45.8248162Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8248315Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8248404Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8248605Z graph_break [] 2025-12-04T09:28:45.8248718Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8248907Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8248998Z frames [('total', 1)] 2025-12-04T09:28:45.8249094Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8249340Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8253962Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8254062Z graph_break [] 2025-12-04T09:28:45.8254162Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8254322Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8254400Z frames [('total', 1)] 2025-12-04T09:28:45.8254476Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8254641Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8254724Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8254791Z graph_break [] 2025-12-04T09:28:45.8254887Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8255068Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8255158Z Traceback (most recent call last): 2025-12-04T09:28:45.8255486Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8255547Z f(a, b) 2025-12-04T09:28:45.8255856Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8256011Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8256311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8256534Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8256890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8257066Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8257389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8257586Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8257894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8257973Z graph.run(*example_inputs) 2025-12-04T09:28:45.8258228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8258305Z return super().run(*args) 2025-12-04T09:28:45.8258562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8258654Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8258928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8259002Z result = super().run_node(n) 2025-12-04T09:28:45.8259260Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8259368Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8259648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8259801Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8260067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8260211Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8260471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8260552Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8260821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8260904Z return autotune_select_algorithm( 2025-12-04T09:28:45.8261255Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8261328Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8261621Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8261788Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8262544Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8262626Z target: aten.mm.default 2025-12-04T09:28:45.8262691Z args[0]: TensorBox( 2025-12-04T09:28:45.8262757Z ReinterpretView( 2025-12-04T09:28:45.8262838Z StorageBox( 2025-12-04T09:28:45.8263087Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8263146Z ), 2025-12-04T09:28:45.8263309Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8263384Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8263449Z stack_traces = {, 2025-12-04T09:28:45.8263730Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8263810Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8263875Z , 2025-12-04T09:28:45.8263967Z } 2025-12-04T09:28:45.8264024Z ) 2025-12-04T09:28:45.8264086Z ) 2025-12-04T09:28:45.8264161Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8264465Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8264529Z )) 2025-12-04T09:28:45.8264533Z 2025-12-04T09:28:45.8264959Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8264963Z 2025-12-04T09:28:45.8264966Z 2025-12-04T09:28:45.8265110Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8265359Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8265365Z 2025-12-04T09:28:45.8265534Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8265683Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8265752Z frames [('total', 1)] 2025-12-04T09:28:45.8265841Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8265920Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8266057Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8266137Z graph_break [] 2025-12-04T09:28:45.8266222Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8266355Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8266428Z frames [('total', 1)] 2025-12-04T09:28:45.8266498Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8266627Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8266712Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8266773Z graph_break [] 2025-12-04T09:28:45.8266856Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8266991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8267061Z frames [('total', 1)] 2025-12-04T09:28:45.8267142Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8267270Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8267343Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8267410Z graph_break [] 2025-12-04T09:28:45.8267489Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8267617Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8267690Z frames [('total', 1)] 2025-12-04T09:28:45.8267758Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8267892Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8267969Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8268030Z graph_break [] 2025-12-04T09:28:45.8268118Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8268245Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8268312Z frames [('total', 1)] 2025-12-04T09:28:45.8268390Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8268523Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8268599Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8268669Z graph_break [] 2025-12-04T09:28:45.8268748Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8268878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8268943Z frames [('total', 1)] 2025-12-04T09:28:45.8269015Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8269154Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8269226Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8269345Z graph_break [] 2025-12-04T09:28:45.8269433Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8269557Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8269658Z frames [('total', 1)] 2025-12-04T09:28:45.8269737Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8269925Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8270007Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8270068Z graph_break [] 2025-12-04T09:28:45.8270145Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8270277Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8270339Z frames [('total', 1)] 2025-12-04T09:28:45.8270409Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8270538Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8270614Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8270675Z graph_break [] 2025-12-04T09:28:45.8270761Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8270886Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8270956Z frames [('total', 1)] 2025-12-04T09:28:45.8271026Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8271157Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8271236Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8271296Z graph_break [] 2025-12-04T09:28:45.8271373Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8271503Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8271578Z frames [('total', 1)] 2025-12-04T09:28:45.8271649Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8271779Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8271851Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8271916Z graph_break [] 2025-12-04T09:28:45.8272001Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8272124Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8272191Z frames [('total', 1)] 2025-12-04T09:28:45.8272266Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8272397Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8272475Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8272535Z graph_break [] 2025-12-04T09:28:45.8272614Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8272747Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8272809Z frames [('total', 1)] 2025-12-04T09:28:45.8272878Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8273013Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8273084Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8273146Z graph_break [] 2025-12-04T09:28:45.8273233Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8273355Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8273424Z frames [('total', 1)] 2025-12-04T09:28:45.8273491Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8273621Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8273698Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8273762Z graph_break [] 2025-12-04T09:28:45.8273848Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8273979Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8274041Z frames [('total', 1)] 2025-12-04T09:28:45.8274108Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8274240Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8274315Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8274381Z graph_break [] 2025-12-04T09:28:45.8274507Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8274633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8274701Z frames [('total', 1)] 2025-12-04T09:28:45.8274805Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8274995Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8275080Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8275139Z graph_break [] 2025-12-04T09:28:45.8275220Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8275351Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8275414Z frames [('total', 1)] 2025-12-04T09:28:45.8275489Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8275613Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8275687Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8275750Z graph_break [] 2025-12-04T09:28:45.8275829Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8275954Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8276023Z frames [('total', 1)] 2025-12-04T09:28:45.8276094Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8276217Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8276302Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8276363Z graph_break [] 2025-12-04T09:28:45.8276440Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8276571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8276633Z frames [('total', 1)] 2025-12-04T09:28:45.8276713Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8276837Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8276909Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8276974Z graph_break [] 2025-12-04T09:28:45.8277051Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8277178Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8277246Z frames [('total', 1)] 2025-12-04T09:28:45.8277319Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8277447Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8277532Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8277600Z graph_break [] 2025-12-04T09:28:45.8277687Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8277812Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8277873Z frames [('total', 1)] 2025-12-04T09:28:45.8277947Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8278072Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8278143Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8278207Z graph_break [] 2025-12-04T09:28:45.8278283Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8278407Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8278476Z frames [('total', 1)] 2025-12-04T09:28:45.8278543Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8278675Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8278750Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8278809Z graph_break [] 2025-12-04T09:28:45.8278890Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8279014Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8279074Z frames [('total', 1)] 2025-12-04T09:28:45.8279147Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8279272Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8279345Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8279412Z graph_break [] 2025-12-04T09:28:45.8279488Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8279682Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8279744Z frames [('total', 1)] 2025-12-04T09:28:45.8279815Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8279980Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8280114Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8280174Z graph_break [] 2025-12-04T09:28:45.8280257Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8280382Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8280444Z frames [('total', 1)] 2025-12-04T09:28:45.8280518Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8280641Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8280720Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8280779Z graph_break [] 2025-12-04T09:28:45.8280855Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8280988Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8281052Z frames [('total', 1)] 2025-12-04T09:28:45.8281120Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8281267Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8281341Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8281404Z graph_break [] 2025-12-04T09:28:45.8281487Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8281611Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8281679Z frames [('total', 1)] 2025-12-04T09:28:45.8281749Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8281873Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8281950Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8282010Z graph_break [] 2025-12-04T09:28:45.8282087Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8282220Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8282282Z frames [('total', 1)] 2025-12-04T09:28:45.8282353Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8282485Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8282558Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8282625Z graph_break [] 2025-12-04T09:28:45.8282701Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8282829Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8282898Z frames [('total', 1)] 2025-12-04T09:28:45.8282966Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8283092Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8283171Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8283232Z graph_break [] 2025-12-04T09:28:45.8283309Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8283441Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8283506Z frames [('total', 1)] 2025-12-04T09:28:45.8283581Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8283705Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8283779Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8283852Z graph_break [] 2025-12-04T09:28:45.8283930Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8284096Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8284182Z Traceback (most recent call last): 2025-12-04T09:28:45.8284501Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8284560Z f(a, b) 2025-12-04T09:28:45.8284858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8285010Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8285370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8285523Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8285881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8285981Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8286294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8286498Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8286798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8286875Z graph.run(*example_inputs) 2025-12-04T09:28:45.8287116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8287184Z return super().run(*args) 2025-12-04T09:28:45.8287416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8287511Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8287774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8287849Z result = super().run_node(n) 2025-12-04T09:28:45.8288095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8288208Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8288482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8288629Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8288894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8289032Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8289293Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8289368Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8289629Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8289709Z return autotune_select_algorithm( 2025-12-04T09:28:45.8290057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8290127Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8290416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8290579Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8291328Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8291406Z target: aten.mm.default 2025-12-04T09:28:45.8291470Z args[0]: TensorBox( 2025-12-04T09:28:45.8291536Z ReinterpretView( 2025-12-04T09:28:45.8291600Z StorageBox( 2025-12-04T09:28:45.8291841Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8291902Z ), 2025-12-04T09:28:45.8292052Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8292125Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8292240Z stack_traces = {, 2025-12-04T09:28:45.8292468Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8292545Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8292641Z , 2025-12-04T09:28:45.8292695Z } 2025-12-04T09:28:45.8292750Z ) 2025-12-04T09:28:45.8292872Z ) 2025-12-04T09:28:45.8292945Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8293309Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8293365Z )) 2025-12-04T09:28:45.8293369Z 2025-12-04T09:28:45.8293788Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8293791Z 2025-12-04T09:28:45.8293795Z 2025-12-04T09:28:45.8293936Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8294185Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8294189Z 2025-12-04T09:28:45.8294356Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8294494Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8294557Z frames [('total', 1)] 2025-12-04T09:28:45.8294637Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8294711Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8294844Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8294913Z graph_break [] 2025-12-04T09:28:45.8294994Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8295127Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8295189Z frames [('total', 1)] 2025-12-04T09:28:45.8295259Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8295401Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8295475Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8295534Z graph_break [] 2025-12-04T09:28:45.8295620Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8295746Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8295815Z frames [('total', 1)] 2025-12-04T09:28:45.8295891Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8296017Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8296094Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8296153Z graph_break [] 2025-12-04T09:28:45.8296232Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8296363Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8296426Z frames [('total', 1)] 2025-12-04T09:28:45.8296495Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8296627Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8296702Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8296761Z graph_break [] 2025-12-04T09:28:45.8296846Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8296973Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8297046Z frames [('total', 1)] 2025-12-04T09:28:45.8297122Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8297248Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8297328Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8297391Z graph_break [] 2025-12-04T09:28:45.8297468Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8297599Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8297665Z frames [('total', 1)] 2025-12-04T09:28:45.8297735Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8297867Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8297989Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8298058Z graph_break [] 2025-12-04T09:28:45.8298146Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8298308Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8298455Z frames [('total', 1)] 2025-12-04T09:28:45.8298525Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8298659Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8298739Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8298801Z graph_break [] 2025-12-04T09:28:45.8298885Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8299008Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8299071Z frames [('total', 1)] 2025-12-04T09:28:45.8299146Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8299273Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8299348Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8299412Z graph_break [] 2025-12-04T09:28:45.8299489Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8299620Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8299690Z frames [('total', 1)] 2025-12-04T09:28:45.8299762Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8299904Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8299977Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8300037Z graph_break [] 2025-12-04T09:28:45.8300120Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8300243Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8300304Z frames [('total', 1)] 2025-12-04T09:28:45.8300377Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8300502Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8300576Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8300644Z graph_break [] 2025-12-04T09:28:45.8300719Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8300846Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8300909Z frames [('total', 1)] 2025-12-04T09:28:45.8300980Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8301109Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8301181Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8301240Z graph_break [] 2025-12-04T09:28:45.8301324Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8301445Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8301507Z frames [('total', 1)] 2025-12-04T09:28:45.8301582Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8301713Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8301794Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8301854Z graph_break [] 2025-12-04T09:28:45.8301931Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8302067Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8302133Z frames [('total', 1)] 2025-12-04T09:28:45.8302209Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8302344Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8302417Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8302475Z graph_break [] 2025-12-04T09:28:45.8302560Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8302686Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8302756Z frames [('total', 1)] 2025-12-04T09:28:45.8302824Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8302951Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8303075Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8303133Z graph_break [] 2025-12-04T09:28:45.8303213Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8303354Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8303453Z frames [('total', 1)] 2025-12-04T09:28:45.8303643Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8303776Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8303849Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8303907Z graph_break [] 2025-12-04T09:28:45.8303991Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8304113Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8304184Z frames [('total', 1)] 2025-12-04T09:28:45.8304254Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8304379Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8304456Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8304804Z graph_break [] 2025-12-04T09:28:45.8304928Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8305065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8305132Z frames [('total', 1)] 2025-12-04T09:28:45.8305201Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8305337Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8305410Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8305474Z graph_break [] 2025-12-04T09:28:45.8305551Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8305683Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8305751Z frames [('total', 1)] 2025-12-04T09:28:45.8305822Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8305954Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8306037Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8306100Z graph_break [] 2025-12-04T09:28:45.8306180Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8306315Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8306382Z frames [('total', 1)] 2025-12-04T09:28:45.8306456Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8306588Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8306661Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8306725Z graph_break [] 2025-12-04T09:28:45.8306802Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8306928Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8306995Z frames [('total', 1)] 2025-12-04T09:28:45.8307072Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8307201Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8307281Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8307343Z graph_break [] 2025-12-04T09:28:45.8307427Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8307552Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8307617Z frames [('total', 1)] 2025-12-04T09:28:45.8307692Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8307824Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8307898Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8307964Z graph_break [] 2025-12-04T09:28:45.8308041Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8308167Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8308236Z frames [('total', 1)] 2025-12-04T09:28:45.8308315Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8308452Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8308524Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8308682Z graph_break [] 2025-12-04T09:28:45.8308767Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8308893Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8309010Z frames [('total', 1)] 2025-12-04T09:28:45.8309084Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8309306Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8309382Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8309461Z graph_break [] 2025-12-04T09:28:45.8309541Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8309673Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8309734Z frames [('total', 1)] 2025-12-04T09:28:45.8309800Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8309931Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8310002Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8310060Z graph_break [] 2025-12-04T09:28:45.8310144Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8310267Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8310329Z frames [('total', 1)] 2025-12-04T09:28:45.8310405Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8310534Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8310609Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8310669Z graph_break [] 2025-12-04T09:28:45.8310746Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8310876Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8310938Z frames [('total', 1)] 2025-12-04T09:28:45.8311005Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8311137Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8311208Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8311268Z graph_break [] 2025-12-04T09:28:45.8311351Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8311475Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8311543Z frames [('total', 1)] 2025-12-04T09:28:45.8311614Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8311743Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8311821Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8311879Z graph_break [] 2025-12-04T09:28:45.8311954Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8312085Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8312148Z frames [('total', 1)] 2025-12-04T09:28:45.8312216Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8312347Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8312419Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8312479Z graph_break [] 2025-12-04T09:28:45.8312564Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8312691Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8312759Z frames [('total', 1)] 2025-12-04T09:28:45.8312831Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8312956Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8313050Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8313111Z graph_break [] 2025-12-04T09:28:45.8313188Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8313318Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8313380Z frames [('total', 1)] 2025-12-04T09:28:45.8313448Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8313578Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8313651Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8313718Z graph_break [] 2025-12-04T09:28:45.8313794Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8314189Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8314271Z Traceback (most recent call last): 2025-12-04T09:28:45.8314624Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8314746Z f(a, b) 2025-12-04T09:28:45.8315047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8315201Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8315512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8315631Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8315925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8316026Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8316336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8316545Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8316861Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8316935Z graph.run(*example_inputs) 2025-12-04T09:28:45.8317176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8317245Z return super().run(*args) 2025-12-04T09:28:45.8317476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8317560Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8317812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8317893Z result = super().run_node(n) 2025-12-04T09:28:45.8318139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8318249Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8318522Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8318672Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8318939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8319076Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8319332Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8319414Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8319672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8319757Z return autotune_select_algorithm( 2025-12-04T09:28:45.8320107Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8320181Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8320473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8320630Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8321369Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8321492Z target: aten.mm.default 2025-12-04T09:28:45.8321555Z args[0]: TensorBox( 2025-12-04T09:28:45.8321624Z ReinterpretView( 2025-12-04T09:28:45.8321726Z StorageBox( 2025-12-04T09:28:45.8322051Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8322117Z ), 2025-12-04T09:28:45.8322270Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8322343Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8322412Z stack_traces = {, 2025-12-04T09:28:45.8322649Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8322725Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8322789Z , 2025-12-04T09:28:45.8322844Z } 2025-12-04T09:28:45.8322897Z ) 2025-12-04T09:28:45.8322957Z ) 2025-12-04T09:28:45.8323032Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8323271Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8323326Z )) 2025-12-04T09:28:45.8323332Z 2025-12-04T09:28:45.8323755Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8323759Z 2025-12-04T09:28:45.8323762Z 2025-12-04T09:28:45.8323903Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8324146Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8324150Z 2025-12-04T09:28:45.8324318Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8324452Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8324515Z frames [('total', 1)] 2025-12-04T09:28:45.8324593Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8324669Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8324804Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8324875Z graph_break [] 2025-12-04T09:28:45.8324957Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8325097Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8325167Z frames [('total', 1)] 2025-12-04T09:28:45.8325237Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8325371Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8325444Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8325504Z graph_break [] 2025-12-04T09:28:45.8325588Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8325712Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8325773Z frames [('total', 1)] 2025-12-04T09:28:45.8325853Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8325981Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8326060Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8326122Z graph_break [] 2025-12-04T09:28:45.8326201Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8326334Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8326396Z frames [('total', 1)] 2025-12-04T09:28:45.8326464Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8326594Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8326666Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8326726Z graph_break [] 2025-12-04T09:28:45.8326809Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8326934Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8327001Z frames [('total', 1)] 2025-12-04T09:28:45.8327114Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8327242Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8327318Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8327412Z graph_break [] 2025-12-04T09:28:45.8327490Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8327683Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8327749Z frames [('total', 1)] 2025-12-04T09:28:45.8327819Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8327950Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8328034Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8328095Z graph_break [] 2025-12-04T09:28:45.8328179Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8328304Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8328372Z frames [('total', 1)] 2025-12-04T09:28:45.8328441Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8328568Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8328647Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8328706Z graph_break [] 2025-12-04T09:28:45.8328785Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8328923Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8328984Z frames [('total', 1)] 2025-12-04T09:28:45.8329054Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8329184Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8329256Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8329319Z graph_break [] 2025-12-04T09:28:45.8329396Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8329519Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8329591Z frames [('total', 1)] 2025-12-04T09:28:45.8329661Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8329789Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8329870Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8329931Z graph_break [] 2025-12-04T09:28:45.8330009Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8330146Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8330210Z frames [('total', 1)] 2025-12-04T09:28:45.8330286Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8330412Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8330485Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8330548Z graph_break [] 2025-12-04T09:28:45.8330625Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8330751Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8330821Z frames [('total', 1)] 2025-12-04T09:28:45.8330892Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8331017Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8331095Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8331156Z graph_break [] 2025-12-04T09:28:45.8331239Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8331367Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8331432Z frames [('total', 1)] 2025-12-04T09:28:45.8331505Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8331631Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8331702Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8331766Z graph_break [] 2025-12-04T09:28:45.8331842Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8331967Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8332045Z frames [('total', 1)] 2025-12-04T09:28:45.8332114Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8332294Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8332365Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8332421Z graph_break [] 2025-12-04T09:28:45.8332502Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8332664Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8332793Z frames [('total', 1)] 2025-12-04T09:28:45.8332870Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8332997Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8333069Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8333238Z graph_break [] 2025-12-04T09:28:45.8333317Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8333447Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8333508Z frames [('total', 1)] 2025-12-04T09:28:45.8333575Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8333702Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8333776Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8333834Z graph_break [] 2025-12-04T09:28:45.8333916Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8334043Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8334110Z frames [('total', 1)] 2025-12-04T09:28:45.8334183Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8334306Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8334382Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8334441Z graph_break [] 2025-12-04T09:28:45.8334516Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8334647Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8334709Z frames [('total', 1)] 2025-12-04T09:28:45.8334777Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8334907Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8334980Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8335038Z graph_break [] 2025-12-04T09:28:45.8335120Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8335247Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8335319Z frames [('total', 1)] 2025-12-04T09:28:45.8335398Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8335524Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8335600Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8335660Z graph_break [] 2025-12-04T09:28:45.8335736Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8335869Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8335932Z frames [('total', 1)] 2025-12-04T09:28:45.8336000Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8336130Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8336205Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8336263Z graph_break [] 2025-12-04T09:28:45.8336344Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8336471Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8336539Z frames [('total', 1)] 2025-12-04T09:28:45.8336612Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8336737Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8336815Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8336875Z graph_break [] 2025-12-04T09:28:45.8336951Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8337082Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8337145Z frames [('total', 1)] 2025-12-04T09:28:45.8337212Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8337345Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8337468Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8337538Z graph_break [] 2025-12-04T09:28:45.8337615Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8337739Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8337854Z frames [('total', 1)] 2025-12-04T09:28:45.8337984Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8338112Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8338197Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8338261Z graph_break [] 2025-12-04T09:28:45.8338338Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8338469Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8338534Z frames [('total', 1)] 2025-12-04T09:28:45.8338610Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8338735Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8338810Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8338876Z graph_break [] 2025-12-04T09:28:45.8338953Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8339077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8339150Z frames [('total', 1)] 2025-12-04T09:28:45.8339223Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8339349Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8339430Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8339491Z graph_break [] 2025-12-04T09:28:45.8339575Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8339699Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8339764Z frames [('total', 1)] 2025-12-04T09:28:45.8339841Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8339966Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8340041Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8340107Z graph_break [] 2025-12-04T09:28:45.8340186Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8340318Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8340391Z frames [('total', 1)] 2025-12-04T09:28:45.8340460Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8340593Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8340664Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8340724Z graph_break [] 2025-12-04T09:28:45.8340807Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8340931Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8340996Z frames [('total', 1)] 2025-12-04T09:28:45.8341069Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8341194Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8341266Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8341336Z graph_break [] 2025-12-04T09:28:45.8341411Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8341540Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8341605Z frames [('total', 1)] 2025-12-04T09:28:45.8341672Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8341807Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8341880Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8341940Z graph_break [] 2025-12-04T09:28:45.8342023Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8342147Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8342209Z frames [('total', 1)] 2025-12-04T09:28:45.8342291Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8342418Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8342495Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8342606Z graph_break [] 2025-12-04T09:28:45.8342682Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8342815Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8342933Z frames [('total', 1)] 2025-12-04T09:28:45.8343001Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8343195Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8343269Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8343328Z graph_break [] 2025-12-04T09:28:45.8343412Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8343535Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8343603Z frames [('total', 1)] 2025-12-04T09:28:45.8343673Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8343798Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8343875Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8343937Z graph_break [] 2025-12-04T09:28:45.8344013Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8344182Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8344262Z Traceback (most recent call last): 2025-12-04T09:28:45.8344581Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8344646Z f(a, b) 2025-12-04T09:28:45.8344936Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8345094Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8345401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8345520Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8345822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8345918Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8346238Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8346441Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8346744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8346836Z graph.run(*example_inputs) 2025-12-04T09:28:45.8347077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8347147Z return super().run(*args) 2025-12-04T09:28:45.8347389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8347471Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8347732Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8347802Z result = super().run_node(n) 2025-12-04T09:28:45.8348050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8348166Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8348432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8348579Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8348850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8348981Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8349246Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8349364Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8349626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8349747Z return autotune_select_algorithm( 2025-12-04T09:28:45.8350151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8350232Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8350520Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8350680Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8351434Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8351521Z target: aten.mm.default 2025-12-04T09:28:45.8351588Z args[0]: TensorBox( 2025-12-04T09:28:45.8351663Z ReinterpretView( 2025-12-04T09:28:45.8351724Z StorageBox( 2025-12-04T09:28:45.8351977Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8352033Z ), 2025-12-04T09:28:45.8352181Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8352262Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8352328Z stack_traces = {, 2025-12-04T09:28:45.8352554Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8352640Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8352697Z , 2025-12-04T09:28:45.8352751Z } 2025-12-04T09:28:45.8352817Z ) 2025-12-04T09:28:45.8352871Z ) 2025-12-04T09:28:45.8352960Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8353194Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8353251Z )) 2025-12-04T09:28:45.8353254Z 2025-12-04T09:28:45.8353683Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8353687Z 2025-12-04T09:28:45.8353690Z 2025-12-04T09:28:45.8353828Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8354075Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8354079Z 2025-12-04T09:28:45.8354241Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8354371Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8354442Z frames [('total', 1)] 2025-12-04T09:28:45.8354512Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8354588Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8354732Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8354793Z graph_break [] 2025-12-04T09:28:45.8354884Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8355009Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8355072Z frames [('total', 1)] 2025-12-04T09:28:45.8355145Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8355273Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8355348Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8355411Z graph_break [] 2025-12-04T09:28:45.8355488Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8355612Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8355724Z frames [('total', 1)] 2025-12-04T09:28:45.8355792Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8355922Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8356041Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8356103Z graph_break [] 2025-12-04T09:28:45.8356249Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8356377Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8356440Z frames [('total', 1)] 2025-12-04T09:28:45.8356514Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8356640Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8356711Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8356780Z graph_break [] 2025-12-04T09:28:45.8356857Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8356986Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8357052Z frames [('total', 1)] 2025-12-04T09:28:45.8357119Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8357251Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8357326Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8357384Z graph_break [] 2025-12-04T09:28:45.8357471Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8357598Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8357660Z frames [('total', 1)] 2025-12-04T09:28:45.8357744Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8357873Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8357952Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8358012Z graph_break [] 2025-12-04T09:28:45.8358089Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8358219Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8358283Z frames [('total', 1)] 2025-12-04T09:28:45.8358352Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8358486Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8358558Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8358620Z graph_break [] 2025-12-04T09:28:45.8358704Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8358833Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8358898Z frames [('total', 1)] 2025-12-04T09:28:45.8358978Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8359108Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8359187Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8359246Z graph_break [] 2025-12-04T09:28:45.8359323Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8359454Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8359528Z frames [('total', 1)] 2025-12-04T09:28:45.8359601Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8359734Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8359806Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8359869Z graph_break [] 2025-12-04T09:28:45.8359951Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8360077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8360147Z frames [('total', 1)] 2025-12-04T09:28:45.8360215Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8360339Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8360419Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8360478Z graph_break [] 2025-12-04T09:28:45.8360558Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8360689Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8360751Z frames [('total', 1)] 2025-12-04T09:28:45.8360866Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8361000Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8361077Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8361182Z graph_break [] 2025-12-04T09:28:45.8361261Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8361449Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8361522Z frames [('total', 1)] 2025-12-04T09:28:45.8361589Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8361714Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8361795Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8361856Z graph_break [] 2025-12-04T09:28:45.8361933Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8362067Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8362133Z frames [('total', 1)] 2025-12-04T09:28:45.8362212Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8362338Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8362411Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8362486Z graph_break [] 2025-12-04T09:28:45.8362569Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8362698Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8362768Z frames [('total', 1)] 2025-12-04T09:28:45.8362836Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8362962Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8363038Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8363097Z graph_break [] 2025-12-04T09:28:45.8363178Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8363302Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8363363Z frames [('total', 1)] 2025-12-04T09:28:45.8363437Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8363566Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8363636Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8363700Z graph_break [] 2025-12-04T09:28:45.8363778Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8363907Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8363975Z frames [('total', 1)] 2025-12-04T09:28:45.8364044Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8364178Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8364250Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8364312Z graph_break [] 2025-12-04T09:28:45.8364393Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8364519Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8364580Z frames [('total', 1)] 2025-12-04T09:28:45.8364662Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8364793Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8364866Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8364931Z graph_break [] 2025-12-04T09:28:45.8365009Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8365141Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8365204Z frames [('total', 1)] 2025-12-04T09:28:45.8365271Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8365400Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8365475Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8365533Z graph_break [] 2025-12-04T09:28:45.8365616Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8365738Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8365800Z frames [('total', 1)] 2025-12-04T09:28:45.8365873Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8366059Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8366138Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8366208Z graph_break [] 2025-12-04T09:28:45.8366323Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8366518Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8366582Z frames [('total', 1)] 2025-12-04T09:28:45.8366653Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8366784Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8366856Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8366916Z graph_break [] 2025-12-04T09:28:45.8366999Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8367124Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8367194Z frames [('total', 1)] 2025-12-04T09:28:45.8367265Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8367391Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8367469Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8367529Z graph_break [] 2025-12-04T09:28:45.8367605Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8367740Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8367807Z frames [('total', 1)] 2025-12-04T09:28:45.8367877Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8368007Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8368080Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8368138Z graph_break [] 2025-12-04T09:28:45.8368221Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8368356Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8368427Z frames [('total', 1)] 2025-12-04T09:28:45.8368494Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8368622Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8368706Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8368766Z graph_break [] 2025-12-04T09:28:45.8368844Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8368979Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8369047Z frames [('total', 1)] 2025-12-04T09:28:45.8369117Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8369249Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8369323Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8369389Z graph_break [] 2025-12-04T09:28:45.8369466Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8369591Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8369658Z frames [('total', 1)] 2025-12-04T09:28:45.8369729Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8369856Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8369938Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8370002Z graph_break [] 2025-12-04T09:28:45.8370078Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8370212Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8370279Z frames [('total', 1)] 2025-12-04T09:28:45.8370353Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8370478Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8370553Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8370620Z graph_break [] 2025-12-04T09:28:45.8370700Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8370823Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8370891Z frames [('total', 1)] 2025-12-04T09:28:45.8370960Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8371086Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8371210Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8371270Z graph_break [] 2025-12-04T09:28:45.8371352Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8371512Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8371574Z frames [('total', 1)] 2025-12-04T09:28:45.8371721Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8371850Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8371923Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8371989Z graph_break [] 2025-12-04T09:28:45.8372064Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8372189Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8372256Z frames [('total', 1)] 2025-12-04T09:28:45.8372323Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8372452Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8372526Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8372584Z graph_break [] 2025-12-04T09:28:45.8372664Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8372791Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8372853Z frames [('total', 1)] 2025-12-04T09:28:45.8372929Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8373055Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8373240Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8373311Z graph_break [] 2025-12-04T09:28:45.8373390Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8373521Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8373582Z frames [('total', 1)] 2025-12-04T09:28:45.8373650Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8373782Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8373856Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8373916Z graph_break [] 2025-12-04T09:28:45.8373998Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8374121Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8374187Z frames [('total', 1)] 2025-12-04T09:28:45.8374266Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8374391Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8374470Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8374529Z graph_break [] 2025-12-04T09:28:45.8374611Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8374782Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8374857Z Traceback (most recent call last): 2025-12-04T09:28:45.8375168Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8375235Z f(a, b) 2025-12-04T09:28:45.8375523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8375673Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8375987Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8376106Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8376408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8376500Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8376822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8377024Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8377374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8377455Z graph.run(*example_inputs) 2025-12-04T09:28:45.8377729Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8377861Z return super().run(*args) 2025-12-04T09:28:45.8378101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8378179Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8378429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8378508Z result = super().run_node(n) 2025-12-04T09:28:45.8378756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8378863Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8379146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8379301Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8379580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8379712Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8379981Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8380056Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8380317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8380405Z return autotune_select_algorithm( 2025-12-04T09:28:45.8380756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8380830Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8381129Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8381297Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8382047Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8382119Z target: aten.mm.default 2025-12-04T09:28:45.8382182Z args[0]: TensorBox( 2025-12-04T09:28:45.8382256Z ReinterpretView( 2025-12-04T09:28:45.8382316Z StorageBox( 2025-12-04T09:28:45.8382566Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8382635Z ), 2025-12-04T09:28:45.8382784Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8382864Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8382932Z stack_traces = {, 2025-12-04T09:28:45.8383165Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8383253Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8383309Z , 2025-12-04T09:28:45.8383364Z } 2025-12-04T09:28:45.8383427Z ) 2025-12-04T09:28:45.8383482Z ) 2025-12-04T09:28:45.8383552Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8383792Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8383849Z )) 2025-12-04T09:28:45.8383852Z 2025-12-04T09:28:45.8384279Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8384329Z 2025-12-04T09:28:45.8384332Z 2025-12-04T09:28:45.8384468Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8384747Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8384837Z 2025-12-04T09:28:45.8385003Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8385135Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8385204Z frames [('total', 1)] 2025-12-04T09:28:45.8385274Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8385346Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8385486Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8385556Z graph_break [] 2025-12-04T09:28:45.8385638Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8385773Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8385837Z frames [('total', 1)] 2025-12-04T09:28:45.8385913Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8386044Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8386117Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8386185Z graph_break [] 2025-12-04T09:28:45.8386262Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8386392Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8386460Z frames [('total', 1)] 2025-12-04T09:28:45.8386527Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8386652Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8386730Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8386788Z graph_break [] 2025-12-04T09:28:45.8386870Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8386995Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8387061Z frames [('total', 1)] 2025-12-04T09:28:45.8387134Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8387259Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8387335Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8387403Z graph_break [] 2025-12-04T09:28:45.8387482Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8387607Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8387675Z frames [('total', 1)] 2025-12-04T09:28:45.8387748Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8387882Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8387956Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8388015Z graph_break [] 2025-12-04T09:28:45.8388099Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8388223Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8388288Z frames [('total', 1)] 2025-12-04T09:28:45.8388362Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8388487Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8388563Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8388641Z graph_break [] 2025-12-04T09:28:45.8388732Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8388872Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8388937Z frames [('total', 1)] 2025-12-04T09:28:45.8389010Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8389151Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8389231Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8389294Z graph_break [] 2025-12-04T09:28:45.8389383Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8389513Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8389627Z frames [('total', 1)] 2025-12-04T09:28:45.8389704Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8389834Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8389954Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8390023Z graph_break [] 2025-12-04T09:28:45.8390170Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8390307Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8390369Z frames [('total', 1)] 2025-12-04T09:28:45.8390439Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8390571Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8390643Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8390703Z graph_break [] 2025-12-04T09:28:45.8390789Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8390914Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8390986Z frames [('total', 1)] 2025-12-04T09:28:45.8391056Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8391185Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8391267Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8391327Z graph_break [] 2025-12-04T09:28:45.8391411Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8391543Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8391607Z frames [('total', 1)] 2025-12-04T09:28:45.8391676Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8391819Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8391895Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8391954Z graph_break [] 2025-12-04T09:28:45.8392038Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8392164Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8392236Z frames [('total', 1)] 2025-12-04T09:28:45.8392305Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8392435Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8392518Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8392581Z graph_break [] 2025-12-04T09:28:45.8392662Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8392799Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8392862Z frames [('total', 1)] 2025-12-04T09:28:45.8392934Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8393068Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8393140Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8393204Z graph_break [] 2025-12-04T09:28:45.8393280Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8393408Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8393477Z frames [('total', 1)] 2025-12-04T09:28:45.8393546Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8393671Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8393754Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8393813Z graph_break [] 2025-12-04T09:28:45.8393895Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8394027Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8394090Z frames [('total', 1)] 2025-12-04T09:28:45.8394162Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8394288Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8394361Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8394425Z graph_break [] 2025-12-04T09:28:45.8394502Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8394628Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8394748Z frames [('total', 1)] 2025-12-04T09:28:45.8394816Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8394941Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8395019Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8395117Z graph_break [] 2025-12-04T09:28:45.8395198Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8395441Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8395505Z frames [('total', 1)] 2025-12-04T09:28:45.8395577Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8395703Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8395774Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8395842Z graph_break [] 2025-12-04T09:28:45.8395920Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8396045Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8396113Z frames [('total', 1)] 2025-12-04T09:28:45.8396183Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8396315Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8396386Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8396449Z graph_break [] 2025-12-04T09:28:45.8396532Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8396662Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8396724Z frames [('total', 1)] 2025-12-04T09:28:45.8396799Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8396926Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8396998Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8397064Z graph_break [] 2025-12-04T09:28:45.8397141Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8397272Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8397336Z frames [('total', 1)] 2025-12-04T09:28:45.8397406Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8397539Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8397612Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8397674Z graph_break [] 2025-12-04T09:28:45.8397761Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8397890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8397954Z frames [('total', 1)] 2025-12-04T09:28:45.8398031Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8398156Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8398235Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8398295Z graph_break [] 2025-12-04T09:28:45.8398372Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8398512Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8398577Z frames [('total', 1)] 2025-12-04T09:28:45.8398647Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8398779Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8398852Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8398914Z graph_break [] 2025-12-04T09:28:45.8398999Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8399128Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8399197Z frames [('total', 1)] 2025-12-04T09:28:45.8399267Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8399394Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8399473Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8399533Z graph_break [] 2025-12-04T09:28:45.8399611Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8399743Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8399806Z frames [('total', 1)] 2025-12-04T09:28:45.8399873Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8400056Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8400128Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8400188Z graph_break [] 2025-12-04T09:28:45.8400307Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8400507Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8400581Z frames [('total', 1)] 2025-12-04T09:28:45.8400653Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8400781Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8400859Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8400919Z graph_break [] 2025-12-04T09:28:45.8400995Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8401127Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8401191Z frames [('total', 1)] 2025-12-04T09:28:45.8401259Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8401395Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8401467Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8401531Z graph_break [] 2025-12-04T09:28:45.8401609Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8401739Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8401809Z frames [('total', 1)] 2025-12-04T09:28:45.8401876Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8402001Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8402080Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8402139Z graph_break [] 2025-12-04T09:28:45.8402216Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8402351Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8402413Z frames [('total', 1)] 2025-12-04T09:28:45.8402484Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8402612Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8402686Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8402750Z graph_break [] 2025-12-04T09:28:45.8402832Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8402960Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8403030Z frames [('total', 1)] 2025-12-04T09:28:45.8403097Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8403223Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8403302Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8403372Z graph_break [] 2025-12-04T09:28:45.8403457Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8403580Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8403642Z frames [('total', 1)] 2025-12-04T09:28:45.8403714Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8403841Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8403913Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8403977Z graph_break [] 2025-12-04T09:28:45.8404054Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8404181Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8404254Z frames [('total', 1)] 2025-12-04T09:28:45.8404324Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8404461Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8404836Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8404958Z graph_break [] 2025-12-04T09:28:45.8405053Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8405188Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8405251Z frames [('total', 1)] 2025-12-04T09:28:45.8405332Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8405466Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8405639Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8405707Z graph_break [] 2025-12-04T09:28:45.8405785Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8405977Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8406158Z frames [('total', 1)] 2025-12-04T09:28:45.8406230Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8406365Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8406437Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8406498Z graph_break [] 2025-12-04T09:28:45.8406582Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8406749Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8406828Z Traceback (most recent call last): 2025-12-04T09:28:45.8407160Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8407220Z f(a, b) 2025-12-04T09:28:45.8407530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8407701Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8408017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8408147Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8408446Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8408617Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8408937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8409143Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8409459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8409547Z graph.run(*example_inputs) 2025-12-04T09:28:45.8409797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8409876Z return super().run(*args) 2025-12-04T09:28:45.8410130Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8410226Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8410493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8410584Z result = super().run_node(n) 2025-12-04T09:28:45.8410834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8410949Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8411220Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8411373Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8411651Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8411785Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8412054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8412130Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8412394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8412488Z return autotune_select_algorithm( 2025-12-04T09:28:45.8412839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8412965Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8413380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8413649Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8414408Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8414480Z target: aten.mm.default 2025-12-04T09:28:45.8414544Z args[0]: TensorBox( 2025-12-04T09:28:45.8414618Z ReinterpretView( 2025-12-04T09:28:45.8414678Z StorageBox( 2025-12-04T09:28:45.8414920Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8414985Z ), 2025-12-04T09:28:45.8415136Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8415220Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8415285Z stack_traces = {, 2025-12-04T09:28:45.8415516Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8415599Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8415655Z , 2025-12-04T09:28:45.8415712Z } 2025-12-04T09:28:45.8415773Z ) 2025-12-04T09:28:45.8415827Z ) 2025-12-04T09:28:45.8415902Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8416144Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8416198Z )) 2025-12-04T09:28:45.8416202Z 2025-12-04T09:28:45.8416631Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8416637Z 2025-12-04T09:28:45.8416640Z 2025-12-04T09:28:45.8416780Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8417026Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8417034Z 2025-12-04T09:28:45.8417208Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8417342Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8417410Z frames [('total', 1)] 2025-12-04T09:28:45.8417481Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8417556Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8417696Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8417757Z graph_break [] 2025-12-04T09:28:45.8417839Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8417986Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8418051Z frames [('total', 1)] 2025-12-04T09:28:45.8418128Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8418261Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8418338Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8418404Z graph_break [] 2025-12-04T09:28:45.8418482Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8418607Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8418677Z frames [('total', 1)] 2025-12-04T09:28:45.8418748Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8418876Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8418957Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8419019Z graph_break [] 2025-12-04T09:28:45.8419102Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8419274Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8419338Z frames [('total', 1)] 2025-12-04T09:28:45.8419412Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8419574Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8419714Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8419784Z graph_break [] 2025-12-04T09:28:45.8419861Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8419986Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8420054Z frames [('total', 1)] 2025-12-04T09:28:45.8420123Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8420255Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8420327Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8420387Z graph_break [] 2025-12-04T09:28:45.8420470Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8420599Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8420662Z frames [('total', 1)] 2025-12-04T09:28:45.8420736Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8420866Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8420940Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8421008Z graph_break [] 2025-12-04T09:28:45.8421099Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8421228Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8421296Z frames [('total', 1)] 2025-12-04T09:28:45.8421364Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8421496Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8421568Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8421628Z graph_break [] 2025-12-04T09:28:45.8421710Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8421836Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8421906Z frames [('total', 1)] 2025-12-04T09:28:45.8431794Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8432093Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8432201Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8432303Z graph_break [] 2025-12-04T09:28:45.8432419Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8432617Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8432706Z frames [('total', 1)] 2025-12-04T09:28:45.8432805Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8432996Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8433096Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8433175Z graph_break [] 2025-12-04T09:28:45.8433293Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8433472Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8433556Z frames [('total', 1)] 2025-12-04T09:28:45.8433655Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8433827Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8433937Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8434020Z graph_break [] 2025-12-04T09:28:45.8434127Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8434302Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8434384Z frames [('total', 1)] 2025-12-04T09:28:45.8434477Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8434664Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8434768Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8434845Z graph_break [] 2025-12-04T09:28:45.8434952Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8435112Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8435295Z frames [('total', 1)] 2025-12-04T09:28:45.8435383Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8435551Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8435711Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8435874Z graph_break [] 2025-12-04T09:28:45.8435986Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8436169Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8436251Z frames [('total', 1)] 2025-12-04T09:28:45.8436344Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8436527Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8436623Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8436710Z graph_break [] 2025-12-04T09:28:45.8436814Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8437045Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8437122Z frames [('total', 1)] 2025-12-04T09:28:45.8437195Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8437331Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8437423Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8437485Z graph_break [] 2025-12-04T09:28:45.8437571Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8437707Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8437778Z frames [('total', 1)] 2025-12-04T09:28:45.8437853Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8437997Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8438077Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8438149Z graph_break [] 2025-12-04T09:28:45.8438230Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8438378Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8438453Z frames [('total', 1)] 2025-12-04T09:28:45.8438525Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8438655Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8438740Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8438802Z graph_break [] 2025-12-04T09:28:45.8438890Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8439031Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8439097Z frames [('total', 1)] 2025-12-04T09:28:45.8439175Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8439304Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8439384Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8439452Z graph_break [] 2025-12-04T09:28:45.8439531Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8439671Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8439747Z frames [('total', 1)] 2025-12-04T09:28:45.8439818Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8439950Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8440037Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8440097Z graph_break [] 2025-12-04T09:28:45.8440186Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8440314Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8440380Z frames [('total', 1)] 2025-12-04T09:28:45.8440456Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8440585Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8440665Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8440731Z graph_break [] 2025-12-04T09:28:45.8440809Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8440938Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8441056Z frames [('total', 1)] 2025-12-04T09:28:45.8441129Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8441267Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8441379Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8441440Z graph_break [] 2025-12-04T09:28:45.8441590Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8441723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8441788Z frames [('total', 1)] 2025-12-04T09:28:45.8441865Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8441992Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8442074Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8442145Z graph_break [] 2025-12-04T09:28:45.8442227Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8442366Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8442435Z frames [('total', 1)] 2025-12-04T09:28:45.8442506Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8442644Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8442721Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8442788Z graph_break [] 2025-12-04T09:28:45.8463607Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8463871Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8463967Z frames [('total', 1)] 2025-12-04T09:28:45.8464063Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8464245Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8464344Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8464418Z graph_break [] 2025-12-04T09:28:45.8464520Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8464690Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8464773Z frames [('total', 1)] 2025-12-04T09:28:45.8464860Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8465028Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8465116Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8465193Z graph_break [] 2025-12-04T09:28:45.8465297Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8465465Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8465545Z frames [('total', 1)] 2025-12-04T09:28:45.8465629Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8465787Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8465879Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8465953Z graph_break [] 2025-12-04T09:28:45.8466045Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8466206Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8466282Z frames [('total', 1)] 2025-12-04T09:28:45.8466365Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8466521Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8466607Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8466684Z graph_break [] 2025-12-04T09:28:45.8466777Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8466934Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8467015Z frames [('total', 1)] 2025-12-04T09:28:45.8467095Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8467251Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8467355Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8467413Z graph_break [] 2025-12-04T09:28:45.8467485Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8467611Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8467671Z frames [('total', 1)] 2025-12-04T09:28:45.8467801Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8467932Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8468002Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8468111Z graph_break [] 2025-12-04T09:28:45.8468191Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8468386Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8468449Z frames [('total', 1)] 2025-12-04T09:28:45.8468514Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8468634Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8468707Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8468765Z graph_break [] 2025-12-04T09:28:45.8468842Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8468978Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8469041Z frames [('total', 1)] 2025-12-04T09:28:45.8469115Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8469243Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8469313Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8469375Z graph_break [] 2025-12-04T09:28:45.8469452Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8469584Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8469650Z frames [('total', 1)] 2025-12-04T09:28:45.8469717Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8469846Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8469922Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8469979Z graph_break [] 2025-12-04T09:28:45.8470055Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8470180Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8470241Z frames [('total', 1)] 2025-12-04T09:28:45.8470310Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8470441Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8470510Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8470571Z graph_break [] 2025-12-04T09:28:45.8470647Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8470780Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8470852Z frames [('total', 1)] 2025-12-04T09:28:45.8470923Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8471052Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8471123Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8471178Z graph_break [] 2025-12-04T09:28:45.8471254Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8471378Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8471437Z frames [('total', 1)] 2025-12-04T09:28:45.8471545Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8471671Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8471740Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8471800Z graph_break [] 2025-12-04T09:28:45.8471877Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8472048Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8472123Z Traceback (most recent call last): 2025-12-04T09:28:45.8472457Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8472516Z f(a, b) 2025-12-04T09:28:45.8472818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8472974Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8473283Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8473453Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8473752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8473881Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8474277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8474487Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8474787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8474860Z graph.run(*example_inputs) 2025-12-04T09:28:45.8475093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8475160Z return super().run(*args) 2025-12-04T09:28:45.8475397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8475475Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8475738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8475814Z result = super().run_node(n) 2025-12-04T09:28:45.8476071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8476186Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8476477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8476627Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8476898Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8477034Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8477309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8477383Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8477658Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8477746Z return autotune_select_algorithm( 2025-12-04T09:28:45.8478089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8478167Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8478482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8478652Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8479439Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8479515Z target: aten.mm.default 2025-12-04T09:28:45.8479583Z args[0]: TensorBox( 2025-12-04T09:28:45.8479658Z ReinterpretView( 2025-12-04T09:28:45.8479719Z StorageBox( 2025-12-04T09:28:45.8479976Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8480036Z ), 2025-12-04T09:28:45.8480192Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8480276Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8480343Z stack_traces = {, 2025-12-04T09:28:45.8480581Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8480714Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8480781Z , 2025-12-04T09:28:45.8480837Z } 2025-12-04T09:28:45.8480898Z ) 2025-12-04T09:28:45.8480952Z ) 2025-12-04T09:28:45.8481069Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8481384Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8481439Z )) 2025-12-04T09:28:45.8481444Z 2025-12-04T09:28:45.8481877Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8481887Z 2025-12-04T09:28:45.8481890Z 2025-12-04T09:28:45.8482032Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8482290Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8482294Z 2025-12-04T09:28:45.8482469Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8482603Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8482673Z frames [('total', 1)] 2025-12-04T09:28:45.8482752Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8482829Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8482979Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8483042Z graph_break [] 2025-12-04T09:28:45.8483126Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8483266Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8483331Z frames [('total', 1)] 2025-12-04T09:28:45.8483400Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8483533Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8483606Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8483676Z graph_break [] 2025-12-04T09:28:45.8483770Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8483896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8483965Z frames [('total', 1)] 2025-12-04T09:28:45.8484037Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8484168Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8484249Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8484309Z graph_break [] 2025-12-04T09:28:45.8484386Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8484521Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8484585Z frames [('total', 1)] 2025-12-04T09:28:45.8484654Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8484791Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8484864Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8484931Z graph_break [] 2025-12-04T09:28:45.8485011Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8485134Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8485203Z frames [('total', 1)] 2025-12-04T09:28:45.8485275Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8485411Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8485498Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8485558Z graph_break [] 2025-12-04T09:28:45.8485636Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8485772Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8485834Z frames [('total', 1)] 2025-12-04T09:28:45.8485913Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8486038Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8486112Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8486178Z graph_break [] 2025-12-04T09:28:45.8486306Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8486433Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8486503Z frames [('total', 1)] 2025-12-04T09:28:45.8486605Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8486733Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8486873Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8486932Z graph_break [] 2025-12-04T09:28:45.8487021Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8487148Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8487209Z frames [('total', 1)] 2025-12-04T09:28:45.8487282Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8487405Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8487478Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8487541Z graph_break [] 2025-12-04T09:28:45.8487617Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8487749Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8487815Z frames [('total', 1)] 2025-12-04T09:28:45.8487881Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8488021Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8488100Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8488166Z graph_break [] 2025-12-04T09:28:45.8488247Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8488373Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8488433Z frames [('total', 1)] 2025-12-04T09:28:45.8488506Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8488631Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8488702Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8488767Z graph_break [] 2025-12-04T09:28:45.8488842Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8488972Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8489032Z frames [('total', 1)] 2025-12-04T09:28:45.8489098Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8489232Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8489306Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8489363Z graph_break [] 2025-12-04T09:28:45.8489445Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8489570Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8489633Z frames [('total', 1)] 2025-12-04T09:28:45.8489705Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8489833Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8489903Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8489970Z graph_break [] 2025-12-04T09:28:45.8490049Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8490182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8490243Z frames [('total', 1)] 2025-12-04T09:28:45.8490311Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8490447Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8490523Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8490584Z graph_break [] 2025-12-04T09:28:45.8490667Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8490792Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8490857Z frames [('total', 1)] 2025-12-04T09:28:45.8490929Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8491055Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8491148Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8491207Z graph_break [] 2025-12-04T09:28:45.8491285Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8491464Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8491525Z frames [('total', 1)] 2025-12-04T09:28:45.8491596Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8491763Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8491835Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8491960Z graph_break [] 2025-12-04T09:28:45.8492048Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8492176Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8492242Z frames [('total', 1)] 2025-12-04T09:28:45.8492309Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8492436Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8492511Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8492570Z graph_break [] 2025-12-04T09:28:45.8492647Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8492782Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8492845Z frames [('total', 1)] 2025-12-04T09:28:45.8492913Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8493045Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8493252Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8493329Z graph_break [] 2025-12-04T09:28:45.8493411Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8493535Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8493603Z frames [('total', 1)] 2025-12-04T09:28:45.8493671Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8493803Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8493882Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8493941Z graph_break [] 2025-12-04T09:28:45.8494020Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8494159Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8494226Z frames [('total', 1)] 2025-12-04T09:28:45.8494298Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8494428Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8494503Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8494567Z graph_break [] 2025-12-04T09:28:45.8494648Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8494773Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8494847Z frames [('total', 1)] 2025-12-04T09:28:45.8494917Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8495041Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8495115Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8495175Z graph_break [] 2025-12-04T09:28:45.8495255Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8495383Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8495445Z frames [('total', 1)] 2025-12-04T09:28:45.8495516Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8495641Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8495716Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8495784Z graph_break [] 2025-12-04T09:28:45.8495864Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8495991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8496057Z frames [('total', 1)] 2025-12-04T09:28:45.8496126Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8496255Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8496328Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8496387Z graph_break [] 2025-12-04T09:28:45.8496468Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8496591Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8496714Z frames [('total', 1)] 2025-12-04T09:28:45.8496786Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8496910Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8497039Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8497101Z graph_break [] 2025-12-04T09:28:45.8497239Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8497373Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8497434Z frames [('total', 1)] 2025-12-04T09:28:45.8497501Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8497629Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8497704Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8497762Z graph_break [] 2025-12-04T09:28:45.8497846Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8497974Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8498038Z frames [('total', 1)] 2025-12-04T09:28:45.8498112Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8498235Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8498311Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8498374Z graph_break [] 2025-12-04T09:28:45.8498453Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8498582Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8498644Z frames [('total', 1)] 2025-12-04T09:28:45.8498712Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8498845Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8498915Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8498975Z graph_break [] 2025-12-04T09:28:45.8499062Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8499193Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8499260Z frames [('total', 1)] 2025-12-04T09:28:45.8499337Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8499460Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8499541Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8499601Z graph_break [] 2025-12-04T09:28:45.8499683Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8499815Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8499877Z frames [('total', 1)] 2025-12-04T09:28:45.8499945Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8500075Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8500145Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8500203Z graph_break [] 2025-12-04T09:28:45.8500287Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8500409Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8500481Z frames [('total', 1)] 2025-12-04T09:28:45.8500550Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8500678Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8500766Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8500827Z graph_break [] 2025-12-04T09:28:45.8500904Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8501039Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8501100Z frames [('total', 1)] 2025-12-04T09:28:45.8501168Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8501298Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8501368Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8501436Z graph_break [] 2025-12-04T09:28:45.8501513Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8501636Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8501709Z frames [('total', 1)] 2025-12-04T09:28:45.8501824Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8501949Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8502028Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8502167Z graph_break [] 2025-12-04T09:28:45.8502246Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8502448Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8502513Z frames [('total', 1)] 2025-12-04T09:28:45.8502590Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8502724Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8502799Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8502865Z graph_break [] 2025-12-04T09:28:45.8502948Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8503081Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8503156Z frames [('total', 1)] 2025-12-04T09:28:45.8503229Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8503363Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8503442Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8503503Z graph_break [] 2025-12-04T09:28:45.8503599Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8503733Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8503799Z frames [('total', 1)] 2025-12-04T09:28:45.8503876Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8504006Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8504080Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8504147Z graph_break [] 2025-12-04T09:28:45.8504225Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8504357Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8504419Z frames [('total', 1)] 2025-12-04T09:28:45.8504490Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8505013Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8505092Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8505158Z graph_break [] 2025-12-04T09:28:45.8505247Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8505430Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8505510Z Traceback (most recent call last): 2025-12-04T09:28:45.8505847Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8505907Z f(a, b) 2025-12-04T09:28:45.8506214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8506370Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8506675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8506803Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8507113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8507217Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8507531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8507729Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8508038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8508115Z graph.run(*example_inputs) 2025-12-04T09:28:45.8508353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8508547Z return super().run(*args) 2025-12-04T09:28:45.8508785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8508869Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8509286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8509362Z result = super().run_node(n) 2025-12-04T09:28:45.8509616Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8509720Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8509989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8510145Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8510409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8510553Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8510815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8510890Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8511176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8511264Z return autotune_select_algorithm( 2025-12-04T09:28:45.8511653Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8511731Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8512023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8512190Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8512960Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8513043Z target: aten.mm.default 2025-12-04T09:28:45.8513108Z args[0]: TensorBox( 2025-12-04T09:28:45.8513174Z ReinterpretView( 2025-12-04T09:28:45.8513238Z StorageBox( 2025-12-04T09:28:45.8513485Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8513540Z ), 2025-12-04T09:28:45.8513697Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8513770Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8513835Z stack_traces = {, 2025-12-04T09:28:45.8514068Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8514148Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8514209Z , 2025-12-04T09:28:45.8514262Z } 2025-12-04T09:28:45.8514321Z ) 2025-12-04T09:28:45.8514383Z ) 2025-12-04T09:28:45.8514456Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8514696Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8514756Z )) 2025-12-04T09:28:45.8514760Z 2025-12-04T09:28:45.8515191Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8515195Z 2025-12-04T09:28:45.8515198Z 2025-12-04T09:28:45.8515343Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8515591Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8515640Z 2025-12-04T09:28:45.8515810Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8515952Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8516051Z frames [('total', 1)] 2025-12-04T09:28:45.8516129Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8516275Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8516416Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8516484Z graph_break [] 2025-12-04T09:28:45.8516567Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8516694Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8516767Z frames [('total', 1)] 2025-12-04T09:28:45.8516836Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8516965Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8517045Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8517107Z graph_break [] 2025-12-04T09:28:45.8517187Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8517321Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8517387Z frames [('total', 1)] 2025-12-04T09:28:45.8517463Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8517606Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8517689Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8517755Z graph_break [] 2025-12-04T09:28:45.8517833Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8517962Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8518031Z frames [('total', 1)] 2025-12-04T09:28:45.8518102Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8518229Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8518309Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8518371Z graph_break [] 2025-12-04T09:28:45.8518454Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8518583Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8518653Z frames [('total', 1)] 2025-12-04T09:28:45.8518727Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8518855Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8518929Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8518996Z graph_break [] 2025-12-04T09:28:45.8519076Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8519206Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8519278Z frames [('total', 1)] 2025-12-04T09:28:45.8519349Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8519479Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8519552Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8519615Z graph_break [] 2025-12-04T09:28:45.8519699Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8519829Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8519893Z frames [('total', 1)] 2025-12-04T09:28:45.8519969Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8520100Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8520174Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8520240Z graph_break [] 2025-12-04T09:28:45.8520318Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8520450Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8520513Z frames [('total', 1)] 2025-12-04T09:28:45.8520581Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8520711Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8520784Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8520843Z graph_break [] 2025-12-04T09:28:45.8521059Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8521188Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8521251Z frames [('total', 1)] 2025-12-04T09:28:45.8521367Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8521557Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8521636Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8521697Z graph_break [] 2025-12-04T09:28:45.8521775Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8521906Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8521968Z frames [('total', 1)] 2025-12-04T09:28:45.8522035Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8522167Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8522240Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8522297Z graph_break [] 2025-12-04T09:28:45.8522385Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8522510Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8522576Z frames [('total', 1)] 2025-12-04T09:28:45.8522646Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8522775Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8522862Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8522923Z graph_break [] 2025-12-04T09:28:45.8523001Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8523130Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8523192Z frames [('total', 1)] 2025-12-04T09:28:45.8523260Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8523390Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8523463Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8523530Z graph_break [] 2025-12-04T09:28:45.8523607Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8523735Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8523802Z frames [('total', 1)] 2025-12-04T09:28:45.8523874Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8524002Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8524086Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8524146Z graph_break [] 2025-12-04T09:28:45.8524223Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8524351Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8524414Z frames [('total', 1)] 2025-12-04T09:28:45.8524488Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8524618Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8524696Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8524761Z graph_break [] 2025-12-04T09:28:45.8524839Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8524967Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8525036Z frames [('total', 1)] 2025-12-04T09:28:45.8525106Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8525235Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8525316Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8525376Z graph_break [] 2025-12-04T09:28:45.8525451Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8525586Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8525659Z frames [('total', 1)] 2025-12-04T09:28:45.8525735Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8525863Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8525935Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8526000Z graph_break [] 2025-12-04T09:28:45.8526079Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8526252Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8526321Z frames [('total', 1)] 2025-12-04T09:28:45.8526391Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8526555Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8526702Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8526764Z graph_break [] 2025-12-04T09:28:45.8526847Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8526971Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8527035Z frames [('total', 1)] 2025-12-04T09:28:45.8527108Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8527233Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8527306Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8527386Z graph_break [] 2025-12-04T09:28:45.8527466Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8527596Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8527665Z frames [('total', 1)] 2025-12-04T09:28:45.8527736Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8527875Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8527947Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8528014Z graph_break [] 2025-12-04T09:28:45.8528096Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8528222Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8528289Z frames [('total', 1)] 2025-12-04T09:28:45.8528366Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8528499Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8528576Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8528641Z graph_break [] 2025-12-04T09:28:45.8528719Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8528861Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8528924Z frames [('total', 1)] 2025-12-04T09:28:45.8528993Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8529132Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8529213Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8529275Z graph_break [] 2025-12-04T09:28:45.8529361Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8529493Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8529554Z frames [('total', 1)] 2025-12-04T09:28:45.8529628Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8529756Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8529833Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8529890Z graph_break [] 2025-12-04T09:28:45.8529969Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8530101Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8530166Z frames [('total', 1)] 2025-12-04T09:28:45.8530232Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8530372Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8530446Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8530507Z graph_break [] 2025-12-04T09:28:45.8530590Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8530714Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8530781Z frames [('total', 1)] 2025-12-04T09:28:45.8530848Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8530978Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8531057Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8531115Z graph_break [] 2025-12-04T09:28:45.8531193Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8531323Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8531464Z frames [('total', 1)] 2025-12-04T09:28:45.8531531Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8531664Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8531773Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8531838Z graph_break [] 2025-12-04T09:28:45.8531980Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8532104Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8532172Z frames [('total', 1)] 2025-12-04T09:28:45.8532242Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8532368Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8532448Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8532508Z graph_break [] 2025-12-04T09:28:45.8532584Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8532718Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8532784Z frames [('total', 1)] 2025-12-04T09:28:45.8532861Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8532993Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8533065Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8533237Z graph_break [] 2025-12-04T09:28:45.8533322Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8533451Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8533521Z frames [('total', 1)] 2025-12-04T09:28:45.8533588Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8533715Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8533792Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8533851Z graph_break [] 2025-12-04T09:28:45.8533929Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8534058Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8534121Z frames [('total', 1)] 2025-12-04T09:28:45.8534194Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8534324Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8534403Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8534467Z graph_break [] 2025-12-04T09:28:45.8534547Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8534671Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8534744Z frames [('total', 1)] 2025-12-04T09:28:45.8534811Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8534947Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8535028Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8535086Z graph_break [] 2025-12-04T09:28:45.8535168Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8535293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8535356Z frames [('total', 1)] 2025-12-04T09:28:45.8535431Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8535558Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8535631Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8535697Z graph_break [] 2025-12-04T09:28:45.8535778Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8535904Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8535974Z frames [('total', 1)] 2025-12-04T09:28:45.8536042Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8536176Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8536248Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8536311Z graph_break [] 2025-12-04T09:28:45.8536399Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8536526Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8536642Z frames [('total', 1)] 2025-12-04T09:28:45.8536718Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8536853Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8536960Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8537027Z graph_break [] 2025-12-04T09:28:45.8537189Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8537324Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8537387Z frames [('total', 1)] 2025-12-04T09:28:45.8537455Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8537590Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8537663Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8537722Z graph_break [] 2025-12-04T09:28:45.8537808Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8537938Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8538002Z frames [('total', 1)] 2025-12-04T09:28:45.8538075Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8538199Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8538286Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8538346Z graph_break [] 2025-12-04T09:28:45.8538423Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8538561Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8538633Z frames [('total', 1)] 2025-12-04T09:28:45.8538702Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8538836Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8538906Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8538965Z graph_break [] 2025-12-04T09:28:45.8539047Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8539212Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8539291Z Traceback (most recent call last): 2025-12-04T09:28:45.8539614Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8539670Z f(a, b) 2025-12-04T09:28:45.8539970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8540129Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8540433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8540560Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8540859Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8540953Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8541274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8541478Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8541788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8541868Z graph.run(*example_inputs) 2025-12-04T09:28:45.8542114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8542183Z return super().run(*args) 2025-12-04T09:28:45.8542414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8542501Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8542753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8542823Z result = super().run_node(n) 2025-12-04T09:28:45.8543133Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8543235Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8543546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8543760Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8544026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8544167Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8544423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8544496Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8544764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8544845Z return autotune_select_algorithm( 2025-12-04T09:28:45.8545203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8545280Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8545569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8545737Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8546496Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8546572Z target: aten.mm.default 2025-12-04T09:28:45.8546634Z args[0]: TensorBox( 2025-12-04T09:28:45.8546698Z ReinterpretView( 2025-12-04T09:28:45.8546766Z StorageBox( 2025-12-04T09:28:45.8547011Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8547069Z ), 2025-12-04T09:28:45.8547224Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8547301Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8547371Z stack_traces = {, 2025-12-04T09:28:45.8547598Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8547674Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8547735Z , 2025-12-04T09:28:45.8547788Z } 2025-12-04T09:28:45.8547842Z ) 2025-12-04T09:28:45.8547905Z ) 2025-12-04T09:28:45.8547980Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8548220Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8548284Z )) 2025-12-04T09:28:45.8548287Z 2025-12-04T09:28:45.8548714Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8548719Z 2025-12-04T09:28:45.8548722Z 2025-12-04T09:28:45.8548882Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8549131Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8549135Z 2025-12-04T09:28:45.8549305Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8549446Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8549510Z frames [('total', 1)] 2025-12-04T09:28:45.8549588Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8549663Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8549801Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8549915Z graph_break [] 2025-12-04T09:28:45.8550001Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8550129Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8550233Z frames [('total', 1)] 2025-12-04T09:28:45.8550365Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8550498Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8550579Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8550637Z graph_break [] 2025-12-04T09:28:45.8550723Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8550854Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8550915Z frames [('total', 1)] 2025-12-04T09:28:45.8550988Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8551114Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8551186Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8551254Z graph_break [] 2025-12-04T09:28:45.8551332Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8551458Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8551529Z frames [('total', 1)] 2025-12-04T09:28:45.8551599Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8551734Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8551805Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8551865Z graph_break [] 2025-12-04T09:28:45.8551950Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8552076Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8552138Z frames [('total', 1)] 2025-12-04T09:28:45.8552222Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8552349Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8552420Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8552487Z graph_break [] 2025-12-04T09:28:45.8552561Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8552691Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8552755Z frames [('total', 1)] 2025-12-04T09:28:45.8552822Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8552956Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8553030Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8553087Z graph_break [] 2025-12-04T09:28:45.8553173Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8553303Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8553365Z frames [('total', 1)] 2025-12-04T09:28:45.8553449Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8553575Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8553654Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8553714Z graph_break [] 2025-12-04T09:28:45.8553788Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8553922Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8553986Z frames [('total', 1)] 2025-12-04T09:28:45.8554054Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8554192Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8554262Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8554320Z graph_break [] 2025-12-04T09:28:45.8554402Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8554530Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8554599Z frames [('total', 1)] 2025-12-04T09:28:45.8554667Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8554791Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8554869Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8554978Z graph_break [] 2025-12-04T09:28:45.8555057Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8555192Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8555288Z frames [('total', 1)] 2025-12-04T09:28:45.8555357Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8555552Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8555627Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8555695Z graph_break [] 2025-12-04T09:28:45.8555773Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8555902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8555971Z frames [('total', 1)] 2025-12-04T09:28:45.8556039Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8556166Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8556247Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8556307Z graph_break [] 2025-12-04T09:28:45.8556386Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8556517Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8556579Z frames [('total', 1)] 2025-12-04T09:28:45.8556651Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8556790Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8556862Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8556929Z graph_break [] 2025-12-04T09:28:45.8557015Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8557146Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8557214Z frames [('total', 1)] 2025-12-04T09:28:45.8557283Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8557416Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8557495Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8557554Z graph_break [] 2025-12-04T09:28:45.8557633Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8557767Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8557839Z frames [('total', 1)] 2025-12-04T09:28:45.8557919Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8558050Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8558123Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8558190Z graph_break [] 2025-12-04T09:28:45.8558266Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8558391Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8558464Z frames [('total', 1)] 2025-12-04T09:28:45.8558537Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8558665Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8558744Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8558802Z graph_break [] 2025-12-04T09:28:45.8558884Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8559009Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8559072Z frames [('total', 1)] 2025-12-04T09:28:45.8559157Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8559290Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8559372Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8559438Z graph_break [] 2025-12-04T09:28:45.8559516Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8559643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8559711Z frames [('total', 1)] 2025-12-04T09:28:45.8559778Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8559908Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8559982Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8560041Z graph_break [] 2025-12-04T09:28:45.8560124Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8560294Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8560355Z frames [('total', 1)] 2025-12-04T09:28:45.8560428Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8560588Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8560753Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8560819Z graph_break [] 2025-12-04T09:28:45.8560897Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8561026Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8561086Z frames [('total', 1)] 2025-12-04T09:28:45.8561157Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8561286Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8561358Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8561417Z graph_break [] 2025-12-04T09:28:45.8561498Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8561623Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8561686Z frames [('total', 1)] 2025-12-04T09:28:45.8561760Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8561888Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8561970Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8562030Z graph_break [] 2025-12-04T09:28:45.8562106Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8562236Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8562297Z frames [('total', 1)] 2025-12-04T09:28:45.8562366Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8562497Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8562568Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8562625Z graph_break [] 2025-12-04T09:28:45.8562708Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8562841Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8562906Z frames [('total', 1)] 2025-12-04T09:28:45.8562975Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8563108Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8563189Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8563259Z graph_break [] 2025-12-04T09:28:45.8563337Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8563469Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8563532Z frames [('total', 1)] 2025-12-04T09:28:45.8563601Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8563732Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8563804Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8563869Z graph_break [] 2025-12-04T09:28:45.8563946Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8564076Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8564145Z frames [('total', 1)] 2025-12-04T09:28:45.8564213Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8564340Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8564421Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8564483Z graph_break [] 2025-12-04T09:28:45.8564561Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8564696Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8564760Z frames [('total', 1)] 2025-12-04T09:28:45.8564832Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8564967Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8565055Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8565122Z graph_break [] 2025-12-04T09:28:45.8565198Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8565374Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8565443Z frames [('total', 1)] 2025-12-04T09:28:45.8565512Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8565638Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8565753Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8565878Z graph_break [] 2025-12-04T09:28:45.8565960Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8566097Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8566161Z frames [('total', 1)] 2025-12-04T09:28:45.8566236Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8566361Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8566435Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8566502Z graph_break [] 2025-12-04T09:28:45.8566579Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8566708Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8566779Z frames [('total', 1)] 2025-12-04T09:28:45.8566848Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8566974Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8567069Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8567136Z graph_break [] 2025-12-04T09:28:45.8567220Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8567350Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8567413Z frames [('total', 1)] 2025-12-04T09:28:45.8567494Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8567626Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8567701Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8567772Z graph_break [] 2025-12-04T09:28:45.8567849Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8567976Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8568047Z frames [('total', 1)] 2025-12-04T09:28:45.8568113Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8568243Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8568323Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8568381Z graph_break [] 2025-12-04T09:28:45.8568475Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8568608Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8568671Z frames [('total', 1)] 2025-12-04T09:28:45.8568744Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8568875Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8568947Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8569010Z graph_break [] 2025-12-04T09:28:45.8569085Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8569219Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8569283Z frames [('total', 1)] 2025-12-04T09:28:45.8569352Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8569487Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8569561Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8569622Z graph_break [] 2025-12-04T09:28:45.8569709Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8569837Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8569899Z frames [('total', 1)] 2025-12-04T09:28:45.8569971Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8570097Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8570181Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8570251Z graph_break [] 2025-12-04T09:28:45.8570330Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8570463Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8570576Z frames [('total', 1)] 2025-12-04T09:28:45.8570644Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8570778Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8570885Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8570943Z graph_break [] 2025-12-04T09:28:45.8571091Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8571219Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8571286Z frames [('total', 1)] 2025-12-04T09:28:45.8571354Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8571481Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8571558Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8571617Z graph_break [] 2025-12-04T09:28:45.8571694Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8571830Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8571896Z frames [('total', 1)] 2025-12-04T09:28:45.8571963Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8572102Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8572177Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8572246Z graph_break [] 2025-12-04T09:28:45.8572331Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8572467Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8572536Z frames [('total', 1)] 2025-12-04T09:28:45.8572604Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8572727Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8572808Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8572867Z graph_break [] 2025-12-04T09:28:45.8572942Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8573216Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8573299Z Traceback (most recent call last): 2025-12-04T09:28:45.8573627Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8573687Z f(a, b) 2025-12-04T09:28:45.8573982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8574143Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8574442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8574565Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8574867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8574960Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8575281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8575477Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8575784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8575868Z graph.run(*example_inputs) 2025-12-04T09:28:45.8576106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8576184Z return super().run(*args) 2025-12-04T09:28:45.8576415Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8576495Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8576755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8576827Z result = super().run_node(n) 2025-12-04T09:28:45.8577128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8577238Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8577548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8577771Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8578039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8578176Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8578441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8578512Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8578783Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8578879Z return autotune_select_algorithm( 2025-12-04T09:28:45.8579225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8579308Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8579596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8579762Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8580514Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8580584Z target: aten.mm.default 2025-12-04T09:28:45.8580654Z args[0]: TensorBox( 2025-12-04T09:28:45.8580722Z ReinterpretView( 2025-12-04T09:28:45.8580781Z StorageBox( 2025-12-04T09:28:45.8581028Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8581086Z ), 2025-12-04T09:28:45.8581233Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8581316Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8581380Z stack_traces = {, 2025-12-04T09:28:45.8581609Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8581694Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8581749Z , 2025-12-04T09:28:45.8581807Z } 2025-12-04T09:28:45.8581859Z ) 2025-12-04T09:28:45.8581911Z ) 2025-12-04T09:28:45.8581990Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8582221Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8582276Z )) 2025-12-04T09:28:45.8582279Z 2025-12-04T09:28:45.8582703Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8582708Z 2025-12-04T09:28:45.8582711Z 2025-12-04T09:28:45.8582846Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8583099Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8583102Z 2025-12-04T09:28:45.8583275Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8583408Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8583479Z frames [('total', 1)] 2025-12-04T09:28:45.8583548Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8583628Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8583808Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8583871Z graph_break [] 2025-12-04T09:28:45.8583956Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8584084Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8584202Z frames [('total', 1)] 2025-12-04T09:28:45.8584369Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8584499Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8584574Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8584642Z graph_break [] 2025-12-04T09:28:45.8584719Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8584851Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8584913Z frames [('total', 1)] 2025-12-04T09:28:45.8584983Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8585113Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8585188Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8585248Z graph_break [] 2025-12-04T09:28:45.8585334Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8585463Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8585531Z frames [('total', 1)] 2025-12-04T09:28:45.8585609Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8585737Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8585827Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8585889Z graph_break [] 2025-12-04T09:28:45.8585967Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8586099Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8586160Z frames [('total', 1)] 2025-12-04T09:28:45.8586230Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8586363Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8586437Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8586495Z graph_break [] 2025-12-04T09:28:45.8586581Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8586706Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8586776Z frames [('total', 1)] 2025-12-04T09:28:45.8586844Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8586973Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8587051Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8587110Z graph_break [] 2025-12-04T09:28:45.8587187Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8587323Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8587386Z frames [('total', 1)] 2025-12-04T09:28:45.8587455Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8587589Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8587660Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8587720Z graph_break [] 2025-12-04T09:28:45.8587803Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8587934Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8588006Z frames [('total', 1)] 2025-12-04T09:28:45.8588075Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8588216Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8588298Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8588360Z graph_break [] 2025-12-04T09:28:45.8588435Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8588570Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8588634Z frames [('total', 1)] 2025-12-04T09:28:45.8588704Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8588838Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8588913Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8589029Z graph_break [] 2025-12-04T09:28:45.8589109Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8589238Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8589341Z frames [('total', 1)] 2025-12-04T09:28:45.8589415Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8589607Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8589689Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8589749Z graph_break [] 2025-12-04T09:28:45.8589826Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8589969Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8590033Z frames [('total', 1)] 2025-12-04T09:28:45.8590108Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8590233Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8590303Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8590369Z graph_break [] 2025-12-04T09:28:45.8590443Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8590571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8590641Z frames [('total', 1)] 2025-12-04T09:28:45.8590710Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8590839Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8590917Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8590975Z graph_break [] 2025-12-04T09:28:45.8591059Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8591185Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8591246Z frames [('total', 1)] 2025-12-04T09:28:45.8591320Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8591444Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8591516Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8591586Z graph_break [] 2025-12-04T09:28:45.8591662Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8591787Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8591857Z frames [('total', 1)] 2025-12-04T09:28:45.8591924Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8592064Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8592136Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8592197Z graph_break [] 2025-12-04T09:28:45.8592280Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8592413Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8592477Z frames [('total', 1)] 2025-12-04T09:28:45.8592551Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8592676Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8592749Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8592815Z graph_break [] 2025-12-04T09:28:45.8592895Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8593025Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8593087Z frames [('total', 1)] 2025-12-04T09:28:45.8593159Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8593293Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8593365Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8593425Z graph_break [] 2025-12-04T09:28:45.8593508Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8593634Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8593697Z frames [('total', 1)] 2025-12-04T09:28:45.8593771Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8593900Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8593977Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8594036Z graph_break [] 2025-12-04T09:28:45.8594160Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8594294Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8594357Z frames [('total', 1)] 2025-12-04T09:28:45.8594463Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8594598Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8594734Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8594796Z graph_break [] 2025-12-04T09:28:45.8594879Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8595007Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8595077Z frames [('total', 1)] 2025-12-04T09:28:45.8595144Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8595279Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8595360Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8595423Z graph_break [] 2025-12-04T09:28:45.8595505Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8595651Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8595716Z frames [('total', 1)] 2025-12-04T09:28:45.8595789Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8595930Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8596011Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8596077Z graph_break [] 2025-12-04T09:28:45.8596157Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8596288Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8596357Z frames [('total', 1)] 2025-12-04T09:28:45.8596426Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8596557Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8596637Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8596697Z graph_break [] 2025-12-04T09:28:45.8596780Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8596919Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8596982Z frames [('total', 1)] 2025-12-04T09:28:45.8597050Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8597187Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8597268Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8597335Z graph_break [] 2025-12-04T09:28:45.8597414Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8597546Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8597619Z frames [('total', 1)] 2025-12-04T09:28:45.8597690Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8597825Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8597904Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8597964Z graph_break [] 2025-12-04T09:28:45.8598045Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8598185Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8598248Z frames [('total', 1)] 2025-12-04T09:28:45.8598322Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8598452Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8598528Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8598596Z graph_break [] 2025-12-04T09:28:45.8598673Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8598798Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8598866Z frames [('total', 1)] 2025-12-04T09:28:45.8598935Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8599062Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8599143Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8599201Z graph_break [] 2025-12-04T09:28:45.8599284Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8599460Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8599524Z frames [('total', 1)] 2025-12-04T09:28:45.8599597Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8599762Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8599910Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8599977Z graph_break [] 2025-12-04T09:28:45.8600054Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8600180Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8600247Z frames [('total', 1)] 2025-12-04T09:28:45.8600314Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8600447Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8600518Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8600576Z graph_break [] 2025-12-04T09:28:45.8600657Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8600785Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8600848Z frames [('total', 1)] 2025-12-04T09:28:45.8600925Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8601054Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8601133Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8601202Z graph_break [] 2025-12-04T09:28:45.8601280Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8601416Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8601477Z frames [('total', 1)] 2025-12-04T09:28:45.8601545Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8601678Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8601751Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8601810Z graph_break [] 2025-12-04T09:28:45.8601891Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8602021Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8602098Z frames [('total', 1)] 2025-12-04T09:28:45.8602174Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8602302Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8602392Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8602456Z graph_break [] 2025-12-04T09:28:45.8602533Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8602665Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8602727Z frames [('total', 1)] 2025-12-04T09:28:45.8602795Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8602928Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8603000Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8603060Z graph_break [] 2025-12-04T09:28:45.8603145Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8603271Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8603344Z frames [('total', 1)] 2025-12-04T09:28:45.8603411Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8603538Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8603619Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8603679Z graph_break [] 2025-12-04T09:28:45.8603760Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8603895Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8603956Z frames [('total', 1)] 2025-12-04T09:28:45.8604025Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8604158Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8604232Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8604299Z graph_break [] 2025-12-04T09:28:45.8604375Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8604889Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8605120Z frames [('total', 1)] 2025-12-04T09:28:45.8605196Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8605344Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8605497Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8605563Z graph_break [] 2025-12-04T09:28:45.8605800Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8605947Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8606011Z frames [('total', 1)] 2025-12-04T09:28:45.8606083Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8606228Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8606304Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8609327Z graph_break [] 2025-12-04T09:28:45.8609472Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8609623Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8609699Z frames [('total', 1)] 2025-12-04T09:28:45.8609782Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8609927Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8610019Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8610084Z graph_break [] 2025-12-04T09:28:45.8610177Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8610328Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8610398Z frames [('total', 1)] 2025-12-04T09:28:45.8610471Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8610616Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8610692Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8610752Z graph_break [] 2025-12-04T09:28:45.8610871Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8611001Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8611066Z frames [('total', 1)] 2025-12-04T09:28:45.8611143Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8611272Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8611355Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8611414Z graph_break [] 2025-12-04T09:28:45.8611502Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8611676Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8611753Z Traceback (most recent call last): 2025-12-04T09:28:45.8612079Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8612143Z f(a, b) 2025-12-04T09:28:45.8612446Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8612605Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8612919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8613040Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8613476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8613572Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8613883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8614084Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8614386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8614465Z graph.run(*example_inputs) 2025-12-04T09:28:45.8614700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8614837Z return super().run(*args) 2025-12-04T09:28:45.8615084Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8615204Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8615504Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8615582Z result = super().run_node(n) 2025-12-04T09:28:45.8615831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8615942Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8616299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8616451Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8616726Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8616871Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8617153Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8617228Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8617503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8617589Z return autotune_select_algorithm( 2025-12-04T09:28:45.8617931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8618005Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8618302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8618466Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8619240Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8619314Z target: aten.mm.default 2025-12-04T09:28:45.8619378Z args[0]: TensorBox( 2025-12-04T09:28:45.8619450Z ReinterpretView( 2025-12-04T09:28:45.8619508Z StorageBox( 2025-12-04T09:28:45.8619752Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8619815Z ), 2025-12-04T09:28:45.8619970Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8620058Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8620121Z stack_traces = {, 2025-12-04T09:28:45.8620356Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8620441Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8620499Z , 2025-12-04T09:28:45.8620554Z } 2025-12-04T09:28:45.8620615Z ) 2025-12-04T09:28:45.8620669Z ) 2025-12-04T09:28:45.8620746Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8620986Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8621040Z )) 2025-12-04T09:28:45.8621045Z 2025-12-04T09:28:45.8626069Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8626079Z 2025-12-04T09:28:45.8626088Z 2025-12-04T09:28:45.8626262Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8626522Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8626609Z 2025-12-04T09:28:45.8626784Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8626965Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8627037Z frames [('total', 1)] 2025-12-04T09:28:45.8627152Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8627232Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8627378Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8627439Z graph_break [] 2025-12-04T09:28:45.8627523Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8627664Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8627784Z frames [('total', 1)] 2025-12-04T09:28:45.8627861Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8627992Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8628069Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8628133Z graph_break [] 2025-12-04T09:28:45.8628212Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8628353Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8628423Z frames [('total', 1)] 2025-12-04T09:28:45.8628494Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8628619Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8628693Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8628752Z graph_break [] 2025-12-04T09:28:45.8628827Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8628956Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8629017Z frames [('total', 1)] 2025-12-04T09:28:45.8629092Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8629216Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8629291Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8629354Z graph_break [] 2025-12-04T09:28:45.8629432Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8629557Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8629628Z frames [('total', 1)] 2025-12-04T09:28:45.8629698Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8629824Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8629900Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8629957Z graph_break [] 2025-12-04T09:28:45.8630039Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8630163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8630225Z frames [('total', 1)] 2025-12-04T09:28:45.8630299Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8630423Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8630497Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8630561Z graph_break [] 2025-12-04T09:28:45.8630635Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8630757Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8630827Z frames [('total', 1)] 2025-12-04T09:28:45.8630897Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8631026Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8631096Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8631157Z graph_break [] 2025-12-04T09:28:45.8631242Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8631371Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8631434Z frames [('total', 1)] 2025-12-04T09:28:45.8631512Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8631645Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8631722Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8631838Z graph_break [] 2025-12-04T09:28:45.8631918Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8632056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8632157Z frames [('total', 1)] 2025-12-04T09:28:45.8632226Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8632421Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8632497Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8632557Z graph_break [] 2025-12-04T09:28:45.8632639Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8632763Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8632824Z frames [('total', 1)] 2025-12-04T09:28:45.8632937Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8633066Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8633152Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8633216Z graph_break [] 2025-12-04T09:28:45.8633293Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8633422Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8633485Z frames [('total', 1)] 2025-12-04T09:28:45.8633550Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8633680Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8633750Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8633808Z graph_break [] 2025-12-04T09:28:45.8633889Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8634013Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8634080Z frames [('total', 1)] 2025-12-04T09:28:45.8634148Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8634276Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8634350Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8634411Z graph_break [] 2025-12-04T09:28:45.8634487Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8634615Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8634682Z frames [('total', 1)] 2025-12-04T09:28:45.8634749Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8634882Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8634953Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8635015Z graph_break [] 2025-12-04T09:28:45.8635092Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8635214Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8635280Z frames [('total', 1)] 2025-12-04T09:28:45.8635347Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8635474Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8635549Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8635609Z graph_break [] 2025-12-04T09:28:45.8635683Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8635814Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8635880Z frames [('total', 1)] 2025-12-04T09:28:45.8635957Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8636088Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8636161Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8636229Z graph_break [] 2025-12-04T09:28:45.8636315Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8636446Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8636517Z frames [('total', 1)] 2025-12-04T09:28:45.8636585Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8636715Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8636796Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8636858Z graph_break [] 2025-12-04T09:28:45.8636982Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8637121Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8637183Z frames [('total', 1)] 2025-12-04T09:28:45.8637294Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8637455Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8637531Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8637597Z graph_break [] 2025-12-04T09:28:45.8637676Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8637800Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8637871Z frames [('total', 1)] 2025-12-04T09:28:45.8637939Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8638116Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8638200Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8638260Z graph_break [] 2025-12-04T09:28:45.8638346Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8638472Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8638535Z frames [('total', 1)] 2025-12-04T09:28:45.8638612Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8638737Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8638809Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8638874Z graph_break [] 2025-12-04T09:28:45.8638953Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8639078Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8639147Z frames [('total', 1)] 2025-12-04T09:28:45.8639216Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8639345Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8639431Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8639495Z graph_break [] 2025-12-04T09:28:45.8639579Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8639704Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8639769Z frames [('total', 1)] 2025-12-04T09:28:45.8639845Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8639970Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8640045Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8640111Z graph_break [] 2025-12-04T09:28:45.8640190Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8640314Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8640389Z frames [('total', 1)] 2025-12-04T09:28:45.8640458Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8640591Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8640663Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8640721Z graph_break [] 2025-12-04T09:28:45.8640803Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8640932Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8640994Z frames [('total', 1)] 2025-12-04T09:28:45.8641070Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8641193Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8641267Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8641333Z graph_break [] 2025-12-04T09:28:45.8641410Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8641541Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8641603Z frames [('total', 1)] 2025-12-04T09:28:45.8641671Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8641801Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8641878Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8641939Z graph_break [] 2025-12-04T09:28:45.8642021Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8642191Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8642254Z frames [('total', 1)] 2025-12-04T09:28:45.8642330Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8642501Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8642612Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8642674Z graph_break [] 2025-12-04T09:28:45.8642752Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8642882Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8642945Z frames [('total', 1)] 2025-12-04T09:28:45.8643014Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8643148Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8643263Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8643324Z graph_break [] 2025-12-04T09:28:45.8643408Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8643538Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8643601Z frames [('total', 1)] 2025-12-04T09:28:45.8643674Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8643803Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8643884Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8643941Z graph_break [] 2025-12-04T09:28:45.8644021Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8644150Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8644211Z frames [('total', 1)] 2025-12-04T09:28:45.8644279Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8644407Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8644481Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8644540Z graph_break [] 2025-12-04T09:28:45.8644622Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8644749Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8644818Z frames [('total', 1)] 2025-12-04T09:28:45.8644885Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8645013Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8645093Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8645153Z graph_break [] 2025-12-04T09:28:45.8645232Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8645362Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8645424Z frames [('total', 1)] 2025-12-04T09:28:45.8645493Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8645621Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8645696Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8645772Z graph_break [] 2025-12-04T09:28:45.8645848Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8645978Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8646045Z frames [('total', 1)] 2025-12-04T09:28:45.8646112Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8646234Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8646314Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8646373Z graph_break [] 2025-12-04T09:28:45.8646450Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8646580Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8646642Z frames [('total', 1)] 2025-12-04T09:28:45.8646714Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8646838Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8646910Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8646976Z graph_break [] 2025-12-04T09:28:45.8647053Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8647177Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8647296Z frames [('total', 1)] 2025-12-04T09:28:45.8647366Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8647490Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8647603Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8647698Z graph_break [] 2025-12-04T09:28:45.8647782Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8647910Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8647974Z frames [('total', 1)] 2025-12-04T09:28:45.8648047Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8648172Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8648253Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8648418Z graph_break [] 2025-12-04T09:28:45.8648498Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8648623Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8648694Z frames [('total', 1)] 2025-12-04T09:28:45.8648761Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8648890Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8648965Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8649024Z graph_break [] 2025-12-04T09:28:45.8649111Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8649235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8649298Z frames [('total', 1)] 2025-12-04T09:28:45.8649372Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8649495Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8649567Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8649635Z graph_break [] 2025-12-04T09:28:45.8649710Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8649841Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8649906Z frames [('total', 1)] 2025-12-04T09:28:45.8649975Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8650106Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8650181Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8650240Z graph_break [] 2025-12-04T09:28:45.8650325Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8650452Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8650515Z frames [('total', 1)] 2025-12-04T09:28:45.8650593Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8650717Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8650798Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8650858Z graph_break [] 2025-12-04T09:28:45.8650942Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8651083Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8651151Z frames [('total', 1)] 2025-12-04T09:28:45.8651221Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8651360Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8651436Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8651495Z graph_break [] 2025-12-04T09:28:45.8651583Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8651753Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8651839Z Traceback (most recent call last): 2025-12-04T09:28:45.8652169Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8652227Z f(a, b) 2025-12-04T09:28:45.8652532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8652692Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8653044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8653299Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8653692Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8653793Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8654108Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8654307Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8654650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8654726Z graph.run(*example_inputs) 2025-12-04T09:28:45.8654961Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8655041Z return super().run(*args) 2025-12-04T09:28:45.8655280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8655368Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8655622Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8655693Z result = super().run_node(n) 2025-12-04T09:28:45.8655941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8656046Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8656320Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8656471Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8656736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8656878Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8657144Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8657219Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8657480Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8657559Z return autotune_select_algorithm( 2025-12-04T09:28:45.8657911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8657986Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8658275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8658443Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8659191Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8659269Z target: aten.mm.default 2025-12-04T09:28:45.8659334Z args[0]: TensorBox( 2025-12-04T09:28:45.8659400Z ReinterpretView( 2025-12-04T09:28:45.8659465Z StorageBox( 2025-12-04T09:28:45.8659707Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8659763Z ), 2025-12-04T09:28:45.8659924Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8659998Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8660062Z stack_traces = {, 2025-12-04T09:28:45.8660336Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8660415Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8660513Z , 2025-12-04T09:28:45.8660581Z } 2025-12-04T09:28:45.8660637Z ) 2025-12-04T09:28:45.8660698Z ) 2025-12-04T09:28:45.8660806Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8661043Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8661104Z )) 2025-12-04T09:28:45.8661108Z 2025-12-04T09:28:45.8661529Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8661533Z 2025-12-04T09:28:45.8661574Z 2025-12-04T09:28:45.8661716Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8661961Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8661971Z 2025-12-04T09:28:45.8662139Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8662281Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8662348Z frames [('total', 1)] 2025-12-04T09:28:45.8662426Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8662501Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8662639Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8662706Z graph_break [] 2025-12-04T09:28:45.8662791Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8662921Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8662994Z frames [('total', 1)] 2025-12-04T09:28:45.8663065Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8663195Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8663276Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8663337Z graph_break [] 2025-12-04T09:28:45.8663421Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8663551Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8663614Z frames [('total', 1)] 2025-12-04T09:28:45.8663694Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8663821Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8663896Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8663969Z graph_break [] 2025-12-04T09:28:45.8664047Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8664174Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8664246Z frames [('total', 1)] 2025-12-04T09:28:45.8664328Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8664466Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8664541Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8664600Z graph_break [] 2025-12-04T09:28:45.8664683Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8664808Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8664873Z frames [('total', 1)] 2025-12-04T09:28:45.8664952Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8665079Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8665151Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8665220Z graph_break [] 2025-12-04T09:28:45.8665298Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8665430Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8665494Z frames [('total', 1)] 2025-12-04T09:28:45.8665565Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8665694Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8665822Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8665881Z graph_break [] 2025-12-04T09:28:45.8665965Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8666090Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8666197Z frames [('total', 1)] 2025-12-04T09:28:45.8666308Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8666435Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8666515Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8666575Z graph_break [] 2025-12-04T09:28:45.8666652Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8666786Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8666851Z frames [('total', 1)] 2025-12-04T09:28:45.8666958Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8667102Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8667183Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8667248Z graph_break [] 2025-12-04T09:28:45.8667327Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8667453Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8667524Z frames [('total', 1)] 2025-12-04T09:28:45.8667593Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8667722Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8667801Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8667860Z graph_break [] 2025-12-04T09:28:45.8667938Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8668070Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8668131Z frames [('total', 1)] 2025-12-04T09:28:45.8668212Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8668346Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8668418Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8668486Z graph_break [] 2025-12-04T09:28:45.8668561Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8668685Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8668756Z frames [('total', 1)] 2025-12-04T09:28:45.8668824Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8668950Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8669029Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8669087Z graph_break [] 2025-12-04T09:28:45.8669169Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8669293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8669358Z frames [('total', 1)] 2025-12-04T09:28:45.8669435Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8669567Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8669645Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8669714Z graph_break [] 2025-12-04T09:28:45.8669792Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8669923Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8669993Z frames [('total', 1)] 2025-12-04T09:28:45.8670062Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8670194Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8670283Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8670344Z graph_break [] 2025-12-04T09:28:45.8670428Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8670555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8670618Z frames [('total', 1)] 2025-12-04T09:28:45.8670694Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8670821Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8670894Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8671005Z graph_break [] 2025-12-04T09:28:45.8671084Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8671211Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8671317Z frames [('total', 1)] 2025-12-04T09:28:45.8671386Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8671551Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8671625Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8671684Z graph_break [] 2025-12-04T09:28:45.8671768Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8671894Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8671957Z frames [('total', 1)] 2025-12-04T09:28:45.8672031Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8672192Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8672267Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8672334Z graph_break [] 2025-12-04T09:28:45.8672411Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8672545Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8672610Z frames [('total', 1)] 2025-12-04T09:28:45.8672681Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8672814Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8672887Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8672948Z graph_break [] 2025-12-04T09:28:45.8673032Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8673156Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8673219Z frames [('total', 1)] 2025-12-04T09:28:45.8673296Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8673424Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8673518Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8673580Z graph_break [] 2025-12-04T09:28:45.8673660Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8673794Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8673856Z frames [('total', 1)] 2025-12-04T09:28:45.8673926Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8674061Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8674133Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8674193Z graph_break [] 2025-12-04T09:28:45.8674277Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8674402Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8674470Z frames [('total', 1)] 2025-12-04T09:28:45.8674539Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8674664Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8674742Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8674801Z graph_break [] 2025-12-04T09:28:45.8674881Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8675012Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8675076Z frames [('total', 1)] 2025-12-04T09:28:45.8675145Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8675281Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8675355Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8675422Z graph_break [] 2025-12-04T09:28:45.8675500Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8675625Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8675693Z frames [('total', 1)] 2025-12-04T09:28:45.8675762Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8675889Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8675966Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8676027Z graph_break [] 2025-12-04T09:28:45.8676176Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8676306Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8676369Z frames [('total', 1)] 2025-12-04T09:28:45.8676480Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8676605Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8676711Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8676777Z graph_break [] 2025-12-04T09:28:45.8676853Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8676979Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8677047Z frames [('total', 1)] 2025-12-04T09:28:45.8677117Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8677286Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8677366Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8677425Z graph_break [] 2025-12-04T09:28:45.8677507Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8677633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8677695Z frames [('total', 1)] 2025-12-04T09:28:45.8677769Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8677898Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8677973Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8678042Z graph_break [] 2025-12-04T09:28:45.8678119Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8678245Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8678312Z frames [('total', 1)] 2025-12-04T09:28:45.8678378Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8678508Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8678581Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8678641Z graph_break [] 2025-12-04T09:28:45.8678721Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8678847Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8678908Z frames [('total', 1)] 2025-12-04T09:28:45.8678980Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8679105Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8679178Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8679243Z graph_break [] 2025-12-04T09:28:45.8679320Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8679446Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8679514Z frames [('total', 1)] 2025-12-04T09:28:45.8679594Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8679726Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8679801Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8679861Z graph_break [] 2025-12-04T09:28:45.8679946Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8680073Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8680136Z frames [('total', 1)] 2025-12-04T09:28:45.8680210Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8680337Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8680411Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8680476Z graph_break [] 2025-12-04T09:28:45.8680553Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8680683Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8680745Z frames [('total', 1)] 2025-12-04T09:28:45.8680814Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8680944Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8681021Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8681081Z graph_break [] 2025-12-04T09:28:45.8681167Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8681339Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8681403Z frames [('total', 1)] 2025-12-04T09:28:45.8681480Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8681639Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8681717Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8681810Z graph_break [] 2025-12-04T09:28:45.8681899Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8682033Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8682096Z frames [('total', 1)] 2025-12-04T09:28:45.8682163Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8682292Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8682402Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8682463Z graph_break [] 2025-12-04T09:28:45.8682546Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8682670Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8682738Z frames [('total', 1)] 2025-12-04T09:28:45.8682808Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8682933Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8683017Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8683079Z graph_break [] 2025-12-04T09:28:45.8683156Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8683288Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8683356Z frames [('total', 1)] 2025-12-04T09:28:45.8683425Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8683553Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8683627Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8683693Z graph_break [] 2025-12-04T09:28:45.8683773Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8683898Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8683970Z frames [('total', 1)] 2025-12-04T09:28:45.8684039Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8684163Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8684249Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8684312Z graph_break [] 2025-12-04T09:28:45.8684391Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8684523Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8684586Z frames [('total', 1)] 2025-12-04T09:28:45.8684664Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8684789Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8684861Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8684939Z graph_break [] 2025-12-04T09:28:45.8685017Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8685142Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8685211Z frames [('total', 1)] 2025-12-04T09:28:45.8685278Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8685403Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8685484Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8685544Z graph_break [] 2025-12-04T09:28:45.8685628Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8685750Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8685812Z frames [('total', 1)] 2025-12-04T09:28:45.8685884Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8686007Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8686077Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8686145Z graph_break [] 2025-12-04T09:28:45.8686220Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8686343Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8686457Z frames [('total', 1)] 2025-12-04T09:28:45.8686525Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8686655Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8686762Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8686821Z graph_break [] 2025-12-04T09:28:45.8686942Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8687067Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8687132Z frames [('total', 1)] 2025-12-04T09:28:45.8687204Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8687328Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8687400Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8687463Z graph_break [] 2025-12-04T09:28:45.8687578Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8687756Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8687843Z Traceback (most recent call last): 2025-12-04T09:28:45.8688166Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8688234Z f(a, b) 2025-12-04T09:28:45.8688529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8688685Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8688991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8689113Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8689413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8689506Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8689821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8690020Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8690323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8690398Z graph.run(*example_inputs) 2025-12-04T09:28:45.8690636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8690704Z return super().run(*args) 2025-12-04T09:28:45.8690940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8691021Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8691275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8691355Z result = super().run_node(n) 2025-12-04T09:28:45.8691601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8691703Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8691980Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8692129Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8692397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8692530Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8692788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8692866Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8693203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8693340Z return autotune_select_algorithm( 2025-12-04T09:28:45.8693678Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8693788Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8694112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8694276Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8695084Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8695158Z target: aten.mm.default 2025-12-04T09:28:45.8695222Z args[0]: TensorBox( 2025-12-04T09:28:45.8695294Z ReinterpretView( 2025-12-04T09:28:45.8695354Z StorageBox( 2025-12-04T09:28:45.8695607Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8695673Z ), 2025-12-04T09:28:45.8695826Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8695902Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8695976Z stack_traces = {, 2025-12-04T09:28:45.8696202Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8696287Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8696346Z , 2025-12-04T09:28:45.8696402Z } 2025-12-04T09:28:45.8696466Z ) 2025-12-04T09:28:45.8696523Z ) 2025-12-04T09:28:45.8696598Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8696839Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8696898Z )) 2025-12-04T09:28:45.8696903Z 2025-12-04T09:28:45.8697327Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8697344Z 2025-12-04T09:28:45.8697347Z 2025-12-04T09:28:45.8697485Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8697731Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8697735Z 2025-12-04T09:28:45.8697905Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8698039Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8698111Z frames [('total', 1)] 2025-12-04T09:28:45.8698184Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8698262Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8698406Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8698469Z graph_break [] 2025-12-04T09:28:45.8698560Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8698696Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8698757Z frames [('total', 1)] 2025-12-04T09:28:45.8698828Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8698960Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8699033Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8699097Z graph_break [] 2025-12-04T09:28:45.8699176Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8699304Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8699375Z frames [('total', 1)] 2025-12-04T09:28:45.8699445Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8699575Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8699702Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8699764Z graph_break [] 2025-12-04T09:28:45.8699845Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8700016Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8700078Z frames [('total', 1)] 2025-12-04T09:28:45.8700186Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8700317Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8700390Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8700458Z graph_break [] 2025-12-04T09:28:45.8700536Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8700661Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8700782Z frames [('total', 1)] 2025-12-04T09:28:45.8700854Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8700981Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8701061Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8701121Z graph_break [] 2025-12-04T09:28:45.8701198Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8701331Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8701394Z frames [('total', 1)] 2025-12-04T09:28:45.8701471Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8701596Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8701668Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8701733Z graph_break [] 2025-12-04T09:28:45.8701810Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8701936Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8702006Z frames [('total', 1)] 2025-12-04T09:28:45.8702076Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8702203Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8702286Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8702357Z graph_break [] 2025-12-04T09:28:45.8702442Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8702569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8702634Z frames [('total', 1)] 2025-12-04T09:28:45.8702711Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8702838Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8702912Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8702979Z graph_break [] 2025-12-04T09:28:45.8703054Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8703180Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8703252Z frames [('total', 1)] 2025-12-04T09:28:45.8703323Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8703455Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8703529Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8703589Z graph_break [] 2025-12-04T09:28:45.8703673Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8703799Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8703864Z frames [('total', 1)] 2025-12-04T09:28:45.8703940Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8704068Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8704141Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8704211Z graph_break [] 2025-12-04T09:28:45.8704288Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8704421Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8704486Z frames [('total', 1)] 2025-12-04T09:28:45.8704840Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8704976Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8705050Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8705198Z graph_break [] 2025-12-04T09:28:45.8705287Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8705414Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8705577Z frames [('total', 1)] 2025-12-04T09:28:45.8705657Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8705832Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8705920Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8705981Z graph_break [] 2025-12-04T09:28:45.8706058Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8706190Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8706253Z frames [('total', 1)] 2025-12-04T09:28:45.8706322Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8706520Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8706596Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8706658Z graph_break [] 2025-12-04T09:28:45.8706747Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8706876Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8706953Z frames [('total', 1)] 2025-12-04T09:28:45.8707026Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8707160Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8707241Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8707302Z graph_break [] 2025-12-04T09:28:45.8707379Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8707509Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8707581Z frames [('total', 1)] 2025-12-04T09:28:45.8707651Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8707787Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8707858Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8707927Z graph_break [] 2025-12-04T09:28:45.8708003Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8708125Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8708197Z frames [('total', 1)] 2025-12-04T09:28:45.8708264Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8708391Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8708468Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8708528Z graph_break [] 2025-12-04T09:28:45.8708604Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8708736Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8708799Z frames [('total', 1)] 2025-12-04T09:28:45.8708873Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8709000Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8709074Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8709142Z graph_break [] 2025-12-04T09:28:45.8709218Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8709343Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8709413Z frames [('total', 1)] 2025-12-04T09:28:45.8709480Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8709619Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8709700Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8709761Z graph_break [] 2025-12-04T09:28:45.8709837Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8709966Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8710029Z frames [('total', 1)] 2025-12-04T09:28:45.8710102Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8710229Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8710304Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8710367Z graph_break [] 2025-12-04T09:28:45.8710490Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8710614Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8710685Z frames [('total', 1)] 2025-12-04T09:28:45.8710791Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8710948Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8711028Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8711087Z graph_break [] 2025-12-04T09:28:45.8711171Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8711295Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8711357Z frames [('total', 1)] 2025-12-04T09:28:45.8711429Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8711601Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8711676Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8711742Z graph_break [] 2025-12-04T09:28:45.8711826Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8711952Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8712036Z frames [('total', 1)] 2025-12-04T09:28:45.8712109Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8712243Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8712319Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8712379Z graph_break [] 2025-12-04T09:28:45.8712464Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8712590Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8712654Z frames [('total', 1)] 2025-12-04T09:28:45.8712731Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8712857Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8712931Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8712998Z graph_break [] 2025-12-04T09:28:45.8713077Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8713210Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8713275Z frames [('total', 1)] 2025-12-04T09:28:45.8713346Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8713478Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8713553Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8713615Z graph_break [] 2025-12-04T09:28:45.8713698Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8713824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8713888Z frames [('total', 1)] 2025-12-04T09:28:45.8713964Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8714094Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8714174Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8714235Z graph_break [] 2025-12-04T09:28:45.8714313Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8714447Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8714512Z frames [('total', 1)] 2025-12-04T09:28:45.8714581Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8714717Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8714802Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8714865Z graph_break [] 2025-12-04T09:28:45.8714952Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8715078Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8715151Z frames [('total', 1)] 2025-12-04T09:28:45.8715220Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8715347Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8715430Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8715490Z graph_break [] 2025-12-04T09:28:45.8715567Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8715746Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8715810Z frames [('total', 1)] 2025-12-04T09:28:45.8715878Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8716051Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8716174Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8716241Z graph_break [] 2025-12-04T09:28:45.8716317Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8716442Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8716510Z frames [('total', 1)] 2025-12-04T09:28:45.8716585Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8716732Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8716869Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8716946Z graph_break [] 2025-12-04T09:28:45.8717037Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8717194Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8717267Z frames [('total', 1)] 2025-12-04T09:28:45.8717353Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8717503Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8717588Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8717666Z graph_break [] 2025-12-04T09:28:45.8717756Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8717912Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8717989Z frames [('total', 1)] 2025-12-04T09:28:45.8718069Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8718216Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8718310Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8718382Z graph_break [] 2025-12-04T09:28:45.8718471Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8718628Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8718703Z frames [('total', 1)] 2025-12-04T09:28:45.8718788Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8718937Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8719013Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8719082Z graph_break [] 2025-12-04T09:28:45.8719166Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8719295Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8719365Z frames [('total', 1)] 2025-12-04T09:28:45.8719434Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8719559Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8719642Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8719704Z graph_break [] 2025-12-04T09:28:45.8719787Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8719911Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8719984Z frames [('total', 1)] 2025-12-04T09:28:45.8720060Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8720187Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8720265Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8720332Z graph_break [] 2025-12-04T09:28:45.8720410Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8720540Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8720609Z frames [('total', 1)] 2025-12-04T09:28:45.8720677Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8720809Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8720883Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8720945Z graph_break [] 2025-12-04T09:28:45.8721029Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8721155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8721272Z frames [('total', 1)] 2025-12-04T09:28:45.8721347Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8721473Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8721583Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8721649Z graph_break [] 2025-12-04T09:28:45.8721758Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8721892Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8721957Z frames [('total', 1)] 2025-12-04T09:28:45.8722026Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8722156Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8722228Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8722327Z graph_break [] 2025-12-04T09:28:45.8722420Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8722548Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8722612Z frames [('total', 1)] 2025-12-04T09:28:45.8722688Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8722813Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8722897Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8722957Z graph_break [] 2025-12-04T09:28:45.8723039Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8723170Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8723233Z frames [('total', 1)] 2025-12-04T09:28:45.8723303Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8723439Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8723511Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8723573Z graph_break [] 2025-12-04T09:28:45.8723659Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8723784Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8723856Z frames [('total', 1)] 2025-12-04T09:28:45.8723925Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8724050Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8724135Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8724195Z graph_break [] 2025-12-04T09:28:45.8724276Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8724411Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8724472Z frames [('total', 1)] 2025-12-04T09:28:45.8724541Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8724674Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8724748Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8724813Z graph_break [] 2025-12-04T09:28:45.8724892Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8725056Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8725152Z Traceback (most recent call last): 2025-12-04T09:28:45.8725470Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8725531Z f(a, b) 2025-12-04T09:28:45.8725837Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8725991Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8726297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8726418Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8726714Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8726811Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8727121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8727365Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8727742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8727819Z graph.run(*example_inputs) 2025-12-04T09:28:45.8728060Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8728132Z return super().run(*args) 2025-12-04T09:28:45.8728364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8728449Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8728735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8728822Z result = super().run_node(n) 2025-12-04T09:28:45.8729082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8729187Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8729467Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8729614Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8729880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8730018Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8730280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8730359Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8730618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8730701Z return autotune_select_algorithm( 2025-12-04T09:28:45.8731047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8731124Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8731413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8731582Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8732339Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8732421Z target: aten.mm.default 2025-12-04T09:28:45.8732484Z args[0]: TensorBox( 2025-12-04T09:28:45.8732553Z ReinterpretView( 2025-12-04T09:28:45.8732619Z StorageBox( 2025-12-04T09:28:45.8732860Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8732924Z ), 2025-12-04T09:28:45.8733077Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8733235Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8733306Z stack_traces = {, 2025-12-04T09:28:45.8733531Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8733607Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8733673Z , 2025-12-04T09:28:45.8733728Z } 2025-12-04T09:28:45.8733784Z ) 2025-12-04T09:28:45.8733854Z ) 2025-12-04T09:28:45.8733933Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8734168Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8734279Z )) 2025-12-04T09:28:45.8734283Z 2025-12-04T09:28:45.8734705Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8734746Z 2025-12-04T09:28:45.8734750Z 2025-12-04T09:28:45.8734926Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8735171Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8735176Z 2025-12-04T09:28:45.8735345Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8735489Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8735589Z frames [('total', 1)] 2025-12-04T09:28:45.8735672Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8735748Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8735890Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8735960Z graph_break [] 2025-12-04T09:28:45.8736044Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8736178Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8736248Z frames [('total', 1)] 2025-12-04T09:28:45.8736322Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8736461Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8736536Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8736596Z graph_break [] 2025-12-04T09:28:45.8736681Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8736809Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8736877Z frames [('total', 1)] 2025-12-04T09:28:45.8736953Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8737081Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8737155Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8737230Z graph_break [] 2025-12-04T09:28:45.8737308Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8737441Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8737507Z frames [('total', 1)] 2025-12-04T09:28:45.8737579Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8737712Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8737785Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8737846Z graph_break [] 2025-12-04T09:28:45.8737940Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8738072Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8738135Z frames [('total', 1)] 2025-12-04T09:28:45.8738215Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8738341Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8738422Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8738481Z graph_break [] 2025-12-04T09:28:45.8738557Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8738689Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8738755Z frames [('total', 1)] 2025-12-04T09:28:45.8738825Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8738960Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8739031Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8739095Z graph_break [] 2025-12-04T09:28:45.8739177Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8739300Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8739369Z frames [('total', 1)] 2025-12-04T09:28:45.8739438Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8739564Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8739692Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8739751Z graph_break [] 2025-12-04T09:28:45.8739828Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8739956Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8740074Z frames [('total', 1)] 2025-12-04T09:28:45.8740176Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8740310Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8740382Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8740446Z graph_break [] 2025-12-04T09:28:45.8740522Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8740657Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8740724Z frames [('total', 1)] 2025-12-04T09:28:45.8740826Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8740955Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8741033Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8741093Z graph_break [] 2025-12-04T09:28:45.8741168Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8741298Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8741362Z frames [('total', 1)] 2025-12-04T09:28:45.8741439Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8741564Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8741637Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8741703Z graph_break [] 2025-12-04T09:28:45.8741779Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8741907Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8741974Z frames [('total', 1)] 2025-12-04T09:28:45.8742047Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8742172Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8742250Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8742313Z graph_break [] 2025-12-04T09:28:45.8742389Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8742520Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8742584Z frames [('total', 1)] 2025-12-04T09:28:45.8742657Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8742787Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8742859Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8742929Z graph_break [] 2025-12-04T09:28:45.8743005Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8743130Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8743210Z frames [('total', 1)] 2025-12-04T09:28:45.8743282Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8743415Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8743495Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8743557Z graph_break [] 2025-12-04T09:28:45.8743640Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8743764Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8743829Z frames [('total', 1)] 2025-12-04T09:28:45.8743905Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8744032Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8744105Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8744172Z graph_break [] 2025-12-04T09:28:45.8744250Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8744375Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8744446Z frames [('total', 1)] 2025-12-04T09:28:45.8744513Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8744651Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8744724Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8744828Z graph_break [] 2025-12-04T09:28:45.8744910Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8745038Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8745136Z frames [('total', 1)] 2025-12-04T09:28:45.8745211Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8745370Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8745449Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8745516Z graph_break [] 2025-12-04T09:28:45.8745594Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8745728Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8745793Z frames [('total', 1)] 2025-12-04T09:28:45.8745863Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8746038Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8746117Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8746178Z graph_break [] 2025-12-04T09:28:45.8746265Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8746391Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8746454Z frames [('total', 1)] 2025-12-04T09:28:45.8746532Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8746664Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8746748Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8746810Z graph_break [] 2025-12-04T09:28:45.8746891Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8747029Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8747092Z frames [('total', 1)] 2025-12-04T09:28:45.8747165Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8747301Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8747374Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8747436Z graph_break [] 2025-12-04T09:28:45.8747520Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8747643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8747711Z frames [('total', 1)] 2025-12-04T09:28:45.8747782Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8747912Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8747989Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8748048Z graph_break [] 2025-12-04T09:28:45.8748125Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8748257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8748332Z frames [('total', 1)] 2025-12-04T09:28:45.8748403Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8748541Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8748612Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8748676Z graph_break [] 2025-12-04T09:28:45.8748754Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8748878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8748946Z frames [('total', 1)] 2025-12-04T09:28:45.8749016Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8749141Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8749220Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8749279Z graph_break [] 2025-12-04T09:28:45.8749356Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8749487Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8749548Z frames [('total', 1)] 2025-12-04T09:28:45.8749623Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8749752Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8749825Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8749892Z graph_break [] 2025-12-04T09:28:45.8749969Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8750143Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8750211Z frames [('total', 1)] 2025-12-04T09:28:45.8750315Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8750443Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8750554Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8750614Z graph_break [] 2025-12-04T09:28:45.8750692Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8750824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8750885Z frames [('total', 1)] 2025-12-04T09:28:45.8750959Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8751085Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8751190Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8751257Z graph_break [] 2025-12-04T09:28:45.8751334Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8751464Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8751545Z frames [('total', 1)] 2025-12-04T09:28:45.8751615Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8751745Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8751826Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8751885Z graph_break [] 2025-12-04T09:28:45.8751967Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8752093Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8752155Z frames [('total', 1)] 2025-12-04T09:28:45.8752230Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8752355Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8752431Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8752498Z graph_break [] 2025-12-04T09:28:45.8752573Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8752702Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8752776Z frames [('total', 1)] 2025-12-04T09:28:45.8752845Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8752981Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8753059Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8753123Z graph_break [] 2025-12-04T09:28:45.8753215Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8753345Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8753408Z frames [('total', 1)] 2025-12-04T09:28:45.8753484Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8753609Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8753685Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8753752Z graph_break [] 2025-12-04T09:28:45.8753829Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8753963Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8754028Z frames [('total', 1)] 2025-12-04T09:28:45.8754099Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8754237Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8754310Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8754371Z graph_break [] 2025-12-04T09:28:45.8754454Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8754592Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8754656Z frames [('total', 1)] 2025-12-04T09:28:45.8754732Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8754859Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8754943Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8755004Z graph_break [] 2025-12-04T09:28:45.8755081Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8755262Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8755326Z frames [('total', 1)] 2025-12-04T09:28:45.8755396Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8755531Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8755639Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8755734Z graph_break [] 2025-12-04T09:28:45.8755816Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8755941Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8756008Z frames [('total', 1)] 2025-12-04T09:28:45.8756077Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8756202Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8756281Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8756373Z graph_break [] 2025-12-04T09:28:45.8756453Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8756583Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8756647Z frames [('total', 1)] 2025-12-04T09:28:45.8756715Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8756846Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8756920Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8756987Z graph_break [] 2025-12-04T09:28:45.8757064Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8757189Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8757257Z frames [('total', 1)] 2025-12-04T09:28:45.8757326Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8757453Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8757531Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8757593Z graph_break [] 2025-12-04T09:28:45.8757669Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8757799Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8757865Z frames [('total', 1)] 2025-12-04T09:28:45.8757937Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8758066Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8758143Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8758207Z graph_break [] 2025-12-04T09:28:45.8758287Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8758411Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8758480Z frames [('total', 1)] 2025-12-04T09:28:45.8758547Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8758671Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8758748Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8758810Z graph_break [] 2025-12-04T09:28:45.8758887Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8759017Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8759082Z frames [('total', 1)] 2025-12-04T09:28:45.8759155Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8759283Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8759356Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8759421Z graph_break [] 2025-12-04T09:28:45.8759500Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8759626Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8759695Z frames [('total', 1)] 2025-12-04T09:28:45.8759763Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8759889Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8759968Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8760027Z graph_break [] 2025-12-04T09:28:45.8760110Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8760236Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8760363Z frames [('total', 1)] 2025-12-04T09:28:45.8760446Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8760576Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8760684Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8760754Z graph_break [] 2025-12-04T09:28:45.8760863Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8760990Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8761059Z frames [('total', 1)] 2025-12-04T09:28:45.8761127Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8761257Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8761330Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8761389Z graph_break [] 2025-12-04T09:28:45.8761504Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8761641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8761708Z frames [('total', 1)] 2025-12-04T09:28:45.8761782Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8761909Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8761987Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8762054Z graph_break [] 2025-12-04T09:28:45.8762136Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8762308Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8762386Z Traceback (most recent call last): 2025-12-04T09:28:45.8762703Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8762768Z f(a, b) 2025-12-04T09:28:45.8763059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8763211Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8763523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8763646Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8763954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8764050Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8764361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8764563Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8764872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8764958Z graph.run(*example_inputs) 2025-12-04T09:28:45.8765194Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8765276Z return super().run(*args) 2025-12-04T09:28:45.8765521Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8765604Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8765858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8765938Z result = super().run_node(n) 2025-12-04T09:28:45.8766182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8766292Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8766560Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8766706Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8767016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8767147Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8767435Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8767545Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8767804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8767887Z return autotune_select_algorithm( 2025-12-04T09:28:45.8768227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8768301Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8768630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8768793Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8769546Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8769618Z target: aten.mm.default 2025-12-04T09:28:45.8769680Z args[0]: TensorBox( 2025-12-04T09:28:45.8769751Z ReinterpretView( 2025-12-04T09:28:45.8769810Z StorageBox( 2025-12-04T09:28:45.8770052Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8770116Z ), 2025-12-04T09:28:45.8770267Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8770345Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8770408Z stack_traces = {, 2025-12-04T09:28:45.8770638Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8770718Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8770774Z , 2025-12-04T09:28:45.8770832Z } 2025-12-04T09:28:45.8770905Z ) 2025-12-04T09:28:45.8770960Z ) 2025-12-04T09:28:45.8771038Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8771275Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8771331Z )) 2025-12-04T09:28:45.8771335Z 2025-12-04T09:28:45.8771759Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8771763Z 2025-12-04T09:28:45.8771769Z 2025-12-04T09:28:45.8771902Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8772146Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8772152Z 2025-12-04T09:28:45.8772320Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8772452Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8772524Z frames [('total', 1)] 2025-12-04T09:28:45.8772597Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8772671Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8772811Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8772871Z graph_break [] 2025-12-04T09:28:45.8772953Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8773088Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8773244Z frames [('total', 1)] 2025-12-04T09:28:45.8773313Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8773448Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8773571Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8773635Z graph_break [] 2025-12-04T09:28:45.8773713Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8773840Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8773954Z frames [('total', 1)] 2025-12-04T09:28:45.8774059Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8774190Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8774270Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8774331Z graph_break [] 2025-12-04T09:28:45.8774408Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8774540Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8774603Z frames [('total', 1)] 2025-12-04T09:28:45.8774713Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8774840Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8774914Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8774980Z graph_break [] 2025-12-04T09:28:45.8775055Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8775180Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8775250Z frames [('total', 1)] 2025-12-04T09:28:45.8775322Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8775447Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8775525Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8775585Z graph_break [] 2025-12-04T09:28:45.8775666Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8775792Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8775859Z frames [('total', 1)] 2025-12-04T09:28:45.8775948Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8776075Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8776151Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8776217Z graph_break [] 2025-12-04T09:28:45.8776294Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8776422Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8776496Z frames [('total', 1)] 2025-12-04T09:28:45.8776565Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8776699Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8776775Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8776834Z graph_break [] 2025-12-04T09:28:45.8776921Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8777046Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8777109Z frames [('total', 1)] 2025-12-04T09:28:45.8777188Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8777322Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8777396Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8777467Z graph_break [] 2025-12-04T09:28:45.8777544Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8777674Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8777740Z frames [('total', 1)] 2025-12-04T09:28:45.8777809Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8777940Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8778014Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8778073Z graph_break [] 2025-12-04T09:28:45.8778157Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8778283Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8778345Z frames [('total', 1)] 2025-12-04T09:28:45.8778420Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8778547Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8778625Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8778736Z graph_break [] 2025-12-04T09:28:45.8778813Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8778942Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8779037Z frames [('total', 1)] 2025-12-04T09:28:45.8779105Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8779270Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8779344Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8779402Z graph_break [] 2025-12-04T09:28:45.8779483Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8779606Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8779673Z frames [('total', 1)] 2025-12-04T09:28:45.8779741Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8779928Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8780010Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8780072Z graph_break [] 2025-12-04T09:28:45.8780149Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8780278Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8780343Z frames [('total', 1)] 2025-12-04T09:28:45.8780409Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8780542Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8780614Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8780677Z graph_break [] 2025-12-04T09:28:45.8780754Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8780878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8780947Z frames [('total', 1)] 2025-12-04T09:28:45.8781016Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8781144Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8781223Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8781285Z graph_break [] 2025-12-04T09:28:45.8781365Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8781496Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8781561Z frames [('total', 1)] 2025-12-04T09:28:45.8781628Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8781761Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8781833Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8781901Z graph_break [] 2025-12-04T09:28:45.8781988Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8782113Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8782181Z frames [('total', 1)] 2025-12-04T09:28:45.8782248Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8782377Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8782454Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8782513Z graph_break [] 2025-12-04T09:28:45.8782593Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8782724Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8782786Z frames [('total', 1)] 2025-12-04T09:28:45.8782863Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8782991Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8783064Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8783133Z graph_break [] 2025-12-04T09:28:45.8783209Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8783343Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8783415Z frames [('total', 1)] 2025-12-04T09:28:45.8783484Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8783612Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8783693Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8783751Z graph_break [] 2025-12-04T09:28:45.8783880Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8784005Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8784068Z frames [('total', 1)] 2025-12-04T09:28:45.8787610Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8787867Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8787952Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8788023Z graph_break [] 2025-12-04T09:28:45.8788110Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8788255Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8788326Z frames [('total', 1)] 2025-12-04T09:28:45.8788399Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8788594Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8788677Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8788738Z graph_break [] 2025-12-04T09:28:45.8788830Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8788970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8789039Z frames [('total', 1)] 2025-12-04T09:28:45.8789120Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8789260Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8789341Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8789408Z graph_break [] 2025-12-04T09:28:45.8789490Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8789630Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8789692Z frames [('total', 1)] 2025-12-04T09:28:45.8789760Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8789893Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8789967Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8790027Z graph_break [] 2025-12-04T09:28:45.8790112Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8790245Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8790306Z frames [('total', 1)] 2025-12-04T09:28:45.8790379Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8790506Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8790584Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8790645Z graph_break [] 2025-12-04T09:28:45.8790723Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8790851Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8790912Z frames [('total', 1)] 2025-12-04T09:28:45.8790982Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8791111Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8791187Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8791245Z graph_break [] 2025-12-04T09:28:45.8791328Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8791455Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8791518Z frames [('total', 1)] 2025-12-04T09:28:45.8791588Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8791714Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8791792Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8791850Z graph_break [] 2025-12-04T09:28:45.8791926Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8792055Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8792117Z frames [('total', 1)] 2025-12-04T09:28:45.8792185Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8792315Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8792401Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8792461Z graph_break [] 2025-12-04T09:28:45.8792542Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8792714Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8792779Z frames [('total', 1)] 2025-12-04T09:28:45.8792849Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8793005Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8793117Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8793178Z graph_break [] 2025-12-04T09:28:45.8793259Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8793389Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8793450Z frames [('total', 1)] 2025-12-04T09:28:45.8793516Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8793646Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8793752Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8793816Z graph_break [] 2025-12-04T09:28:45.8793891Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8794015Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8794089Z frames [('total', 1)] 2025-12-04T09:28:45.8794160Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8794288Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8794369Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8794432Z graph_break [] 2025-12-04T09:28:45.8794508Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8794641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8794704Z frames [('total', 1)] 2025-12-04T09:28:45.8794775Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8794900Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8794974Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8795038Z graph_break [] 2025-12-04T09:28:45.8795114Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8795239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8795308Z frames [('total', 1)] 2025-12-04T09:28:45.8795375Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8795498Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8795578Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8795641Z graph_break [] 2025-12-04T09:28:45.8795722Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8795848Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8795911Z frames [('total', 1)] 2025-12-04T09:28:45.8795983Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8796106Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8796178Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8796244Z graph_break [] 2025-12-04T09:28:45.8796318Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8796441Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8796512Z frames [('total', 1)] 2025-12-04T09:28:45.8796579Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8796705Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8796778Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8796836Z graph_break [] 2025-12-04T09:28:45.8796921Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8797057Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8797130Z frames [('total', 1)] 2025-12-04T09:28:45.8797209Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8797336Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8797410Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8797481Z graph_break [] 2025-12-04T09:28:45.8797560Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8797692Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8797847Z frames [('total', 1)] 2025-12-04T09:28:45.8797916Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8798050Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8798160Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8798223Z graph_break [] 2025-12-04T09:28:45.8798341Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8798470Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8798532Z frames [('total', 1)] 2025-12-04T09:28:45.8798607Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8798732Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8798811Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8798904Z graph_break [] 2025-12-04T09:28:45.8798984Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8799118Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8799182Z frames [('total', 1)] 2025-12-04T09:28:45.8799250Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8799384Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8799459Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8799518Z graph_break [] 2025-12-04T09:28:45.8799606Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8799733Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8799795Z frames [('total', 1)] 2025-12-04T09:28:45.8799868Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8799992Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8800064Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8800131Z graph_break [] 2025-12-04T09:28:45.8800223Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8800355Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8800419Z frames [('total', 1)] 2025-12-04T09:28:45.8800486Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8800616Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8800692Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8800753Z graph_break [] 2025-12-04T09:28:45.8800841Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8800968Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8801031Z frames [('total', 1)] 2025-12-04T09:28:45.8801108Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8801233Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8801311Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8801371Z graph_break [] 2025-12-04T09:28:45.8801451Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8801582Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8801647Z frames [('total', 1)] 2025-12-04T09:28:45.8801715Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8801845Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8801928Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8801990Z graph_break [] 2025-12-04T09:28:45.8802077Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8802200Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8802268Z frames [('total', 1)] 2025-12-04T09:28:45.8802335Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8802458Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8802538Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8802598Z graph_break [] 2025-12-04T09:28:45.8802678Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8802811Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8802918Z frames [('total', 1)] 2025-12-04T09:28:45.8802988Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8803122Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8803195Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8803298Z graph_break [] 2025-12-04T09:28:45.8803376Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8803595Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8803684Z Traceback (most recent call last): 2025-12-04T09:28:45.8804013Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8804070Z f(a, b) 2025-12-04T09:28:45.8804427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8804798Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8805113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8805238Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8805533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8805640Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8805957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8806154Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8806479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8806556Z graph.run(*example_inputs) 2025-12-04T09:28:45.8806802Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8806872Z return super().run(*args) 2025-12-04T09:28:45.8807121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8807210Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8807461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8807533Z result = super().run_node(n) 2025-12-04T09:28:45.8807785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8807887Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8808161Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8808309Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8808573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8808714Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8808967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8809047Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8809307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8809385Z return autotune_select_algorithm( 2025-12-04T09:28:45.8809727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8809798Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8810083Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8810249Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8811128Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8811253Z target: aten.mm.default 2025-12-04T09:28:45.8811318Z args[0]: TensorBox( 2025-12-04T09:28:45.8811384Z ReinterpretView( 2025-12-04T09:28:45.8811448Z StorageBox( 2025-12-04T09:28:45.8811689Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8811748Z ), 2025-12-04T09:28:45.8811958Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8812032Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8812103Z stack_traces = {, 2025-12-04T09:28:45.8812331Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8812407Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8812470Z , 2025-12-04T09:28:45.8812526Z } 2025-12-04T09:28:45.8812581Z ) 2025-12-04T09:28:45.8812640Z ) 2025-12-04T09:28:45.8812714Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8812954Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8813014Z )) 2025-12-04T09:28:45.8813019Z 2025-12-04T09:28:45.8813534Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8813539Z 2025-12-04T09:28:45.8813542Z 2025-12-04T09:28:45.8813686Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8813929Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8813935Z 2025-12-04T09:28:45.8814107Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8814239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8814309Z frames [('total', 1)] 2025-12-04T09:28:45.8814389Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8814465Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8814602Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8814670Z graph_break [] 2025-12-04T09:28:45.8814754Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8814880Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8814961Z frames [('total', 1)] 2025-12-04T09:28:45.8815035Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8815170Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8815245Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8815305Z graph_break [] 2025-12-04T09:28:45.8815390Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8815517Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8815581Z frames [('total', 1)] 2025-12-04T09:28:45.8815659Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8815787Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8815861Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8815930Z graph_break [] 2025-12-04T09:28:45.8816007Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8816138Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8816200Z frames [('total', 1)] 2025-12-04T09:28:45.8816271Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8816402Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8816480Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8816587Z graph_break [] 2025-12-04T09:28:45.8816673Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8816803Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8816901Z frames [('total', 1)] 2025-12-04T09:28:45.8816978Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8817140Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8817222Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8817283Z graph_break [] 2025-12-04T09:28:45.8817361Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8817492Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8817555Z frames [('total', 1)] 2025-12-04T09:28:45.8817657Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8817792Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8817867Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8817929Z graph_break [] 2025-12-04T09:28:45.8818017Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8818141Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8818216Z frames [('total', 1)] 2025-12-04T09:28:45.8818290Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8818420Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8818502Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8818562Z graph_break [] 2025-12-04T09:28:45.8818642Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8818775Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8818837Z frames [('total', 1)] 2025-12-04T09:28:45.8818906Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8819041Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8819115Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8819183Z graph_break [] 2025-12-04T09:28:45.8819260Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8819383Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8819454Z frames [('total', 1)] 2025-12-04T09:28:45.8819521Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8819645Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8819722Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8819783Z graph_break [] 2025-12-04T09:28:45.8819859Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8819987Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8820049Z frames [('total', 1)] 2025-12-04T09:28:45.8820117Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8820247Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8820319Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8820385Z graph_break [] 2025-12-04T09:28:45.8820463Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8820588Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8820657Z frames [('total', 1)] 2025-12-04T09:28:45.8820724Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8820850Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8820929Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8820988Z graph_break [] 2025-12-04T09:28:45.8821063Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8821195Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8821257Z frames [('total', 1)] 2025-12-04T09:28:45.8821331Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8821456Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8821529Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8821641Z graph_break [] 2025-12-04T09:28:45.8821718Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8821843Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8821946Z frames [('total', 1)] 2025-12-04T09:28:45.8822014Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8822171Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8822250Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8822312Z graph_break [] 2025-12-04T09:28:45.8822397Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8822521Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8822584Z frames [('total', 1)] 2025-12-04T09:28:45.8822656Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8822824Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8822901Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8822967Z graph_break [] 2025-12-04T09:28:45.8823046Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8823174Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8823243Z frames [('total', 1)] 2025-12-04T09:28:45.8823317Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8823450Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8823523Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8823581Z graph_break [] 2025-12-04T09:28:45.8823665Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8823789Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8823852Z frames [('total', 1)] 2025-12-04T09:28:45.8823925Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8824051Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8824125Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8824191Z graph_break [] 2025-12-04T09:28:45.8824270Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8824399Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8824463Z frames [('total', 1)] 2025-12-04T09:28:45.8824534Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8824665Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8824740Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8824800Z graph_break [] 2025-12-04T09:28:45.8824893Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8825019Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8825081Z frames [('total', 1)] 2025-12-04T09:28:45.8825156Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8825281Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8825360Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8825420Z graph_break [] 2025-12-04T09:28:45.8825498Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8825632Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8825694Z frames [('total', 1)] 2025-12-04T09:28:45.8825764Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8825898Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8825972Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8826033Z graph_break [] 2025-12-04T09:28:45.8826115Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8826239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8826308Z frames [('total', 1)] 2025-12-04T09:28:45.8826377Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8826504Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8826598Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8826669Z graph_break [] 2025-12-04T09:28:45.8826760Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8826968Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8827044Z frames [('total', 1)] 2025-12-04T09:28:45.8827122Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8827347Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8827474Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8827553Z graph_break [] 2025-12-04T09:28:45.8827643Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8827793Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8827881Z frames [('total', 1)] 2025-12-04T09:28:45.8827961Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8828108Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8828245Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8828317Z graph_break [] 2025-12-04T09:28:45.8828416Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8828566Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8828639Z frames [('total', 1)] 2025-12-04T09:28:45.8828737Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8828893Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8828976Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8829042Z graph_break [] 2025-12-04T09:28:45.8829119Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8829243Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8829317Z frames [('total', 1)] 2025-12-04T09:28:45.8829385Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8829515Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8829589Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8829652Z graph_break [] 2025-12-04T09:28:45.8829732Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8829860Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8829924Z frames [('total', 1)] 2025-12-04T09:28:45.8829997Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8830122Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8830195Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8830263Z graph_break [] 2025-12-04T09:28:45.8830340Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8830471Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8830543Z frames [('total', 1)] 2025-12-04T09:28:45.8830613Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8830745Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8830817Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8830875Z graph_break [] 2025-12-04T09:28:45.8830956Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8831082Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8831144Z frames [('total', 1)] 2025-12-04T09:28:45.8831216Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8831340Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8831420Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8831482Z graph_break [] 2025-12-04T09:28:45.8831558Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8831686Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8831747Z frames [('total', 1)] 2025-12-04T09:28:45.8831817Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8831948Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8832018Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8832080Z graph_break [] 2025-12-04T09:28:45.8832162Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8832287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8832394Z frames [('total', 1)] 2025-12-04T09:28:45.8832469Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8832596Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8832707Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8832808Z graph_break [] 2025-12-04T09:28:45.8832888Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8833019Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8833080Z frames [('total', 1)] 2025-12-04T09:28:45.8833147Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8833276Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8833346Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8833440Z graph_break [] 2025-12-04T09:28:45.8833524Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8833648Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8833716Z frames [('total', 1)] 2025-12-04T09:28:45.8833785Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8833911Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8833990Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8834050Z graph_break [] 2025-12-04T09:28:45.8834130Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8834260Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8834322Z frames [('total', 1)] 2025-12-04T09:28:45.8834389Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8834518Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8834589Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8834655Z graph_break [] 2025-12-04T09:28:45.8834732Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8834855Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8834926Z frames [('total', 1)] 2025-12-04T09:28:45.8834996Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8835132Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8835213Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8835274Z graph_break [] 2025-12-04T09:28:45.8835354Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8835484Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8835547Z frames [('total', 1)] 2025-12-04T09:28:45.8835621Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8835746Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8835818Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8835890Z graph_break [] 2025-12-04T09:28:45.8835969Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8836096Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8836168Z frames [('total', 1)] 2025-12-04T09:28:45.8836237Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8836364Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8836445Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8836505Z graph_break [] 2025-12-04T09:28:45.8836591Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8836720Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8836781Z frames [('total', 1)] 2025-12-04T09:28:45.8836855Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8836989Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8837064Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8837139Z graph_break [] 2025-12-04T09:28:45.8837218Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8837342Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8837455Z frames [('total', 1)] 2025-12-04T09:28:45.8837523Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8837654Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8837761Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8837821Z graph_break [] 2025-12-04T09:28:45.8837939Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8838065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8838127Z frames [('total', 1)] 2025-12-04T09:28:45.8838198Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8838325Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8838396Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8838461Z graph_break [] 2025-12-04T09:28:45.8838573Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8838704Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8838768Z frames [('total', 1)] 2025-12-04T09:28:45.8838836Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8838965Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8839040Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8839100Z graph_break [] 2025-12-04T09:28:45.8839184Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8839309Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8839370Z frames [('total', 1)] 2025-12-04T09:28:45.8839442Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8839568Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8839644Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8839703Z graph_break [] 2025-12-04T09:28:45.8839780Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8839909Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8839973Z frames [('total', 1)] 2025-12-04T09:28:45.8840041Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8840169Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8840242Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8840304Z graph_break [] 2025-12-04T09:28:45.8840386Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8840513Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8840585Z frames [('total', 1)] 2025-12-04T09:28:45.8840660Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8840785Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8840860Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8840919Z graph_break [] 2025-12-04T09:28:45.8840997Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8841127Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8841188Z frames [('total', 1)] 2025-12-04T09:28:45.8841259Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8841389Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8841461Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8841521Z graph_break [] 2025-12-04T09:28:45.8841605Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8841730Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8841798Z frames [('total', 1)] 2025-12-04T09:28:45.8841867Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8841991Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8842068Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8842126Z graph_break [] 2025-12-04T09:28:45.8842203Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8842372Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8842449Z Traceback (most recent call last): 2025-12-04T09:28:45.8842809Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8842872Z f(a, b) 2025-12-04T09:28:45.8843197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8843390Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8843691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8843810Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8844111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8844239Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8844564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8844767Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8845067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8845152Z graph.run(*example_inputs) 2025-12-04T09:28:45.8845385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8845454Z return super().run(*args) 2025-12-04T09:28:45.8845686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8845765Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8846023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8846093Z result = super().run_node(n) 2025-12-04T09:28:45.8846342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8846457Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8846722Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8846883Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8847145Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8847281Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8847543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8847616Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8847875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8847964Z return autotune_select_algorithm( 2025-12-04T09:28:45.8848303Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8848388Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8848682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8848844Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8849598Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8849671Z target: aten.mm.default 2025-12-04T09:28:45.8849742Z args[0]: TensorBox( 2025-12-04T09:28:45.8849808Z ReinterpretView( 2025-12-04T09:28:45.8849929Z StorageBox( 2025-12-04T09:28:45.8850176Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8850233Z ), 2025-12-04T09:28:45.8850416Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8850532Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8850597Z stack_traces = {, 2025-12-04T09:28:45.8850821Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8850905Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8850962Z , 2025-12-04T09:28:45.8851024Z } 2025-12-04T09:28:45.8851078Z ) 2025-12-04T09:28:45.8851131Z ) 2025-12-04T09:28:45.8851211Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8851483Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8851542Z )) 2025-12-04T09:28:45.8851547Z 2025-12-04T09:28:45.8851975Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8851982Z 2025-12-04T09:28:45.8851984Z 2025-12-04T09:28:45.8852119Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8852366Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8852370Z 2025-12-04T09:28:45.8852531Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8852664Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8852733Z frames [('total', 1)] 2025-12-04T09:28:45.8852806Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8852886Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8853018Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8853082Z graph_break [] 2025-12-04T09:28:45.8853258Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8853387Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8853452Z frames [('total', 1)] 2025-12-04T09:28:45.8853527Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8853657Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8853730Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8853793Z graph_break [] 2025-12-04T09:28:45.8853871Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8854001Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8854066Z frames [('total', 1)] 2025-12-04T09:28:45.8854134Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8854265Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8854335Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8854396Z graph_break [] 2025-12-04T09:28:45.8854479Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8854602Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8854666Z frames [('total', 1)] 2025-12-04T09:28:45.8854742Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8854869Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8854941Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8855007Z graph_break [] 2025-12-04T09:28:45.8855083Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8855214Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8855281Z frames [('total', 1)] 2025-12-04T09:28:45.8855350Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8855484Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8855557Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8855661Z graph_break [] 2025-12-04T09:28:45.8855747Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8855873Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8855970Z frames [('total', 1)] 2025-12-04T09:28:45.8856054Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8856213Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8856294Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8856354Z graph_break [] 2025-12-04T09:28:45.8856431Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8856562Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8856627Z frames [('total', 1)] 2025-12-04T09:28:45.8856695Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8856865Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8856939Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8857002Z graph_break [] 2025-12-04T09:28:45.8857084Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8857207Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8857277Z frames [('total', 1)] 2025-12-04T09:28:45.8857343Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8857471Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8857548Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8857608Z graph_break [] 2025-12-04T09:28:45.8857684Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8857815Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8857876Z frames [('total', 1)] 2025-12-04T09:28:45.8857944Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8858077Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8858149Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8858214Z graph_break [] 2025-12-04T09:28:45.8858292Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8858417Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8858485Z frames [('total', 1)] 2025-12-04T09:28:45.8858558Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8858685Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8858765Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8858828Z graph_break [] 2025-12-04T09:28:45.8858905Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8859034Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8859098Z frames [('total', 1)] 2025-12-04T09:28:45.8859178Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8859308Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8859383Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8859449Z graph_break [] 2025-12-04T09:28:45.8859531Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8859657Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8859724Z frames [('total', 1)] 2025-12-04T09:28:45.8859794Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8859921Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8860001Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8860062Z graph_break [] 2025-12-04T09:28:45.8860144Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8860271Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8860335Z frames [('total', 1)] 2025-12-04T09:28:45.8860409Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8860534Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8860617Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8860684Z graph_break [] 2025-12-04T09:28:45.8860806Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8860930Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8860997Z frames [('total', 1)] 2025-12-04T09:28:45.8861099Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8861233Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8861338Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8861398Z graph_break [] 2025-12-04T09:28:45.8861479Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8861603Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8861664Z frames [('total', 1)] 2025-12-04T09:28:45.8861739Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8861896Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8861970Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8862035Z graph_break [] 2025-12-04T09:28:45.8862111Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8862245Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8862307Z frames [('total', 1)] 2025-12-04T09:28:45.8862376Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8862513Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8862594Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8862654Z graph_break [] 2025-12-04T09:28:45.8862737Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8862862Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8862923Z frames [('total', 1)] 2025-12-04T09:28:45.8862996Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8863122Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8863202Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8863262Z graph_break [] 2025-12-04T09:28:45.8863337Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8863471Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8863533Z frames [('total', 1)] 2025-12-04T09:28:45.8863599Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8863735Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8863810Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8863872Z graph_break [] 2025-12-04T09:28:45.8863953Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8864078Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8864141Z frames [('total', 1)] 2025-12-04T09:28:45.8864213Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8864338Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8864416Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8864476Z graph_break [] 2025-12-04T09:28:45.8864553Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8864687Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8864749Z frames [('total', 1)] 2025-12-04T09:28:45.8864817Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8864952Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8865025Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8865087Z graph_break [] 2025-12-04T09:28:45.8865170Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8865295Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8865366Z frames [('total', 1)] 2025-12-04T09:28:45.8865434Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8865558Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8865645Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8865705Z graph_break [] 2025-12-04T09:28:45.8865786Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8865961Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8866023Z frames [('total', 1)] 2025-12-04T09:28:45.8866101Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8866267Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8866340Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8866444Z graph_break [] 2025-12-04T09:28:45.8866522Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8866649Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8866717Z frames [('total', 1)] 2025-12-04T09:28:45.8866787Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8866919Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8867052Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8867115Z graph_break [] 2025-12-04T09:28:45.8867194Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8867330Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8867391Z frames [('total', 1)] 2025-12-04T09:28:45.8867469Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8867594Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8867670Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8867740Z graph_break [] 2025-12-04T09:28:45.8867819Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8867943Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8868012Z frames [('total', 1)] 2025-12-04T09:28:45.8868082Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8868208Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8868288Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8868349Z graph_break [] 2025-12-04T09:28:45.8868433Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8868558Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8868623Z frames [('total', 1)] 2025-12-04T09:28:45.8868699Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8868822Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8868896Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8868963Z graph_break [] 2025-12-04T09:28:45.8869040Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8869166Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8869235Z frames [('total', 1)] 2025-12-04T09:28:45.8869304Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8869433Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8869504Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8869565Z graph_break [] 2025-12-04T09:28:45.8869645Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8869780Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8869845Z frames [('total', 1)] 2025-12-04T09:28:45.8869920Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8870044Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8870117Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8870183Z graph_break [] 2025-12-04T09:28:45.8870263Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8870394Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8870456Z frames [('total', 1)] 2025-12-04T09:28:45.8870526Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8870655Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8870728Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8870790Z graph_break [] 2025-12-04T09:28:45.8870871Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8870995Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8871102Z frames [('total', 1)] 2025-12-04T09:28:45.8871176Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8871305Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8871419Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8871479Z graph_break [] 2025-12-04T09:28:45.8871592Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8871730Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8871792Z frames [('total', 1)] 2025-12-04T09:28:45.8871864Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8872001Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8872074Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8872134Z graph_break [] 2025-12-04T09:28:45.8872259Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8872395Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8872463Z frames [('total', 1)] 2025-12-04T09:28:45.8872538Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8872666Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8872748Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8872808Z graph_break [] 2025-12-04T09:28:45.8872890Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8873025Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8873089Z frames [('total', 1)] 2025-12-04T09:28:45.8873158Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8873293Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8873365Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8873430Z graph_break [] 2025-12-04T09:28:45.8873517Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8873642Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8873714Z frames [('total', 1)] 2025-12-04T09:28:45.8873784Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8873908Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8873990Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8874052Z graph_break [] 2025-12-04T09:28:45.8874134Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8874277Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8874342Z frames [('total', 1)] 2025-12-04T09:28:45.8874413Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8874545Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8874619Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8874683Z graph_break [] 2025-12-04T09:28:45.8874763Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8874890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8874961Z frames [('total', 1)] 2025-12-04T09:28:45.8875030Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8875156Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8875240Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8875303Z graph_break [] 2025-12-04T09:28:45.8875382Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8875518Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8875580Z frames [('total', 1)] 2025-12-04T09:28:45.8875658Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8875793Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8875870Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8875938Z graph_break [] 2025-12-04T09:28:45.8876018Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8876144Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8876215Z frames [('total', 1)] 2025-12-04T09:28:45.8876332Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8876458Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8876538Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8876647Z graph_break [] 2025-12-04T09:28:45.8876732Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8876893Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8876957Z frames [('total', 1)] 2025-12-04T09:28:45.8877033Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8877159Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8877229Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8877294Z graph_break [] 2025-12-04T09:28:45.8877371Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8877529Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8877599Z frames [('total', 1)] 2025-12-04T09:28:45.8877684Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8877816Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8877887Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8877948Z graph_break [] 2025-12-04T09:28:45.8878031Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8878156Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8878218Z frames [('total', 1)] 2025-12-04T09:28:45.8878291Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8878415Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8878486Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8878552Z graph_break [] 2025-12-04T09:28:45.8878629Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8878758Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8878821Z frames [('total', 1)] 2025-12-04T09:28:45.8878889Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8879018Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8879091Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8879155Z graph_break [] 2025-12-04T09:28:45.8879237Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8879363Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8879429Z frames [('total', 1)] 2025-12-04T09:28:45.8879504Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8879629Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8879705Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8879765Z graph_break [] 2025-12-04T09:28:45.8879841Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8879972Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8880046Z frames [('total', 1)] 2025-12-04T09:28:45.8880117Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8880247Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8880317Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8880376Z graph_break [] 2025-12-04T09:28:45.8880460Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8880585Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8880648Z frames [('total', 1)] 2025-12-04T09:28:45.8880720Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8880844Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8880927Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8880988Z graph_break [] 2025-12-04T09:28:45.8881068Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8881241Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8881319Z Traceback (most recent call last): 2025-12-04T09:28:45.8881634Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8881751Z f(a, b) 2025-12-04T09:28:45.8882045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8882268Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8882570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8882690Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8882988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8883155Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8883471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8883674Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8883971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8884054Z graph.run(*example_inputs) 2025-12-04T09:28:45.8884289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8884358Z return super().run(*args) 2025-12-04T09:28:45.8884594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8884672Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8884932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8885011Z result = super().run_node(n) 2025-12-04T09:28:45.8885260Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8885372Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8885640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8885795Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8886063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8886196Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8886463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8886535Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8886795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8886883Z return autotune_select_algorithm( 2025-12-04T09:28:45.8887223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8887304Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8887595Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8887757Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8888510Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8888584Z target: aten.mm.default 2025-12-04T09:28:45.8888648Z args[0]: TensorBox( 2025-12-04T09:28:45.8888723Z ReinterpretView( 2025-12-04T09:28:45.8888785Z StorageBox( 2025-12-04T09:28:45.8889075Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8889133Z ), 2025-12-04T09:28:45.8889284Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8889399Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8889518Z stack_traces = {, 2025-12-04T09:28:45.8889750Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8889834Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8889891Z , 2025-12-04T09:28:45.8889947Z } 2025-12-04T09:28:45.8890006Z ) 2025-12-04T09:28:45.8890070Z ) 2025-12-04T09:28:45.8890145Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8890419Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8890474Z )) 2025-12-04T09:28:45.8890480Z 2025-12-04T09:28:45.8890906Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8890913Z 2025-12-04T09:28:45.8890916Z 2025-12-04T09:28:45.8891050Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8891303Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8891307Z 2025-12-04T09:28:45.8891472Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8891605Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8891675Z frames [('total', 1)] 2025-12-04T09:28:45.8891745Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8891820Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8891961Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8892024Z graph_break [] 2025-12-04T09:28:45.8892114Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8892240Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8892305Z frames [('total', 1)] 2025-12-04T09:28:45.8892380Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8892510Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8892582Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8892648Z graph_break [] 2025-12-04T09:28:45.8892726Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8892851Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8892920Z frames [('total', 1)] 2025-12-04T09:28:45.8892988Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8893219Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8893296Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8893359Z graph_break [] 2025-12-04T09:28:45.8893443Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8893568Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8893633Z frames [('total', 1)] 2025-12-04T09:28:45.8893707Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8893835Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8893907Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8893973Z graph_break [] 2025-12-04T09:28:45.8894050Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8894180Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8894243Z frames [('total', 1)] 2025-12-04T09:28:45.8894311Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8894453Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8894529Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8894590Z graph_break [] 2025-12-04T09:28:45.8894718Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8894843Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8894905Z frames [('total', 1)] 2025-12-04T09:28:45.8895017Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8895173Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8895247Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8895312Z graph_break [] 2025-12-04T09:28:45.8895389Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8895522Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8895584Z frames [('total', 1)] 2025-12-04T09:28:45.8895653Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8895825Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8895899Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8895968Z graph_break [] 2025-12-04T09:28:45.8896059Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8896182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8896245Z frames [('total', 1)] 2025-12-04T09:28:45.8896323Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8896449Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8896529Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8896590Z graph_break [] 2025-12-04T09:28:45.8896667Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8896801Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8896863Z frames [('total', 1)] 2025-12-04T09:28:45.8896931Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8897066Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8897140Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8897202Z graph_break [] 2025-12-04T09:28:45.8897290Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8897414Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8897484Z frames [('total', 1)] 2025-12-04T09:28:45.8897556Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8897682Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8897764Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8897825Z graph_break [] 2025-12-04T09:28:45.8897902Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8898033Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8898099Z frames [('total', 1)] 2025-12-04T09:28:45.8898166Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8898300Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8898372Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8898438Z graph_break [] 2025-12-04T09:28:45.8898515Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8898643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8898711Z frames [('total', 1)] 2025-12-04T09:28:45.8898780Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8898909Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8898991Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8899051Z graph_break [] 2025-12-04T09:28:45.8899129Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8899269Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8899332Z frames [('total', 1)] 2025-12-04T09:28:45.8899408Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8899532Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8899607Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8899671Z graph_break [] 2025-12-04T09:28:45.8899746Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8899916Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8899984Z frames [('total', 1)] 2025-12-04T09:28:45.8900052Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8900216Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8900325Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8900386Z graph_break [] 2025-12-04T09:28:45.8900468Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8900594Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8900655Z frames [('total', 1)] 2025-12-04T09:28:45.8900727Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8900851Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8900956Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8901031Z graph_break [] 2025-12-04T09:28:45.8901113Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8901241Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8901311Z frames [('total', 1)] 2025-12-04T09:28:45.8901379Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8901513Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8901587Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8901646Z graph_break [] 2025-12-04T09:28:45.8901728Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8901854Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8901916Z frames [('total', 1)] 2025-12-04T09:28:45.8901989Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8902113Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8902186Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8902252Z graph_break [] 2025-12-04T09:28:45.8902328Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8902459Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8902522Z frames [('total', 1)] 2025-12-04T09:28:45.8902589Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8902725Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8902797Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8902860Z graph_break [] 2025-12-04T09:28:45.8902943Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8903067Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8903129Z frames [('total', 1)] 2025-12-04T09:28:45.8903203Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8903328Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8903401Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8903468Z graph_break [] 2025-12-04T09:28:45.8903544Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8903672Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8903737Z frames [('total', 1)] 2025-12-04T09:28:45.8903807Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8903946Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8904026Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8904087Z graph_break [] 2025-12-04T09:28:45.8904175Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8904299Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8904361Z frames [('total', 1)] 2025-12-04T09:28:45.8904437Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8904769Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8904850Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8904915Z graph_break [] 2025-12-04T09:28:45.8904992Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8905123Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8905274Z frames [('total', 1)] 2025-12-04T09:28:45.8905343Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8905475Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8905600Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8905660Z graph_break [] 2025-12-04T09:28:45.8905795Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8905925Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8905999Z frames [('total', 1)] 2025-12-04T09:28:45.8906067Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8906193Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8906277Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8906402Z graph_break [] 2025-12-04T09:28:45.8906485Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8906615Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8906680Z frames [('total', 1)] 2025-12-04T09:28:45.8906748Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8906881Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8906958Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8907024Z graph_break [] 2025-12-04T09:28:45.8907106Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8907237Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8907307Z frames [('total', 1)] 2025-12-04T09:28:45.8907378Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8907507Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8907584Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8907644Z graph_break [] 2025-12-04T09:28:45.8907728Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8907866Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8907934Z frames [('total', 1)] 2025-12-04T09:28:45.8908007Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8908130Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8908207Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8908271Z graph_break [] 2025-12-04T09:28:45.8908348Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8908471Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8908539Z frames [('total', 1)] 2025-12-04T09:28:45.8908607Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8908729Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8908809Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8908868Z graph_break [] 2025-12-04T09:28:45.8908952Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8909077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8909142Z frames [('total', 1)] 2025-12-04T09:28:45.8909218Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8909346Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8909419Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8909483Z graph_break [] 2025-12-04T09:28:45.8909563Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8909688Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8909768Z frames [('total', 1)] 2025-12-04T09:28:45.8909840Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8909975Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8910047Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8910107Z graph_break [] 2025-12-04T09:28:45.8910191Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8910318Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8910425Z frames [('total', 1)] 2025-12-04T09:28:45.8910500Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8910626Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8910753Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8910820Z graph_break [] 2025-12-04T09:28:45.8910932Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8911066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8911131Z frames [('total', 1)] 2025-12-04T09:28:45.8911199Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8911331Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8911403Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8911461Z graph_break [] 2025-12-04T09:28:45.8911576Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8911715Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8911782Z frames [('total', 1)] 2025-12-04T09:28:45.8911858Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8911986Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8912066Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8912130Z graph_break [] 2025-12-04T09:28:45.8912208Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8912341Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8912404Z frames [('total', 1)] 2025-12-04T09:28:45.8912472Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8912601Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8912673Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8912733Z graph_break [] 2025-12-04T09:28:45.8912817Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8912940Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8913005Z frames [('total', 1)] 2025-12-04T09:28:45.8913086Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8913210Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8913293Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8913356Z graph_break [] 2025-12-04T09:28:45.8913434Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8913569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8913632Z frames [('total', 1)] 2025-12-04T09:28:45.8913702Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8913832Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8913904Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8913964Z graph_break [] 2025-12-04T09:28:45.8914048Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8914174Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8914244Z frames [('total', 1)] 2025-12-04T09:28:45.8914315Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8914440Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8914520Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8914583Z graph_break [] 2025-12-04T09:28:45.8914661Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8914796Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8914858Z frames [('total', 1)] 2025-12-04T09:28:45.8914930Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8915061Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8915134Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8915205Z graph_break [] 2025-12-04T09:28:45.8915292Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8915420Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8915489Z frames [('total', 1)] 2025-12-04T09:28:45.8915608Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8915736Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8915817Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8915913Z graph_break [] 2025-12-04T09:28:45.8915991Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8916156Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8916222Z frames [('total', 1)] 2025-12-04T09:28:45.8916297Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8916425Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8916498Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8916563Z graph_break [] 2025-12-04T09:28:45.8916640Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8916796Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8916865Z frames [('total', 1)] 2025-12-04T09:28:45.8916940Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8917068Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8917147Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8917207Z graph_break [] 2025-12-04T09:28:45.8917290Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8917415Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8917476Z frames [('total', 1)] 2025-12-04T09:28:45.8917549Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8917673Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8917744Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8917808Z graph_break [] 2025-12-04T09:28:45.8917883Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8918006Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8918073Z frames [('total', 1)] 2025-12-04T09:28:45.8918157Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8918292Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8918363Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8918425Z graph_break [] 2025-12-04T09:28:45.8918509Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8918636Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8918698Z frames [('total', 1)] 2025-12-04T09:28:45.8918773Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8918897Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8918972Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8919037Z graph_break [] 2025-12-04T09:28:45.8919113Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8919241Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8919303Z frames [('total', 1)] 2025-12-04T09:28:45.8919372Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8919508Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8919582Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8919642Z graph_break [] 2025-12-04T09:28:45.8919729Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8919855Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8919916Z frames [('total', 1)] 2025-12-04T09:28:45.8919990Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8920115Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8920195Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8920264Z graph_break [] 2025-12-04T09:28:45.8920344Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8920477Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8920540Z frames [('total', 1)] 2025-12-04T09:28:45.8920608Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8920785Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8920857Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8920919Z graph_break [] 2025-12-04T09:28:45.8921002Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8921204Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8921313Z Traceback (most recent call last): 2025-12-04T09:28:45.8921637Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8921696Z f(a, b) 2025-12-04T09:28:45.8921991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8922175Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8922476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8922603Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8922904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8923006Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8923317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8923511Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8923822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8923894Z graph.run(*example_inputs) 2025-12-04T09:28:45.8924126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8924201Z return super().run(*args) 2025-12-04T09:28:45.8924432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8924515Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8924765Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8924842Z result = super().run_node(n) 2025-12-04T09:28:45.8925093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8925195Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8925469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8925617Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8925881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8926022Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8926282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8926357Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8926626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8926706Z return autotune_select_algorithm( 2025-12-04T09:28:45.8927052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8927125Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8927411Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8927577Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8928321Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8928473Z target: aten.mm.default 2025-12-04T09:28:45.8928537Z args[0]: TensorBox( 2025-12-04T09:28:45.8928635Z ReinterpretView( 2025-12-04T09:28:45.8928703Z StorageBox( 2025-12-04T09:28:45.8928947Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8929004Z ), 2025-12-04T09:28:45.8929159Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8929230Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8929326Z stack_traces = {, 2025-12-04T09:28:45.8929559Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8929637Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8929699Z , 2025-12-04T09:28:45.8929757Z } 2025-12-04T09:28:45.8929810Z ) 2025-12-04T09:28:45.8929868Z ) 2025-12-04T09:28:45.8929942Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8930176Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8930234Z )) 2025-12-04T09:28:45.8930237Z 2025-12-04T09:28:45.8930657Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8930661Z 2025-12-04T09:28:45.8930664Z 2025-12-04T09:28:45.8930811Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8931058Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8931061Z 2025-12-04T09:28:45.8931225Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8931364Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8931429Z frames [('total', 1)] 2025-12-04T09:28:45.8931506Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8931577Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8931712Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8931776Z graph_break [] 2025-12-04T09:28:45.8931857Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8931985Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8932055Z frames [('total', 1)] 2025-12-04T09:28:45.8932126Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8932252Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8932330Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8932390Z graph_break [] 2025-12-04T09:28:45.8932476Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8932601Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8932663Z frames [('total', 1)] 2025-12-04T09:28:45.8932739Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8932866Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8932940Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8933004Z graph_break [] 2025-12-04T09:28:45.8933082Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8933287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8933356Z frames [('total', 1)] 2025-12-04T09:28:45.8933424Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8933555Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8933627Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8933687Z graph_break [] 2025-12-04T09:28:45.8933769Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8933962Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8934024Z frames [('total', 1)] 2025-12-04T09:28:45.8934131Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8934257Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8934363Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8934431Z graph_break [] 2025-12-04T09:28:45.8934509Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8934633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8934713Z frames [('total', 1)] 2025-12-04T09:28:45.8934782Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8934913Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8935022Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8935084Z graph_break [] 2025-12-04T09:28:45.8935170Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8935297Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8935359Z frames [('total', 1)] 2025-12-04T09:28:45.8935435Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8935562Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8935637Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8935704Z graph_break [] 2025-12-04T09:28:45.8935781Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8935914Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8935976Z frames [('total', 1)] 2025-12-04T09:28:45.8936049Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8936184Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8936256Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8936315Z graph_break [] 2025-12-04T09:28:45.8936399Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8936527Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8936590Z frames [('total', 1)] 2025-12-04T09:28:45.8936674Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8936805Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8936888Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8936950Z graph_break [] 2025-12-04T09:28:45.8937028Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8937161Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8937223Z frames [('total', 1)] 2025-12-04T09:28:45.8937292Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8937426Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8937501Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8937561Z graph_break [] 2025-12-04T09:28:45.8937648Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8937775Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8937843Z frames [('total', 1)] 2025-12-04T09:28:45.8937913Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8938042Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8938120Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8938183Z graph_break [] 2025-12-04T09:28:45.8938261Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8938394Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8938459Z frames [('total', 1)] 2025-12-04T09:28:45.8938526Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8938657Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8938731Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8938795Z graph_break [] 2025-12-04T09:28:45.8938873Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8939044Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8939111Z frames [('total', 1)] 2025-12-04T09:28:45.8939180Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8939304Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8939416Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8939508Z graph_break [] 2025-12-04T09:28:45.8939587Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8939716Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8939779Z frames [('total', 1)] 2025-12-04T09:28:45.8939852Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8939977Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8940048Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8940153Z graph_break [] 2025-12-04T09:28:45.8940236Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8940359Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8940429Z frames [('total', 1)] 2025-12-04T09:28:45.8940497Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8940622Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8940704Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8940766Z graph_break [] 2025-12-04T09:28:45.8940847Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8940973Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8941035Z frames [('total', 1)] 2025-12-04T09:28:45.8941107Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8941231Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8941305Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8941373Z graph_break [] 2025-12-04T09:28:45.8941450Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8941573Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8941646Z frames [('total', 1)] 2025-12-04T09:28:45.8941717Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8941850Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8941927Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8941990Z graph_break [] 2025-12-04T09:28:45.8942089Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8942220Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8942284Z frames [('total', 1)] 2025-12-04T09:28:45.8942361Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8942488Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8942564Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8942633Z graph_break [] 2025-12-04T09:28:45.8942711Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8942836Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8942907Z frames [('total', 1)] 2025-12-04T09:28:45.8942975Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8943106Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8943183Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8943244Z graph_break [] 2025-12-04T09:28:45.8943328Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8943454Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8943517Z frames [('total', 1)] 2025-12-04T09:28:45.8943596Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8943723Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8943796Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8943863Z graph_break [] 2025-12-04T09:28:45.8943942Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8944072Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8944184Z frames [('total', 1)] 2025-12-04T09:28:45.8944256Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8944398Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8944509Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8944571Z graph_break [] 2025-12-04T09:28:45.8944693Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8944835Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8944905Z frames [('total', 1)] 2025-12-04T09:28:45.8948367Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8948534Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8948621Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8948683Z graph_break [] 2025-12-04T09:28:45.8949059Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8949229Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8949299Z frames [('total', 1)] 2025-12-04T09:28:45.8949371Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8949514Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8949595Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8949657Z graph_break [] 2025-12-04T09:28:45.8949749Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8949882Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8949950Z frames [('total', 1)] 2025-12-04T09:28:45.8950017Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8950144Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8950221Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8950281Z graph_break [] 2025-12-04T09:28:45.8950365Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8950502Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8950567Z frames [('total', 1)] 2025-12-04T09:28:45.8950634Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8950773Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8950847Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8950908Z graph_break [] 2025-12-04T09:28:45.8950989Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8951115Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8951180Z frames [('total', 1)] 2025-12-04T09:28:45.8951247Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8951376Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8951454Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8951519Z graph_break [] 2025-12-04T09:28:45.8951597Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8951729Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8951798Z frames [('total', 1)] 2025-12-04T09:28:45.8951865Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8951993Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8952064Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8952131Z graph_break [] 2025-12-04T09:28:45.8952208Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8952333Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8952400Z frames [('total', 1)] 2025-12-04T09:28:45.8952467Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8952593Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8952668Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8952729Z graph_break [] 2025-12-04T09:28:45.8952806Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8952934Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8952995Z frames [('total', 1)] 2025-12-04T09:28:45.8953115Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8953250Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8953324Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8953421Z graph_break [] 2025-12-04T09:28:45.8953496Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8953652Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8953718Z frames [('total', 1)] 2025-12-04T09:28:45.8953787Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8953911Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8953990Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8954050Z graph_break [] 2025-12-04T09:28:45.8954130Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8954308Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8954369Z frames [('total', 1)] 2025-12-04T09:28:45.8954441Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8954564Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8954634Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8954709Z graph_break [] 2025-12-04T09:28:45.8954784Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8954913Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8954982Z frames [('total', 1)] 2025-12-04T09:28:45.8955048Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8955180Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8955251Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8955310Z graph_break [] 2025-12-04T09:28:45.8955394Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8955527Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8955587Z frames [('total', 1)] 2025-12-04T09:28:45.8955662Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8955791Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8955865Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8955928Z graph_break [] 2025-12-04T09:28:45.8956010Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8956148Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8956208Z frames [('total', 1)] 2025-12-04T09:28:45.8956276Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8956408Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8956479Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8956540Z graph_break [] 2025-12-04T09:28:45.8956623Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8956751Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8956819Z frames [('total', 1)] 2025-12-04T09:28:45.8956895Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8957023Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8957099Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8957155Z graph_break [] 2025-12-04T09:28:45.8957232Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8957364Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8957425Z frames [('total', 1)] 2025-12-04T09:28:45.8957490Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8957616Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8957686Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8957743Z graph_break [] 2025-12-04T09:28:45.8957830Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8957958Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8958029Z frames [('total', 1)] 2025-12-04T09:28:45.8958103Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8958277Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8958355Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8958414Z graph_break [] 2025-12-04T09:28:45.8958525Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8958698Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8958762Z frames [('total', 1)] 2025-12-04T09:28:45.8958830Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8958965Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8959039Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8959111Z graph_break [] 2025-12-04T09:28:45.8959187Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8959347Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8959422Z frames [('total', 1)] 2025-12-04T09:28:45.8959502Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8959632Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8959712Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8959770Z graph_break [] 2025-12-04T09:28:45.8959846Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8959981Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8960045Z frames [('total', 1)] 2025-12-04T09:28:45.8960121Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8960245Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8960318Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8960385Z graph_break [] 2025-12-04T09:28:45.8960460Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8960585Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8960652Z frames [('total', 1)] 2025-12-04T09:28:45.8960722Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8960846Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8960928Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8960987Z graph_break [] 2025-12-04T09:28:45.8961063Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8961197Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8961261Z frames [('total', 1)] 2025-12-04T09:28:45.8961330Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8961461Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8961533Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8961600Z graph_break [] 2025-12-04T09:28:45.8961676Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8961800Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8961868Z frames [('total', 1)] 2025-12-04T09:28:45.8961936Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8962064Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8962153Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8962213Z graph_break [] 2025-12-04T09:28:45.8962289Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8962424Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8962486Z frames [('total', 1)] 2025-12-04T09:28:45.8962561Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8962686Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8962758Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8962822Z graph_break [] 2025-12-04T09:28:45.8962897Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8963021Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8963092Z frames [('total', 1)] 2025-12-04T09:28:45.8963159Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8963283Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8963407Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8963466Z graph_break [] 2025-12-04T09:28:45.8963548Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8963707Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8963805Z frames [('total', 1)] 2025-12-04T09:28:45.8963881Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8964006Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8964078Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8964146Z graph_break [] 2025-12-04T09:28:45.8964221Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8964344Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8964449Z frames [('total', 1)] 2025-12-04T09:28:45.8964518Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8964653Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8964728Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8964786Z graph_break [] 2025-12-04T09:28:45.8964871Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8965040Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.8965115Z Traceback (most recent call last): 2025-12-04T09:28:45.8965456Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8965514Z f(a, b) 2025-12-04T09:28:45.8965821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.8965990Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.8966297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.8966423Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.8966719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.8966811Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.8967132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.8967327Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.8967633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.8967705Z graph.run(*example_inputs) 2025-12-04T09:28:45.8967940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.8968012Z return super().run(*args) 2025-12-04T09:28:45.8968246Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.8968326Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.8968579Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.8968651Z result = super().run_node(n) 2025-12-04T09:28:45.8968901Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.8969002Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.8969269Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.8969421Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.8969682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.8969817Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.8970115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.8970186Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.8970484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.8970595Z return autotune_select_algorithm( 2025-12-04T09:28:45.8970932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.8971011Z return cache(*args, **kwargs) 2025-12-04T09:28:45.8971295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.8971493Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.8972240Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.8972315Z target: aten.mm.default 2025-12-04T09:28:45.8972383Z args[0]: TensorBox( 2025-12-04T09:28:45.8972449Z ReinterpretView( 2025-12-04T09:28:45.8972508Z StorageBox( 2025-12-04T09:28:45.8972752Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.8972808Z ), 2025-12-04T09:28:45.8972961Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.8973032Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.8973190Z stack_traces = {, 2025-12-04T09:28:45.8973429Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.8973506Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.8973565Z , 2025-12-04T09:28:45.8973625Z } 2025-12-04T09:28:45.8973679Z ) 2025-12-04T09:28:45.8973733Z ) 2025-12-04T09:28:45.8973810Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.8974046Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.8974108Z )) 2025-12-04T09:28:45.8974112Z 2025-12-04T09:28:45.8974534Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.8974538Z 2025-12-04T09:28:45.8974541Z 2025-12-04T09:28:45.8974674Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.8974924Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.8974927Z 2025-12-04T09:28:45.8975091Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.8975235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8975310Z frames [('total', 1)] 2025-12-04T09:28:45.8975382Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8975467Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8975610Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8975670Z graph_break [] 2025-12-04T09:28:45.8975759Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8975887Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8975955Z frames [('total', 1)] 2025-12-04T09:28:45.8976023Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8976153Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8976234Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8976293Z graph_break [] 2025-12-04T09:28:45.8976373Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8976557Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8976618Z frames [('total', 1)] 2025-12-04T09:28:45.8976686Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8976854Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8977034Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8977101Z graph_break [] 2025-12-04T09:28:45.8977180Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8977306Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8977373Z frames [('total', 1)] 2025-12-04T09:28:45.8977440Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8977565Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8977677Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8977738Z graph_break [] 2025-12-04T09:28:45.8977820Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8977954Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8978015Z frames [('total', 1)] 2025-12-04T09:28:45.8978090Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8978228Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8978302Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8978372Z graph_break [] 2025-12-04T09:28:45.8978449Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8978573Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8978642Z frames [('total', 1)] 2025-12-04T09:28:45.8978708Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8978834Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8978914Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8978977Z graph_break [] 2025-12-04T09:28:45.8979058Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8979182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8979248Z frames [('total', 1)] 2025-12-04T09:28:45.8979321Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8979446Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8979519Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8979587Z graph_break [] 2025-12-04T09:28:45.8979661Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8979784Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8979849Z frames [('total', 1)] 2025-12-04T09:28:45.8979917Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8980048Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8980120Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8980181Z graph_break [] 2025-12-04T09:28:45.8980259Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8980382Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8980444Z frames [('total', 1)] 2025-12-04T09:28:45.8980522Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8980645Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8980717Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8980784Z graph_break [] 2025-12-04T09:28:45.8980863Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8980991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8981052Z frames [('total', 1)] 2025-12-04T09:28:45.8981123Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8981252Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8981325Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8981385Z graph_break [] 2025-12-04T09:28:45.8981467Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8981589Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8981695Z frames [('total', 1)] 2025-12-04T09:28:45.8981770Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8981893Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8982000Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8982063Z graph_break [] 2025-12-04T09:28:45.8982170Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8982299Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8982363Z frames [('total', 1)] 2025-12-04T09:28:45.8982430Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8982560Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8982630Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8982733Z graph_break [] 2025-12-04T09:28:45.8982820Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8982945Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8983009Z frames [('total', 1)] 2025-12-04T09:28:45.8983083Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8983205Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8983287Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8983346Z graph_break [] 2025-12-04T09:28:45.8983423Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8983555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8983616Z frames [('total', 1)] 2025-12-04T09:28:45.8983683Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8983813Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8983886Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8983945Z graph_break [] 2025-12-04T09:28:45.8984030Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8984153Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8984223Z frames [('total', 1)] 2025-12-04T09:28:45.8984291Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8984414Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8984495Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8984553Z graph_break [] 2025-12-04T09:28:45.8984631Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8984761Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8984822Z frames [('total', 1)] 2025-12-04T09:28:45.8984888Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8985019Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8985091Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8985156Z graph_break [] 2025-12-04T09:28:45.8985236Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8985360Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8985432Z frames [('total', 1)] 2025-12-04T09:28:45.8985500Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8985624Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8985707Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8985766Z graph_break [] 2025-12-04T09:28:45.8985843Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8985984Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8986047Z frames [('total', 1)] 2025-12-04T09:28:45.8986124Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8986250Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8986322Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8986385Z graph_break [] 2025-12-04T09:28:45.8986465Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8986591Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8986705Z frames [('total', 1)] 2025-12-04T09:28:45.8986776Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8986904Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8986982Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8987085Z graph_break [] 2025-12-04T09:28:45.8987169Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8987327Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8987391Z frames [('total', 1)] 2025-12-04T09:28:45.8987467Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8987595Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8987666Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8987731Z graph_break [] 2025-12-04T09:28:45.8987862Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8987993Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8988054Z frames [('total', 1)] 2025-12-04T09:28:45.8988124Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8988253Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8988323Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8988385Z graph_break [] 2025-12-04T09:28:45.8988468Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8988593Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8988654Z frames [('total', 1)] 2025-12-04T09:28:45.8988737Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8988863Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8988940Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8988998Z graph_break [] 2025-12-04T09:28:45.8989077Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8989216Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8989280Z frames [('total', 1)] 2025-12-04T09:28:45.8989353Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8989486Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8989558Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8989619Z graph_break [] 2025-12-04T09:28:45.8989700Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8989822Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8989887Z frames [('total', 1)] 2025-12-04T09:28:45.8989954Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8990078Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8990153Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8990210Z graph_break [] 2025-12-04T09:28:45.8990286Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8990415Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8990476Z frames [('total', 1)] 2025-12-04T09:28:45.8990544Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8990682Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8990753Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8990819Z graph_break [] 2025-12-04T09:28:45.8990894Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8991016Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8991081Z frames [('total', 1)] 2025-12-04T09:28:45.8991147Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8991269Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8991346Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8991404Z graph_break [] 2025-12-04T09:28:45.8991478Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8991609Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8991669Z frames [('total', 1)] 2025-12-04T09:28:45.8991743Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8991915Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8991986Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8992054Z graph_break [] 2025-12-04T09:28:45.8992164Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8992323Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8992391Z frames [('total', 1)] 2025-12-04T09:28:45.8992458Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8992585Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8992661Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8992720Z graph_break [] 2025-12-04T09:28:45.8992795Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8992971Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8993032Z frames [('total', 1)] 2025-12-04T09:28:45.8993103Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8993228Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8993298Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8993362Z graph_break [] 2025-12-04T09:28:45.8993441Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8993567Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8993633Z frames [('total', 1)] 2025-12-04T09:28:45.8993701Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8993825Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8993902Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8993962Z graph_break [] 2025-12-04T09:28:45.8994042Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8994171Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8994233Z frames [('total', 1)] 2025-12-04T09:28:45.8994309Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8994442Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8994516Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8994582Z graph_break [] 2025-12-04T09:28:45.8994666Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8994798Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8994867Z frames [('total', 1)] 2025-12-04T09:28:45.8994936Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8995069Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8995142Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8995201Z graph_break [] 2025-12-04T09:28:45.8995285Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8995413Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8995475Z frames [('total', 1)] 2025-12-04T09:28:45.8995549Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8995674Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8995747Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8995812Z graph_break [] 2025-12-04T09:28:45.8995889Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8996023Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8996089Z frames [('total', 1)] 2025-12-04T09:28:45.8996156Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8996287Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8996360Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8996418Z graph_break [] 2025-12-04T09:28:45.8996503Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8996629Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8996691Z frames [('total', 1)] 2025-12-04T09:28:45.8996776Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8996903Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8997036Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8997095Z graph_break [] 2025-12-04T09:28:45.8997172Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8997356Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8997453Z frames [('total', 1)] 2025-12-04T09:28:45.8997522Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8997651Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8997721Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8997779Z graph_break [] 2025-12-04T09:28:45.8997861Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8997985Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8998089Z frames [('total', 1)] 2025-12-04T09:28:45.8998158Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8998284Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8998362Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8998420Z graph_break [] 2025-12-04T09:28:45.8998493Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8998625Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8998688Z frames [('total', 1)] 2025-12-04T09:28:45.8998753Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8998883Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8998953Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8999019Z graph_break [] 2025-12-04T09:28:45.8999094Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8999216Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8999283Z frames [('total', 1)] 2025-12-04T09:28:45.8999349Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.8999474Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.8999554Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.8999623Z graph_break [] 2025-12-04T09:28:45.8999703Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.8999834Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.8999896Z frames [('total', 1)] 2025-12-04T09:28:45.8999973Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9000096Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9000168Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9000229Z graph_break [] 2025-12-04T09:28:45.9000303Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9000425Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9000492Z frames [('total', 1)] 2025-12-04T09:28:45.9000557Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9000680Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9000757Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9000815Z graph_break [] 2025-12-04T09:28:45.9000892Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9001023Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9001084Z frames [('total', 1)] 2025-12-04T09:28:45.9001157Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9001292Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9001365Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9001428Z graph_break [] 2025-12-04T09:28:45.9001503Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9001624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9001690Z frames [('total', 1)] 2025-12-04T09:28:45.9001759Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9001883Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9002009Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9002066Z graph_break [] 2025-12-04T09:28:45.9002147Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9002271Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9002367Z frames [('total', 1)] 2025-12-04T09:28:45.9002471Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9002598Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9002669Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9002734Z graph_break [] 2025-12-04T09:28:45.9002811Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9002933Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9002999Z frames [('total', 1)] 2025-12-04T09:28:45.9003099Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9003229Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9003302Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9003360Z graph_break [] 2025-12-04T09:28:45.9003456Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9003581Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9003643Z frames [('total', 1)] 2025-12-04T09:28:45.9003718Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9003842Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9003912Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9003974Z graph_break [] 2025-12-04T09:28:45.9004048Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9004176Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9004237Z frames [('total', 1)] 2025-12-04T09:28:45.9004306Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9004437Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9004704Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9004772Z graph_break [] 2025-12-04T09:28:45.9004859Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9004985Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9005048Z frames [('total', 1)] 2025-12-04T09:28:45.9005123Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9005246Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9005323Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9005381Z graph_break [] 2025-12-04T09:28:45.9005454Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9005624Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.9005700Z Traceback (most recent call last): 2025-12-04T09:28:45.9006020Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9006084Z f(a, b) 2025-12-04T09:28:45.9006383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.9006541Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.9006846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.9006966Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.9007264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.9007355Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.9007666Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.9007865Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.9008272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.9008349Z graph.run(*example_inputs) 2025-12-04T09:28:45.9008583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.9008750Z return super().run(*args) 2025-12-04T09:28:45.9008985Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.9009063Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.9009312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.9009387Z result = super().run_node(n) 2025-12-04T09:28:45.9009687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.9009797Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.9010065Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.9010214Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.9010483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.9010615Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.9010878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.9010947Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.9011205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.9011294Z return autotune_select_algorithm( 2025-12-04T09:28:45.9011630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.9011702Z return cache(*args, **kwargs) 2025-12-04T09:28:45.9011991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.9012151Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.9012897Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9012966Z target: aten.mm.default 2025-12-04T09:28:45.9013028Z args[0]: TensorBox( 2025-12-04T09:28:45.9013164Z ReinterpretView( 2025-12-04T09:28:45.9013227Z StorageBox( 2025-12-04T09:28:45.9013474Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9013531Z ), 2025-12-04T09:28:45.9013679Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9013757Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9013836Z stack_traces = {, 2025-12-04T09:28:45.9014068Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9014147Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9014201Z , 2025-12-04T09:28:45.9014257Z } 2025-12-04T09:28:45.9014315Z ) 2025-12-04T09:28:45.9014367Z ) 2025-12-04T09:28:45.9014439Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9014674Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9014728Z )) 2025-12-04T09:28:45.9014732Z 2025-12-04T09:28:45.9015160Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9015209Z 2025-12-04T09:28:45.9015213Z 2025-12-04T09:28:45.9015345Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9015626Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9015631Z 2025-12-04T09:28:45.9015825Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9015957Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9016025Z frames [('total', 1)] 2025-12-04T09:28:45.9016096Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9016169Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9016309Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9016403Z graph_break [] 2025-12-04T09:28:45.9016487Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9016619Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9016686Z frames [('total', 1)] 2025-12-04T09:28:45.9016759Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9016883Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9016960Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9017024Z graph_break [] 2025-12-04T09:28:45.9017106Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9017231Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9017298Z frames [('total', 1)] 2025-12-04T09:28:45.9017365Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9017491Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9017570Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9017631Z graph_break [] 2025-12-04T09:28:45.9017712Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9017834Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9017898Z frames [('total', 1)] 2025-12-04T09:28:45.9017981Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9018111Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9018185Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9018251Z graph_break [] 2025-12-04T09:28:45.9018329Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9018454Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9018523Z frames [('total', 1)] 2025-12-04T09:28:45.9018594Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9018724Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9018796Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9018856Z graph_break [] 2025-12-04T09:28:45.9018942Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9019066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9019130Z frames [('total', 1)] 2025-12-04T09:28:45.9019205Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9019328Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9019403Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9019470Z graph_break [] 2025-12-04T09:28:45.9019548Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9019680Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9019743Z frames [('total', 1)] 2025-12-04T09:28:45.9019812Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9019949Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9020022Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9020079Z graph_break [] 2025-12-04T09:28:45.9020163Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9020296Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9020422Z frames [('total', 1)] 2025-12-04T09:28:45.9020496Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9020622Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9020733Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9020791Z graph_break [] 2025-12-04T09:28:45.9020899Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9021029Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9021092Z frames [('total', 1)] 2025-12-04T09:28:45.9021157Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9021289Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9021361Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9021420Z graph_break [] 2025-12-04T09:28:45.9021547Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9021676Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9021744Z frames [('total', 1)] 2025-12-04T09:28:45.9021811Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9021932Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9022010Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9022067Z graph_break [] 2025-12-04T09:28:45.9022142Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9022269Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9022333Z frames [('total', 1)] 2025-12-04T09:28:45.9022398Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9022526Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9022597Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9022660Z graph_break [] 2025-12-04T09:28:45.9022739Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9022864Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9022931Z frames [('total', 1)] 2025-12-04T09:28:45.9022997Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9023120Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9023196Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9023267Z graph_break [] 2025-12-04T09:28:45.9023346Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9023478Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9023538Z frames [('total', 1)] 2025-12-04T09:28:45.9023606Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9023734Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9023807Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9023873Z graph_break [] 2025-12-04T09:28:45.9023950Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9024075Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9024141Z frames [('total', 1)] 2025-12-04T09:28:45.9024212Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9024336Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9024413Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9024481Z graph_break [] 2025-12-04T09:28:45.9024561Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9024690Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9024751Z frames [('total', 1)] 2025-12-04T09:28:45.9024822Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9024947Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9025019Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9025085Z graph_break [] 2025-12-04T09:28:45.9025160Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9025287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9025355Z frames [('total', 1)] 2025-12-04T09:28:45.9025468Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9025594Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9025670Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9025763Z graph_break [] 2025-12-04T09:28:45.9025843Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9025998Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9026062Z frames [('total', 1)] 2025-12-04T09:28:45.9026133Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9026257Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9026331Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9026396Z graph_break [] 2025-12-04T09:28:45.9026474Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9026629Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9026700Z frames [('total', 1)] 2025-12-04T09:28:45.9026771Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9026912Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9026985Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9027048Z graph_break [] 2025-12-04T09:28:45.9027131Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9027256Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9027318Z frames [('total', 1)] 2025-12-04T09:28:45.9027393Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9027518Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9027589Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9027654Z graph_break [] 2025-12-04T09:28:45.9027730Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9027859Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9027922Z frames [('total', 1)] 2025-12-04T09:28:45.9027994Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9028128Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9028200Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9028258Z graph_break [] 2025-12-04T09:28:45.9028344Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9028469Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9028533Z frames [('total', 1)] 2025-12-04T09:28:45.9028607Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9028730Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9028806Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9028863Z graph_break [] 2025-12-04T09:28:45.9028938Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9029068Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9029129Z frames [('total', 1)] 2025-12-04T09:28:45.9029196Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9029326Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9029396Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9029453Z graph_break [] 2025-12-04T09:28:45.9029535Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9029658Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9029733Z frames [('total', 1)] 2025-12-04T09:28:45.9029802Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9029927Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9030003Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9030061Z graph_break [] 2025-12-04T09:28:45.9030137Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9030267Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9030328Z frames [('total', 1)] 2025-12-04T09:28:45.9030395Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9030569Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9030643Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9030705Z graph_break [] 2025-12-04T09:28:45.9030816Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9030974Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9031041Z frames [('total', 1)] 2025-12-04T09:28:45.9031108Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9031231Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9031306Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9031363Z graph_break [] 2025-12-04T09:28:45.9031441Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9031614Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9031679Z frames [('total', 1)] 2025-12-04T09:28:45.9031747Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9031879Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9031949Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9032012Z graph_break [] 2025-12-04T09:28:45.9032086Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9032213Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9032282Z frames [('total', 1)] 2025-12-04T09:28:45.9032349Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9032473Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9032552Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9032610Z graph_break [] 2025-12-04T09:28:45.9032686Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9032820Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9032882Z frames [('total', 1)] 2025-12-04T09:28:45.9032954Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9033076Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9033154Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9033218Z graph_break [] 2025-12-04T09:28:45.9033293Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9033423Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9033494Z frames [('total', 1)] 2025-12-04T09:28:45.9033562Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9033686Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9033764Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9033820Z graph_break [] 2025-12-04T09:28:45.9033901Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9034029Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9034091Z frames [('total', 1)] 2025-12-04T09:28:45.9034165Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9034291Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9034363Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9034429Z graph_break [] 2025-12-04T09:28:45.9034504Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9034631Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9034703Z frames [('total', 1)] 2025-12-04T09:28:45.9034772Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9034901Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9034984Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9035046Z graph_break [] 2025-12-04T09:28:45.9035130Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9035256Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9035321Z frames [('total', 1)] 2025-12-04T09:28:45.9035397Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9035523Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9035641Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9035708Z graph_break [] 2025-12-04T09:28:45.9035783Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9035948Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9036009Z frames [('total', 1)] 2025-12-04T09:28:45.9036108Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9036244Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9036317Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9036376Z graph_break [] 2025-12-04T09:28:45.9036460Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9036584Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9036709Z frames [('total', 1)] 2025-12-04T09:28:45.9036782Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9036905Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9036984Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9037041Z graph_break [] 2025-12-04T09:28:45.9037120Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9037247Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9037312Z frames [('total', 1)] 2025-12-04T09:28:45.9037380Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9037509Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9037579Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9037635Z graph_break [] 2025-12-04T09:28:45.9037715Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9037836Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9037899Z frames [('total', 1)] 2025-12-04T09:28:45.9037972Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9038105Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9038183Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9038241Z graph_break [] 2025-12-04T09:28:45.9038314Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9038440Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9038502Z frames [('total', 1)] 2025-12-04T09:28:45.9038570Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9038700Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9038772Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9038833Z graph_break [] 2025-12-04T09:28:45.9038907Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9039032Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9039098Z frames [('total', 1)] 2025-12-04T09:28:45.9039166Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9039291Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9039369Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9039426Z graph_break [] 2025-12-04T09:28:45.9039500Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9039626Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9039689Z frames [('total', 1)] 2025-12-04T09:28:45.9039757Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9039886Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9039958Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9040022Z graph_break [] 2025-12-04T09:28:45.9040097Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9040219Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9040286Z frames [('total', 1)] 2025-12-04T09:28:45.9040353Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9040478Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9040554Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9040657Z graph_break [] 2025-12-04T09:28:45.9040733Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9040861Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9040955Z frames [('total', 1)] 2025-12-04T09:28:45.9041040Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9041199Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9041272Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9041336Z graph_break [] 2025-12-04T09:28:45.9041414Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9041537Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9041604Z frames [('total', 1)] 2025-12-04T09:28:45.9041706Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9041832Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9041909Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9041969Z graph_break [] 2025-12-04T09:28:45.9042049Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9042175Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9042238Z frames [('total', 1)] 2025-12-04T09:28:45.9042311Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9042438Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9042521Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9042585Z graph_break [] 2025-12-04T09:28:45.9042662Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9042785Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9042854Z frames [('total', 1)] 2025-12-04T09:28:45.9042921Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9043054Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9043125Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9043187Z graph_break [] 2025-12-04T09:28:45.9043271Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9043395Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9043459Z frames [('total', 1)] 2025-12-04T09:28:45.9043535Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9043661Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9043733Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9043797Z graph_break [] 2025-12-04T09:28:45.9043875Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9044004Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9044066Z frames [('total', 1)] 2025-12-04T09:28:45.9044137Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9044268Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9044340Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9044402Z graph_break [] 2025-12-04T09:28:45.9044485Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9044608Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9044672Z frames [('total', 1)] 2025-12-04T09:28:45.9044746Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9044870Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9044947Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9045006Z graph_break [] 2025-12-04T09:28:45.9045081Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9045221Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9045284Z frames [('total', 1)] 2025-12-04T09:28:45.9045349Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9045483Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9045553Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9045614Z graph_break [] 2025-12-04T09:28:45.9045744Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9045867Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9045931Z frames [('total', 1)] 2025-12-04T09:28:45.9046033Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9046190Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9046267Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9046324Z graph_break [] 2025-12-04T09:28:45.9046401Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9046570Z _____________ TestKernelBenchmark.test_mm_triton_kernel_benchmark ______________ 2025-12-04T09:28:45.9046646Z Traceback (most recent call last): 2025-12-04T09:28:45.9046993Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 188, in test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9047067Z f(a, b) 2025-12-04T09:28:45.9047364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 940, in compile_wrapper 2025-12-04T09:28:45.9047520Z raise e.remove_dynamo_frames() from None # see TORCHDYNAMO_VERBOSE=1 2025-12-04T09:28:45.9047818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1019, in _compile_fx_inner 2025-12-04T09:28:45.9047939Z raise InductorError(e, currentframe()).with_traceback( 2025-12-04T09:28:45.9048239Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1003, in _compile_fx_inner 2025-12-04T09:28:45.9048328Z mb_compiled_graph = fx_codegen_and_compile( 2025-12-04T09:28:45.9048642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1757, in fx_codegen_and_compile 2025-12-04T09:28:45.9048837Z return scheme.codegen_and_compile(gm, example_inputs, inputs_to_check, graph_kwargs) 2025-12-04T09:28:45.9049138Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py", line 1452, in codegen_and_compile 2025-12-04T09:28:45.9049218Z graph.run(*example_inputs) 2025-12-04T09:28:45.9049449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 987, in run 2025-12-04T09:28:45.9049524Z return super().run(*args) 2025-12-04T09:28:45.9049760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 200, in run 2025-12-04T09:28:45.9049837Z self.env[node] = self.run_node(node) 2025-12-04T09:28:45.9050095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1726, in run_node 2025-12-04T09:28:45.9050171Z result = super().run_node(n) 2025-12-04T09:28:45.9050419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/interpreter.py", line 295, in run_node 2025-12-04T09:28:45.9050528Z return getattr(self, n.op)(n.target, args, kwargs) 2025-12-04T09:28:45.9050797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1369, in call_function 2025-12-04T09:28:45.9050944Z raise LoweringException(e, target, args, kwargs).with_traceback( 2025-12-04T09:28:45.9051211Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/graph.py", line 1359, in call_function 2025-12-04T09:28:45.9051345Z out = lowerings[target](*args, **kwargs) # type: ignore[index] 2025-12-04T09:28:45.9051606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py", line 495, in wrapped 2025-12-04T09:28:45.9051677Z out = decomp_fn(*args, **kwargs) 2025-12-04T09:28:45.9051934Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/kernel/mm.py", line 525, in tuned_mm 2025-12-04T09:28:45.9052019Z return autotune_select_algorithm( 2025-12-04T09:28:45.9052357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 4298, in autotune_select_algorithm 2025-12-04T09:28:45.9052435Z return cache(*args, **kwargs) 2025-12-04T09:28:45.9052765Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2777, in __call__ 2025-12-04T09:28:45.9052924Z raise self.create_no_valid_choices(name, "No choices exist for backend.") 2025-12-04T09:28:45.9053858Z torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9053931Z target: aten.mm.default 2025-12-04T09:28:45.9053998Z args[0]: TensorBox( 2025-12-04T09:28:45.9054074Z ReinterpretView( 2025-12-04T09:28:45.9054134Z StorageBox( 2025-12-04T09:28:45.9054416Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9054473Z ), 2025-12-04T09:28:45.9054619Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9054699Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9054773Z stack_traces = {, 2025-12-04T09:28:45.9055000Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9055087Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9055141Z , 2025-12-04T09:28:45.9055195Z } 2025-12-04T09:28:45.9055255Z ) 2025-12-04T09:28:45.9055309Z ) 2025-12-04T09:28:45.9055388Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9055618Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9055672Z )) 2025-12-04T09:28:45.9055675Z 2025-12-04T09:28:45.9056100Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9056106Z 2025-12-04T09:28:45.9056109Z 2025-12-04T09:28:45.9056243Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9056490Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9056496Z 2025-12-04T09:28:45.9056659Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9056789Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9056857Z frames [('total', 1)] 2025-12-04T09:28:45.9056928Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9057002Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9057142Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9057204Z graph_break [] 2025-12-04T09:28:45.9057293Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9057421Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9057484Z frames [('total', 1)] 2025-12-04T09:28:45.9057557Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9057686Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9057758Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9057825Z graph_break [] 2025-12-04T09:28:45.9057905Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9058033Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9058100Z frames [('total', 1)] 2025-12-04T09:28:45.9058168Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9058299Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9058373Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9058432Z graph_break [] 2025-12-04T09:28:45.9058527Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9058653Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9058763Z frames [('total', 1)] 2025-12-04T09:28:45.9058835Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9058958Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9059028Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9059127Z graph_break [] 2025-12-04T09:28:45.9059204Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9059385Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9059449Z frames [('total', 1)] 2025-12-04T09:28:45.9059515Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9059648Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9059718Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9059776Z graph_break [] 2025-12-04T09:28:45.9059893Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9060018Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9060078Z frames [('total', 1)] 2025-12-04T09:28:45.9060153Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9060275Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9060351Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9060410Z graph_break [] 2025-12-04T09:28:45.9060486Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9060617Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9060677Z frames [('total', 1)] 2025-12-04T09:28:45.9060754Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9060888Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9060959Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9061018Z graph_break [] 2025-12-04T09:28:45.9061099Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9061223Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9061294Z frames [('total', 1)] 2025-12-04T09:28:45.9061364Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9061487Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9061563Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9061623Z graph_break [] 2025-12-04T09:28:45.9061699Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9061830Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9061890Z frames [('total', 1)] 2025-12-04T09:28:45.9061956Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9062084Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9062155Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9062219Z graph_break [] 2025-12-04T09:28:45.9062299Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9062423Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9062489Z frames [('total', 1)] 2025-12-04T09:28:45.9062556Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9062682Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9062760Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9062822Z graph_break [] 2025-12-04T09:28:45.9062901Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9063036Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9063101Z frames [('total', 1)] 2025-12-04T09:28:45.9063167Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9063299Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9063371Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9063434Z graph_break [] 2025-12-04T09:28:45.9063509Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9063634Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9063699Z frames [('total', 1)] 2025-12-04T09:28:45.9063767Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9063938Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9064030Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9064090Z graph_break [] 2025-12-04T09:28:45.9064248Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9064409Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9064472Z frames [('total', 1)] 2025-12-04T09:28:45.9064548Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9064672Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9064743Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9064809Z graph_break [] 2025-12-04T09:28:45.9064886Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9065044Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9065115Z frames [('total', 1)] 2025-12-04T09:28:45.9065181Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9065308Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9065385Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9065443Z graph_break [] 2025-12-04T09:28:45.9065530Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9065660Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9065722Z frames [('total', 1)] 2025-12-04T09:28:45.9065796Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9065919Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9065990Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9066055Z graph_break [] 2025-12-04T09:28:45.9066131Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9066257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9066325Z frames [('total', 1)] 2025-12-04T09:28:45.9066392Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9066526Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9066598Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9066655Z graph_break [] 2025-12-04T09:28:45.9066737Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9066865Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9066927Z frames [('total', 1)] 2025-12-04T09:28:45.9067001Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9067136Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9067210Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9067272Z graph_break [] 2025-12-04T09:28:45.9067346Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9067475Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9067536Z frames [('total', 1)] 2025-12-04T09:28:45.9067603Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9067733Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9067808Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9067866Z graph_break [] 2025-12-04T09:28:45.9067947Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9068076Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9068139Z frames [('total', 1)] 2025-12-04T09:28:45.9068211Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9068334Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9068409Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9068478Z graph_break [] 2025-12-04T09:28:45.9068556Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9068687Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9068749Z frames [('total', 1)] 2025-12-04T09:28:45.9068815Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9068943Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9069057Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9069114Z graph_break [] 2025-12-04T09:28:45.9069196Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9069353Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9069453Z frames [('total', 1)] 2025-12-04T09:28:45.9069520Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9069644Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9069720Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9069777Z graph_break [] 2025-12-04T09:28:45.9069853Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9069982Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9070076Z frames [('total', 1)] 2025-12-04T09:28:45.9070144Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9070276Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9070350Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9070417Z graph_break [] 2025-12-04T09:28:45.9070500Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9070627Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9070691Z frames [('total', 1)] 2025-12-04T09:28:45.9070763Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9070887Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9070964Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9071021Z graph_break [] 2025-12-04T09:28:45.9071097Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9071224Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9071288Z frames [('total', 1)] 2025-12-04T09:28:45.9071355Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9071484Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9071557Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9071621Z graph_break [] 2025-12-04T09:28:45.9071697Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9071824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9071891Z frames [('total', 1)] 2025-12-04T09:28:45.9071961Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9072086Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9072163Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9072220Z graph_break [] 2025-12-04T09:28:45.9072306Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9072437Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9072499Z frames [('total', 1)] 2025-12-04T09:28:45.9072573Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9072696Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9072770Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9072833Z graph_break [] 2025-12-04T09:28:45.9072909Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9073032Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9073103Z frames [('total', 1)] 2025-12-04T09:28:45.9073172Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9073299Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9073377Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9073434Z graph_break [] 2025-12-04T09:28:45.9073520Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9073643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9073703Z frames [('total', 1)] 2025-12-04T09:28:45.9073776Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9073899Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9074012Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9074086Z graph_break [] 2025-12-04T09:28:45.9074167Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9074292Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9074396Z frames [('total', 1)] 2025-12-04T09:28:45.9074495Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9074627Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9074699Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9074758Z graph_break [] 2025-12-04T09:28:45.9074846Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9074969Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9075031Z frames [('total', 1)] 2025-12-04T09:28:45.9075146Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9075271Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9075344Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9075408Z graph_break [] 2025-12-04T09:28:45.9075482Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9075611Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9075673Z frames [('total', 1)] 2025-12-04T09:28:45.9075740Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9075874Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9075947Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9076005Z graph_break [] 2025-12-04T09:28:45.9076084Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9076205Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9076266Z frames [('total', 1)] 2025-12-04T09:28:45.9076340Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9076463Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9076538Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9076601Z graph_break [] 2025-12-04T09:28:45.9076678Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9076805Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9076868Z frames [('total', 1)] 2025-12-04T09:28:45.9076934Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9077077Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9077149Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9077209Z graph_break [] 2025-12-04T09:28:45.9077290Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9077412Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9077479Z frames [('total', 1)] 2025-12-04T09:28:45.9077546Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9077676Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9077753Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9077814Z graph_break [] 2025-12-04T09:28:45.9077889Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9078018Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9078082Z frames [('total', 1)] 2025-12-04T09:28:45.9078157Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9078292Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9078362Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9078424Z graph_break [] 2025-12-04T09:28:45.9078500Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9078623Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9078690Z frames [('total', 1)] 2025-12-04T09:28:45.9078757Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9078881Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9078957Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9079059Z graph_break [] 2025-12-04T09:28:45.9079135Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9079263Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9079377Z frames [('total', 1)] 2025-12-04T09:28:45.9079443Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9079606Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9079677Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9079740Z graph_break [] 2025-12-04T09:28:45.9079816Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9079941Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9080007Z frames [('total', 1)] 2025-12-04T09:28:45.9080073Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9080232Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9080311Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9080373Z graph_break [] 2025-12-04T09:28:45.9080449Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9080579Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9080652Z frames [('total', 1)] 2025-12-04T09:28:45.9080725Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9080850Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9080920Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9080984Z graph_break [] 2025-12-04T09:28:45.9081060Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9081182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9081248Z frames [('total', 1)] 2025-12-04T09:28:45.9081314Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9081440Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9081520Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9081577Z graph_break [] 2025-12-04T09:28:45.9081661Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9081784Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9081844Z frames [('total', 1)] 2025-12-04T09:28:45.9081922Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9082048Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9082120Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9082188Z graph_break [] 2025-12-04T09:28:45.9082262Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9082395Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9082466Z frames [('total', 1)] 2025-12-04T09:28:45.9082536Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9082671Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9082741Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9082800Z graph_break [] 2025-12-04T09:28:45.9082885Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9083008Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9083069Z frames [('total', 1)] 2025-12-04T09:28:45.9083145Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9083271Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9083346Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9083411Z graph_break [] 2025-12-04T09:28:45.9083488Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9083618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9083680Z frames [('total', 1)] 2025-12-04T09:28:45.9083745Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9083874Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9083944Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9084002Z graph_break [] 2025-12-04T09:28:45.9084132Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9084259Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9084319Z frames [('total', 1)] 2025-12-04T09:28:45.9084424Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9084546Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9084652Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9084711Z graph_break [] 2025-12-04T09:28:45.9084787Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9084916Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9084976Z frames [('total', 1)] 2025-12-04T09:28:45.9085043Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9085203Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9085276Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9085334Z graph_break [] 2025-12-04T09:28:45.9085414Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9085538Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9085602Z frames [('total', 1)] 2025-12-04T09:28:45.9085667Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9085791Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9085870Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9085930Z graph_break [] 2025-12-04T09:28:45.9086005Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9086132Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9086193Z frames [('total', 1)] 2025-12-04T09:28:45.9086259Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9086389Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9086460Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9086523Z graph_break [] 2025-12-04T09:28:45.9086598Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9086723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9086791Z frames [('total', 1)] 2025-12-04T09:28:45.9086858Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9086984Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9087063Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9087129Z graph_break [] 2025-12-04T09:28:45.9087217Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9087348Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:28:45.9087408Z frames [('total', 1)] 2025-12-04T09:28:45.9087474Z stats [('calls_captured', 2)] 2025-12-04T09:28:45.9087606Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T09:28:45.9087679Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:28:45.9087745Z graph_break [] 2025-12-04T09:28:45.9087823Z aten_mm_info [('aten.mm_2048_2432_1949', 1)] 2025-12-04T09:28:45.9088298Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_kernel_benchmark/inductor.test_kernel_benchmark-8244c39f5fb36d0d.xml - 2025-12-04T09:28:45.9088409Z =========================== short test summary info ============================ 2025-12-04T09:28:45.9089487Z FAILED [1.2294s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9089563Z target: aten.mm.default 2025-12-04T09:28:45.9089626Z args[0]: TensorBox( 2025-12-04T09:28:45.9089692Z ReinterpretView( 2025-12-04T09:28:45.9089757Z StorageBox( 2025-12-04T09:28:45.9089999Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9090097Z ), 2025-12-04T09:28:45.9090251Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9090321Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9090423Z stack_traces = {, 2025-12-04T09:28:45.9090678Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9090755Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9090815Z , 2025-12-04T09:28:45.9090868Z } 2025-12-04T09:28:45.9090921Z ) 2025-12-04T09:28:45.9090978Z ) 2025-12-04T09:28:45.9091048Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9091280Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9091385Z )) 2025-12-04T09:28:45.9091390Z 2025-12-04T09:28:45.9091810Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9091815Z 2025-12-04T09:28:45.9091818Z 2025-12-04T09:28:45.9091957Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9092202Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9092206Z 2025-12-04T09:28:45.9092368Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9093493Z FAILED [0.0606s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9093564Z target: aten.mm.default 2025-12-04T09:28:45.9093631Z args[0]: TensorBox( 2025-12-04T09:28:45.9093698Z ReinterpretView( 2025-12-04T09:28:45.9093757Z StorageBox( 2025-12-04T09:28:45.9094001Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9094058Z ), 2025-12-04T09:28:45.9094211Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9094283Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9094346Z stack_traces = {, 2025-12-04T09:28:45.9094574Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9094647Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9094701Z , 2025-12-04T09:28:45.9094763Z } 2025-12-04T09:28:45.9094821Z ) 2025-12-04T09:28:45.9094878Z ) 2025-12-04T09:28:45.9094956Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9095184Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9095245Z )) 2025-12-04T09:28:45.9095249Z 2025-12-04T09:28:45.9095663Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9095668Z 2025-12-04T09:28:45.9095671Z 2025-12-04T09:28:45.9095801Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9096047Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9096051Z 2025-12-04T09:28:45.9096208Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9097261Z FAILED [0.0606s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9097376Z target: aten.mm.default 2025-12-04T09:28:45.9097437Z args[0]: TensorBox( 2025-12-04T09:28:45.9097539Z ReinterpretView( 2025-12-04T09:28:45.9097597Z StorageBox( 2025-12-04T09:28:45.9097873Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9097928Z ), 2025-12-04T09:28:45.9098074Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9098148Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9098210Z stack_traces = {, 2025-12-04T09:28:45.9098459Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9098540Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9098593Z , 2025-12-04T09:28:45.9098645Z } 2025-12-04T09:28:45.9098704Z ) 2025-12-04T09:28:45.9098757Z ) 2025-12-04T09:28:45.9098829Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9099058Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9099116Z )) 2025-12-04T09:28:45.9099119Z 2025-12-04T09:28:45.9099539Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9099544Z 2025-12-04T09:28:45.9099547Z 2025-12-04T09:28:45.9099673Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9099916Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9099920Z 2025-12-04T09:28:45.9100077Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9101109Z FAILED [0.0887s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9101189Z target: aten.mm.default 2025-12-04T09:28:45.9101249Z args[0]: TensorBox( 2025-12-04T09:28:45.9101320Z ReinterpretView( 2025-12-04T09:28:45.9101385Z StorageBox( 2025-12-04T09:28:45.9101618Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9101678Z ), 2025-12-04T09:28:45.9101822Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9101897Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9101965Z stack_traces = {, 2025-12-04T09:28:45.9102180Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9102259Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9102313Z , 2025-12-04T09:28:45.9102366Z } 2025-12-04T09:28:45.9102427Z ) 2025-12-04T09:28:45.9102479Z ) 2025-12-04T09:28:45.9102551Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9102784Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9102837Z )) 2025-12-04T09:28:45.9102840Z 2025-12-04T09:28:45.9103251Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9103260Z 2025-12-04T09:28:45.9103263Z 2025-12-04T09:28:45.9103390Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9103627Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9103695Z 2025-12-04T09:28:45.9103862Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9105937Z FAILED [0.0859s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9106121Z target: aten.mm.default 2025-12-04T09:28:45.9106187Z args[0]: TensorBox( 2025-12-04T09:28:45.9106254Z ReinterpretView( 2025-12-04T09:28:45.9106319Z StorageBox( 2025-12-04T09:28:45.9106630Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9106690Z ), 2025-12-04T09:28:45.9106848Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9106926Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9107003Z stack_traces = {, 2025-12-04T09:28:45.9107239Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9107330Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9107395Z , 2025-12-04T09:28:45.9107449Z } 2025-12-04T09:28:45.9107504Z ) 2025-12-04T09:28:45.9107562Z ) 2025-12-04T09:28:45.9107637Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9107878Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9107945Z )) 2025-12-04T09:28:45.9107949Z 2025-12-04T09:28:45.9111730Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9111741Z 2025-12-04T09:28:45.9111747Z 2025-12-04T09:28:45.9111936Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9112194Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9112202Z 2025-12-04T09:28:45.9112378Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9113443Z FAILED [0.0613s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9113518Z target: aten.mm.default 2025-12-04T09:28:45.9113589Z args[0]: TensorBox( 2025-12-04T09:28:45.9113659Z ReinterpretView( 2025-12-04T09:28:45.9113719Z StorageBox( 2025-12-04T09:28:45.9113981Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9114039Z ), 2025-12-04T09:28:45.9114202Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9114279Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9114344Z stack_traces = {, 2025-12-04T09:28:45.9114587Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9114671Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9114728Z , 2025-12-04T09:28:45.9114788Z } 2025-12-04T09:28:45.9114843Z ) 2025-12-04T09:28:45.9114897Z ) 2025-12-04T09:28:45.9114978Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9115220Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9115274Z )) 2025-12-04T09:28:45.9115283Z 2025-12-04T09:28:45.9115711Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9115821Z 2025-12-04T09:28:45.9115859Z 2025-12-04T09:28:45.9116005Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9116289Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9116294Z 2025-12-04T09:28:45.9116459Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9117548Z FAILED [0.0625s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9117622Z target: aten.mm.default 2025-12-04T09:28:45.9117685Z args[0]: TensorBox( 2025-12-04T09:28:45.9117757Z ReinterpretView( 2025-12-04T09:28:45.9117816Z StorageBox( 2025-12-04T09:28:45.9118061Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9118121Z ), 2025-12-04T09:28:45.9118278Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9118356Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9118418Z stack_traces = {, 2025-12-04T09:28:45.9118643Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9118725Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9118781Z , 2025-12-04T09:28:45.9118839Z } 2025-12-04T09:28:45.9118895Z ) 2025-12-04T09:28:45.9118948Z ) 2025-12-04T09:28:45.9119020Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9119257Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9119309Z )) 2025-12-04T09:28:45.9119312Z 2025-12-04T09:28:45.9119734Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9119740Z 2025-12-04T09:28:45.9119743Z 2025-12-04T09:28:45.9119873Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9120122Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9120126Z 2025-12-04T09:28:45.9120287Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9121332Z FAILED [0.0604s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9121408Z target: aten.mm.default 2025-12-04T09:28:45.9121470Z args[0]: TensorBox( 2025-12-04T09:28:45.9121537Z ReinterpretView( 2025-12-04T09:28:45.9121596Z StorageBox( 2025-12-04T09:28:45.9121829Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9121890Z ), 2025-12-04T09:28:45.9122038Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9122111Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9122180Z stack_traces = {, 2025-12-04T09:28:45.9122404Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9122480Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9122590Z , 2025-12-04T09:28:45.9122644Z } 2025-12-04T09:28:45.9122701Z ) 2025-12-04T09:28:45.9122753Z ) 2025-12-04T09:28:45.9122828Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9123098Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9123151Z )) 2025-12-04T09:28:45.9123188Z 2025-12-04T09:28:45.9123609Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9123613Z 2025-12-04T09:28:45.9123620Z 2025-12-04T09:28:45.9123750Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9124043Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9124048Z 2025-12-04T09:28:45.9124214Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9125264Z FAILED [0.0600s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9125341Z target: aten.mm.default 2025-12-04T09:28:45.9125403Z args[0]: TensorBox( 2025-12-04T09:28:45.9125466Z ReinterpretView( 2025-12-04T09:28:45.9125529Z StorageBox( 2025-12-04T09:28:45.9125762Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9125816Z ), 2025-12-04T09:28:45.9125969Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9126041Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9126102Z stack_traces = {, 2025-12-04T09:28:45.9126329Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9126401Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9126461Z , 2025-12-04T09:28:45.9126525Z } 2025-12-04T09:28:45.9126580Z ) 2025-12-04T09:28:45.9126638Z ) 2025-12-04T09:28:45.9126710Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9126940Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9126998Z )) 2025-12-04T09:28:45.9127001Z 2025-12-04T09:28:45.9127416Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9127420Z 2025-12-04T09:28:45.9127426Z 2025-12-04T09:28:45.9127557Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9127790Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9127796Z 2025-12-04T09:28:45.9127951Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9128988Z FAILED [0.0614s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9129055Z target: aten.mm.default 2025-12-04T09:28:45.9129127Z args[0]: TensorBox( 2025-12-04T09:28:45.9129190Z ReinterpretView( 2025-12-04T09:28:45.9129247Z StorageBox( 2025-12-04T09:28:45.9129488Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9129545Z ), 2025-12-04T09:28:45.9129733Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9129816Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9129883Z stack_traces = {, 2025-12-04T09:28:45.9130141Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9130247Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9130304Z , 2025-12-04T09:28:45.9130366Z } 2025-12-04T09:28:45.9130422Z ) 2025-12-04T09:28:45.9130476Z ) 2025-12-04T09:28:45.9130555Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9130786Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9130839Z )) 2025-12-04T09:28:45.9130843Z 2025-12-04T09:28:45.9131308Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9131314Z 2025-12-04T09:28:45.9131317Z 2025-12-04T09:28:45.9131445Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9131688Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9131694Z 2025-12-04T09:28:45.9131852Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9132892Z FAILED [0.0619s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9132963Z target: aten.mm.default 2025-12-04T09:28:45.9133026Z args[0]: TensorBox( 2025-12-04T09:28:45.9133238Z ReinterpretView( 2025-12-04T09:28:45.9133320Z StorageBox( 2025-12-04T09:28:45.9133565Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9133621Z ), 2025-12-04T09:28:45.9133771Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9133854Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9133917Z stack_traces = {, 2025-12-04T09:28:45.9134136Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9134216Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9134271Z , 2025-12-04T09:28:45.9134325Z } 2025-12-04T09:28:45.9134384Z ) 2025-12-04T09:28:45.9134441Z ) 2025-12-04T09:28:45.9134511Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9134747Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9134804Z )) 2025-12-04T09:28:45.9134807Z 2025-12-04T09:28:45.9135226Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9135231Z 2025-12-04T09:28:45.9135234Z 2025-12-04T09:28:45.9135362Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9135600Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9135603Z 2025-12-04T09:28:45.9135766Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9136806Z FAILED [0.0593s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9136935Z target: aten.mm.default 2025-12-04T09:28:45.9136996Z args[0]: TensorBox( 2025-12-04T09:28:45.9137059Z ReinterpretView( 2025-12-04T09:28:45.9137164Z StorageBox( 2025-12-04T09:28:45.9137429Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9137491Z ), 2025-12-04T09:28:45.9137637Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9137709Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9137777Z stack_traces = {, 2025-12-04T09:28:45.9137993Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9138100Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9138162Z , 2025-12-04T09:28:45.9138216Z } 2025-12-04T09:28:45.9138271Z ) 2025-12-04T09:28:45.9138333Z ) 2025-12-04T09:28:45.9138405Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9138635Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9138695Z )) 2025-12-04T09:28:45.9138701Z 2025-12-04T09:28:45.9139114Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9139118Z 2025-12-04T09:28:45.9139121Z 2025-12-04T09:28:45.9139257Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9139494Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9139497Z 2025-12-04T09:28:45.9139662Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9140695Z FAILED [0.0589s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9140766Z target: aten.mm.default 2025-12-04T09:28:45.9140837Z args[0]: TensorBox( 2025-12-04T09:28:45.9140903Z ReinterpretView( 2025-12-04T09:28:45.9140971Z StorageBox( 2025-12-04T09:28:45.9141200Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9141268Z ), 2025-12-04T09:28:45.9141423Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9141498Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9141563Z stack_traces = {, 2025-12-04T09:28:45.9141785Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9141861Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9141916Z , 2025-12-04T09:28:45.9141980Z } 2025-12-04T09:28:45.9142033Z ) 2025-12-04T09:28:45.9142086Z ) 2025-12-04T09:28:45.9142163Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9142395Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9142455Z )) 2025-12-04T09:28:45.9142458Z 2025-12-04T09:28:45.9142870Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9142874Z 2025-12-04T09:28:45.9142877Z 2025-12-04T09:28:45.9143000Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9143242Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9143246Z 2025-12-04T09:28:45.9143456Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9144532Z FAILED [0.0591s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9144633Z target: aten.mm.default 2025-12-04T09:28:45.9144700Z args[0]: TensorBox( 2025-12-04T09:28:45.9144764Z ReinterpretView( 2025-12-04T09:28:45.9144822Z StorageBox( 2025-12-04T09:28:45.9145063Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9145151Z ), 2025-12-04T09:28:45.9145301Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9145378Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9145443Z stack_traces = {, 2025-12-04T09:28:45.9145668Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9145748Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9145802Z , 2025-12-04T09:28:45.9145861Z } 2025-12-04T09:28:45.9145915Z ) 2025-12-04T09:28:45.9145967Z ) 2025-12-04T09:28:45.9146043Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9146269Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9146323Z )) 2025-12-04T09:28:45.9146326Z 2025-12-04T09:28:45.9146750Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9146754Z 2025-12-04T09:28:45.9146757Z 2025-12-04T09:28:45.9146886Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9147133Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9147137Z 2025-12-04T09:28:45.9147297Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9148347Z FAILED [0.0594s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9148415Z target: aten.mm.default 2025-12-04T09:28:45.9148477Z args[0]: TensorBox( 2025-12-04T09:28:45.9148547Z ReinterpretView( 2025-12-04T09:28:45.9148606Z StorageBox( 2025-12-04T09:28:45.9148838Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9148900Z ), 2025-12-04T09:28:45.9149045Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9149118Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9149188Z stack_traces = {, 2025-12-04T09:28:45.9149407Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9149485Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9149540Z , 2025-12-04T09:28:45.9149594Z } 2025-12-04T09:28:45.9149653Z ) 2025-12-04T09:28:45.9149706Z ) 2025-12-04T09:28:45.9149778Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9150011Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9150066Z )) 2025-12-04T09:28:45.9150069Z 2025-12-04T09:28:45.9150483Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9150539Z 2025-12-04T09:28:45.9150542Z 2025-12-04T09:28:45.9150672Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9150943Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9150996Z 2025-12-04T09:28:45.9151159Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9152221Z FAILED [0.0598s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9152296Z target: aten.mm.default 2025-12-04T09:28:45.9152358Z args[0]: TensorBox( 2025-12-04T09:28:45.9152427Z ReinterpretView( 2025-12-04T09:28:45.9152492Z StorageBox( 2025-12-04T09:28:45.9152722Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9152778Z ), 2025-12-04T09:28:45.9152931Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9153002Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9153073Z stack_traces = {, 2025-12-04T09:28:45.9153290Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9153364Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9153425Z , 2025-12-04T09:28:45.9153479Z } 2025-12-04T09:28:45.9153532Z ) 2025-12-04T09:28:45.9153591Z ) 2025-12-04T09:28:45.9153665Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9153896Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9153957Z )) 2025-12-04T09:28:45.9153961Z 2025-12-04T09:28:45.9154371Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9154377Z 2025-12-04T09:28:45.9154380Z 2025-12-04T09:28:45.9154514Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9154750Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9154754Z 2025-12-04T09:28:45.9154920Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9155954Z FAILED [0.0663s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9156024Z target: aten.mm.default 2025-12-04T09:28:45.9156092Z args[0]: TensorBox( 2025-12-04T09:28:45.9156159Z ReinterpretView( 2025-12-04T09:28:45.9156218Z StorageBox( 2025-12-04T09:28:45.9156454Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9156508Z ), 2025-12-04T09:28:45.9156662Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9156736Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9156800Z stack_traces = {, 2025-12-04T09:28:45.9157024Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9157109Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9157169Z , 2025-12-04T09:28:45.9157230Z } 2025-12-04T09:28:45.9157334Z ) 2025-12-04T09:28:45.9157387Z ) 2025-12-04T09:28:45.9157474Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9157712Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9157808Z )) 2025-12-04T09:28:45.9157811Z 2025-12-04T09:28:45.9158275Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9158279Z 2025-12-04T09:28:45.9158282Z 2025-12-04T09:28:45.9158426Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9158685Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9158688Z 2025-12-04T09:28:45.9158888Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9159941Z FAILED [0.0628s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9160020Z target: aten.mm.default 2025-12-04T09:28:45.9160088Z args[0]: TensorBox( 2025-12-04T09:28:45.9160153Z ReinterpretView( 2025-12-04T09:28:45.9160212Z StorageBox( 2025-12-04T09:28:45.9160466Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9160523Z ), 2025-12-04T09:28:45.9160676Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9160760Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9160823Z stack_traces = {, 2025-12-04T09:28:45.9161050Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9161136Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9161194Z , 2025-12-04T09:28:45.9161252Z } 2025-12-04T09:28:45.9161308Z ) 2025-12-04T09:28:45.9161361Z ) 2025-12-04T09:28:45.9161438Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9161674Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9161728Z )) 2025-12-04T09:28:45.9161731Z 2025-12-04T09:28:45.9162155Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9162158Z 2025-12-04T09:28:45.9162161Z 2025-12-04T09:28:45.9162297Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9162544Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9162550Z 2025-12-04T09:28:45.9162711Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9163750Z FAILED [0.0617s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9163827Z target: aten.mm.default 2025-12-04T09:28:45.9163891Z args[0]: TensorBox( 2025-12-04T09:28:45.9163961Z ReinterpretView( 2025-12-04T09:28:45.9164020Z StorageBox( 2025-12-04T09:28:45.9164257Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9164319Z ), 2025-12-04T09:28:45.9164469Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9164590Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9164659Z stack_traces = {, 2025-12-04T09:28:45.9164884Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9165000Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9165088Z , 2025-12-04T09:28:45.9165142Z } 2025-12-04T09:28:45.9165201Z ) 2025-12-04T09:28:45.9165254Z ) 2025-12-04T09:28:45.9165326Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9165562Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9165616Z )) 2025-12-04T09:28:45.9165619Z 2025-12-04T09:28:45.9166069Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9166079Z 2025-12-04T09:28:45.9166086Z 2025-12-04T09:28:45.9166219Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9166460Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9166465Z 2025-12-04T09:28:45.9166632Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9167668Z FAILED [0.0599s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9167745Z target: aten.mm.default 2025-12-04T09:28:45.9167809Z args[0]: TensorBox( 2025-12-04T09:28:45.9167873Z ReinterpretView( 2025-12-04T09:28:45.9167938Z StorageBox( 2025-12-04T09:28:45.9168185Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9168244Z ), 2025-12-04T09:28:45.9168402Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9168477Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9168548Z stack_traces = {, 2025-12-04T09:28:45.9168774Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9168849Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9168910Z , 2025-12-04T09:28:45.9168965Z } 2025-12-04T09:28:45.9169020Z ) 2025-12-04T09:28:45.9169080Z ) 2025-12-04T09:28:45.9169152Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9169382Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9169441Z )) 2025-12-04T09:28:45.9169444Z 2025-12-04T09:28:45.9169860Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9169869Z 2025-12-04T09:28:45.9169874Z 2025-12-04T09:28:45.9170014Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9170255Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9170258Z 2025-12-04T09:28:45.9170425Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9171465Z FAILED [0.0617s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9171576Z target: aten.mm.default 2025-12-04T09:28:45.9171646Z args[0]: TensorBox( 2025-12-04T09:28:45.9171710Z ReinterpretView( 2025-12-04T09:28:45.9171770Z StorageBox( 2025-12-04T09:28:45.9172012Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9172104Z ), 2025-12-04T09:28:45.9172305Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9172386Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9172451Z stack_traces = {, 2025-12-04T09:28:45.9172683Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9172759Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9172815Z , 2025-12-04T09:28:45.9172879Z } 2025-12-04T09:28:45.9172967Z ) 2025-12-04T09:28:45.9173021Z ) 2025-12-04T09:28:45.9173205Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9173450Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9173511Z )) 2025-12-04T09:28:45.9173514Z 2025-12-04T09:28:45.9173936Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9173942Z 2025-12-04T09:28:45.9173945Z 2025-12-04T09:28:45.9174090Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9174338Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9174341Z 2025-12-04T09:28:45.9174504Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9175549Z FAILED [0.0595s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9175626Z target: aten.mm.default 2025-12-04T09:28:45.9175696Z args[0]: TensorBox( 2025-12-04T09:28:45.9175762Z ReinterpretView( 2025-12-04T09:28:45.9175826Z StorageBox( 2025-12-04T09:28:45.9176069Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9176125Z ), 2025-12-04T09:28:45.9176273Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9176358Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9176421Z stack_traces = {, 2025-12-04T09:28:45.9176654Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9176732Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9176790Z , 2025-12-04T09:28:45.9176851Z } 2025-12-04T09:28:45.9176906Z ) 2025-12-04T09:28:45.9176969Z ) 2025-12-04T09:28:45.9177052Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9177287Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9177345Z )) 2025-12-04T09:28:45.9177351Z 2025-12-04T09:28:45.9177778Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9177781Z 2025-12-04T09:28:45.9177784Z 2025-12-04T09:28:45.9177918Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9178168Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9178172Z 2025-12-04T09:28:45.9178336Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9179483Z FAILED [0.0588s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9179589Z target: aten.mm.default 2025-12-04T09:28:45.9179652Z args[0]: TensorBox( 2025-12-04T09:28:45.9179725Z ReinterpretView( 2025-12-04T09:28:45.9179783Z StorageBox( 2025-12-04T09:28:45.9180020Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9180081Z ), 2025-12-04T09:28:45.9180267Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9180348Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9180409Z stack_traces = {, 2025-12-04T09:28:45.9180637Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9180722Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9180780Z , 2025-12-04T09:28:45.9180833Z } 2025-12-04T09:28:45.9180891Z ) 2025-12-04T09:28:45.9180944Z ) 2025-12-04T09:28:45.9181019Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9181265Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9181319Z )) 2025-12-04T09:28:45.9181322Z 2025-12-04T09:28:45.9181745Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9181748Z 2025-12-04T09:28:45.9181753Z 2025-12-04T09:28:45.9181884Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9182122Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9182133Z 2025-12-04T09:28:45.9182295Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9183340Z FAILED [0.0583s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9183416Z target: aten.mm.default 2025-12-04T09:28:45.9183479Z args[0]: TensorBox( 2025-12-04T09:28:45.9183549Z ReinterpretView( 2025-12-04T09:28:45.9183608Z StorageBox( 2025-12-04T09:28:45.9183842Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9183901Z ), 2025-12-04T09:28:45.9184051Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9184126Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9184196Z stack_traces = {, 2025-12-04T09:28:45.9184420Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9184495Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9184556Z , 2025-12-04T09:28:45.9184611Z } 2025-12-04T09:28:45.9184666Z ) 2025-12-04T09:28:45.9184724Z ) 2025-12-04T09:28:45.9184794Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9185031Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9185086Z )) 2025-12-04T09:28:45.9185089Z 2025-12-04T09:28:45.9185504Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9185559Z 2025-12-04T09:28:45.9185562Z 2025-12-04T09:28:45.9185708Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9185952Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9185988Z 2025-12-04T09:28:45.9186190Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9187225Z FAILED [0.0580s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9187345Z target: aten.mm.default 2025-12-04T09:28:45.9187410Z args[0]: TensorBox( 2025-12-04T09:28:45.9187473Z ReinterpretView( 2025-12-04T09:28:45.9187539Z StorageBox( 2025-12-04T09:28:45.9187779Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9187836Z ), 2025-12-04T09:28:45.9187996Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9188078Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9188151Z stack_traces = {, 2025-12-04T09:28:45.9188384Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9188462Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9188523Z , 2025-12-04T09:28:45.9188578Z } 2025-12-04T09:28:45.9188633Z ) 2025-12-04T09:28:45.9188692Z ) 2025-12-04T09:28:45.9188765Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9189000Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9189062Z )) 2025-12-04T09:28:45.9189067Z 2025-12-04T09:28:45.9189485Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9189490Z 2025-12-04T09:28:45.9189493Z 2025-12-04T09:28:45.9189634Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9189878Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9189881Z 2025-12-04T09:28:45.9190042Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9191088Z FAILED [0.0588s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9191160Z target: aten.mm.default 2025-12-04T09:28:45.9191230Z args[0]: TensorBox( 2025-12-04T09:28:45.9191295Z ReinterpretView( 2025-12-04T09:28:45.9191355Z StorageBox( 2025-12-04T09:28:45.9191601Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9191656Z ), 2025-12-04T09:28:45.9191807Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9191887Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9191950Z stack_traces = {, 2025-12-04T09:28:45.9192178Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9192254Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9192312Z , 2025-12-04T09:28:45.9192374Z } 2025-12-04T09:28:45.9192430Z ) 2025-12-04T09:28:45.9192485Z ) 2025-12-04T09:28:45.9192618Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9192849Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9192906Z )) 2025-12-04T09:28:45.9192945Z 2025-12-04T09:28:45.9193403Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9193408Z 2025-12-04T09:28:45.9193411Z 2025-12-04T09:28:45.9193542Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9193785Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9193788Z 2025-12-04T09:28:45.9193978Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9195016Z FAILED [0.0596s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9195088Z target: aten.mm.default 2025-12-04T09:28:45.9195150Z args[0]: TensorBox( 2025-12-04T09:28:45.9195220Z ReinterpretView( 2025-12-04T09:28:45.9195279Z StorageBox( 2025-12-04T09:28:45.9195509Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9195569Z ), 2025-12-04T09:28:45.9195715Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9195792Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9195856Z stack_traces = {, 2025-12-04T09:28:45.9196076Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9196159Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9196214Z , 2025-12-04T09:28:45.9196267Z } 2025-12-04T09:28:45.9196327Z ) 2025-12-04T09:28:45.9196379Z ) 2025-12-04T09:28:45.9196451Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9196702Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9196758Z )) 2025-12-04T09:28:45.9196761Z 2025-12-04T09:28:45.9197181Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9197185Z 2025-12-04T09:28:45.9197188Z 2025-12-04T09:28:45.9197314Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9197550Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9197558Z 2025-12-04T09:28:45.9197717Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9198754Z FAILED [0.0599s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9198832Z target: aten.mm.default 2025-12-04T09:28:45.9198893Z args[0]: TensorBox( 2025-12-04T09:28:45.9198962Z ReinterpretView( 2025-12-04T09:28:45.9199019Z StorageBox( 2025-12-04T09:28:45.9199251Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9199315Z ), 2025-12-04T09:28:45.9199461Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9199534Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9199655Z stack_traces = {, 2025-12-04T09:28:45.9199873Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9199982Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9200042Z , 2025-12-04T09:28:45.9200097Z } 2025-12-04T09:28:45.9200184Z ) 2025-12-04T09:28:45.9200244Z ) 2025-12-04T09:28:45.9200319Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9200557Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9200611Z )) 2025-12-04T09:28:45.9200615Z 2025-12-04T09:28:45.9201061Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9201066Z 2025-12-04T09:28:45.9201069Z 2025-12-04T09:28:45.9201211Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9201456Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9201459Z 2025-12-04T09:28:45.9201627Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9202667Z FAILED [0.0627s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9202745Z target: aten.mm.default 2025-12-04T09:28:45.9202810Z args[0]: TensorBox( 2025-12-04T09:28:45.9202874Z ReinterpretView( 2025-12-04T09:28:45.9202940Z StorageBox( 2025-12-04T09:28:45.9203175Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9203232Z ), 2025-12-04T09:28:45.9203390Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9203467Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9203533Z stack_traces = {, 2025-12-04T09:28:45.9203762Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9203838Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9203898Z , 2025-12-04T09:28:45.9203953Z } 2025-12-04T09:28:45.9204011Z ) 2025-12-04T09:28:45.9204072Z ) 2025-12-04T09:28:45.9204142Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9204370Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9204430Z )) 2025-12-04T09:28:45.9204434Z 2025-12-04T09:28:45.9205164Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9205173Z 2025-12-04T09:28:45.9205176Z 2025-12-04T09:28:45.9205319Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9205564Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9205570Z 2025-12-04T09:28:45.9205733Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9206779Z FAILED [0.0598s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9206849Z target: aten.mm.default 2025-12-04T09:28:45.9206920Z args[0]: TensorBox( 2025-12-04T09:28:45.9207113Z ReinterpretView( 2025-12-04T09:28:45.9207176Z StorageBox( 2025-12-04T09:28:45.9207419Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9207530Z ), 2025-12-04T09:28:45.9207726Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9207812Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9207876Z stack_traces = {, 2025-12-04T09:28:45.9208106Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9208183Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9208237Z , 2025-12-04T09:28:45.9208298Z } 2025-12-04T09:28:45.9208354Z ) 2025-12-04T09:28:45.9208408Z ) 2025-12-04T09:28:45.9208542Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9208777Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9208834Z )) 2025-12-04T09:28:45.9208838Z 2025-12-04T09:28:45.9209258Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9209264Z 2025-12-04T09:28:45.9209267Z 2025-12-04T09:28:45.9209405Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9209652Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9209656Z 2025-12-04T09:28:45.9209825Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9210868Z FAILED [0.0598s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9210939Z target: aten.mm.default 2025-12-04T09:28:45.9211001Z args[0]: TensorBox( 2025-12-04T09:28:45.9211074Z ReinterpretView( 2025-12-04T09:28:45.9211132Z StorageBox( 2025-12-04T09:28:45.9211366Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9211426Z ), 2025-12-04T09:28:45.9211575Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9211652Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9211715Z stack_traces = {, 2025-12-04T09:28:45.9211935Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9212019Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9212075Z , 2025-12-04T09:28:45.9212128Z } 2025-12-04T09:28:45.9212190Z ) 2025-12-04T09:28:45.9212245Z ) 2025-12-04T09:28:45.9212316Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9212556Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9212612Z )) 2025-12-04T09:28:45.9212615Z 2025-12-04T09:28:45.9213034Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9213038Z 2025-12-04T09:28:45.9213040Z 2025-12-04T09:28:45.9213268Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9213507Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9213515Z 2025-12-04T09:28:45.9213679Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9214711Z FAILED [0.0587s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9214892Z target: aten.mm.default 2025-12-04T09:28:45.9214953Z args[0]: TensorBox( 2025-12-04T09:28:45.9215016Z ReinterpretView( 2025-12-04T09:28:45.9215083Z StorageBox( 2025-12-04T09:28:45.9215318Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9215383Z ), 2025-12-04T09:28:45.9215537Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9215643Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9215713Z stack_traces = {, 2025-12-04T09:28:45.9215930Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9216006Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9216067Z , 2025-12-04T09:28:45.9216120Z } 2025-12-04T09:28:45.9216177Z ) 2025-12-04T09:28:45.9216237Z ) 2025-12-04T09:28:45.9216310Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9216550Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9216605Z )) 2025-12-04T09:28:45.9216609Z 2025-12-04T09:28:45.9217020Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9217024Z 2025-12-04T09:28:45.9217027Z 2025-12-04T09:28:45.9217164Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9217400Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9217406Z 2025-12-04T09:28:45.9217573Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9218603Z FAILED [0.0580s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9218680Z target: aten.mm.default 2025-12-04T09:28:45.9218742Z args[0]: TensorBox( 2025-12-04T09:28:45.9218806Z ReinterpretView( 2025-12-04T09:28:45.9218871Z StorageBox( 2025-12-04T09:28:45.9219107Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9219162Z ), 2025-12-04T09:28:45.9219313Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9219399Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9219463Z stack_traces = {, 2025-12-04T09:28:45.9219688Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9219765Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9219828Z , 2025-12-04T09:28:45.9219883Z } 2025-12-04T09:28:45.9219937Z ) 2025-12-04T09:28:45.9219996Z ) 2025-12-04T09:28:45.9220065Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9220294Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9220355Z )) 2025-12-04T09:28:45.9220358Z 2025-12-04T09:28:45.9220773Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9220777Z 2025-12-04T09:28:45.9220821Z 2025-12-04T09:28:45.9220955Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9221196Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9221237Z 2025-12-04T09:28:45.9221396Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9222484Z FAILED [0.0604s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9222553Z target: aten.mm.default 2025-12-04T09:28:45.9222655Z args[0]: TensorBox( 2025-12-04T09:28:45.9222721Z ReinterpretView( 2025-12-04T09:28:45.9222781Z StorageBox( 2025-12-04T09:28:45.9223022Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9223081Z ), 2025-12-04T09:28:45.9223231Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9223310Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9223374Z stack_traces = {, 2025-12-04T09:28:45.9223600Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9223674Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9223729Z , 2025-12-04T09:28:45.9223792Z } 2025-12-04T09:28:45.9223850Z ) 2025-12-04T09:28:45.9223903Z ) 2025-12-04T09:28:45.9223981Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9224211Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9224266Z )) 2025-12-04T09:28:45.9224269Z 2025-12-04T09:28:45.9224689Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9224695Z 2025-12-04T09:28:45.9224698Z 2025-12-04T09:28:45.9224829Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9225086Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9225090Z 2025-12-04T09:28:45.9225249Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9226292Z FAILED [0.0589s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9226363Z target: aten.mm.default 2025-12-04T09:28:45.9226424Z args[0]: TensorBox( 2025-12-04T09:28:45.9226495Z ReinterpretView( 2025-12-04T09:28:45.9226554Z StorageBox( 2025-12-04T09:28:45.9226783Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9226849Z ), 2025-12-04T09:28:45.9226999Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9227077Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9227140Z stack_traces = {, 2025-12-04T09:28:45.9227353Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9227432Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9227486Z , 2025-12-04T09:28:45.9227539Z } 2025-12-04T09:28:45.9227601Z ) 2025-12-04T09:28:45.9227652Z ) 2025-12-04T09:28:45.9227722Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9227954Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9228058Z )) 2025-12-04T09:28:45.9228062Z 2025-12-04T09:28:45.9228481Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9228549Z 2025-12-04T09:28:45.9228553Z 2025-12-04T09:28:45.9228678Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9228914Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9228923Z 2025-12-04T09:28:45.9229081Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9230166Z FAILED [0.0587s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9230243Z target: aten.mm.default 2025-12-04T09:28:45.9230306Z args[0]: TensorBox( 2025-12-04T09:28:45.9230368Z ReinterpretView( 2025-12-04T09:28:45.9230433Z StorageBox( 2025-12-04T09:28:45.9230662Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9230721Z ), 2025-12-04T09:28:45.9230866Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9230939Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9231006Z stack_traces = {, 2025-12-04T09:28:45.9231223Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9231294Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9231356Z , 2025-12-04T09:28:45.9231409Z } 2025-12-04T09:28:45.9231462Z ) 2025-12-04T09:28:45.9231522Z ) 2025-12-04T09:28:45.9231592Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9231824Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9231883Z )) 2025-12-04T09:28:45.9231889Z 2025-12-04T09:28:45.9232300Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9232304Z 2025-12-04T09:28:45.9232307Z 2025-12-04T09:28:45.9232437Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9232675Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9232679Z 2025-12-04T09:28:45.9232840Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9233882Z FAILED [0.0624s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9233957Z target: aten.mm.default 2025-12-04T09:28:45.9234018Z args[0]: TensorBox( 2025-12-04T09:28:45.9234081Z ReinterpretView( 2025-12-04T09:28:45.9234153Z StorageBox( 2025-12-04T09:28:45.9234391Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9234447Z ), 2025-12-04T09:28:45.9234599Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9234671Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9234735Z stack_traces = {, 2025-12-04T09:28:45.9235004Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9235078Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9235177Z , 2025-12-04T09:28:45.9235235Z } 2025-12-04T09:28:45.9235291Z ) 2025-12-04T09:28:45.9235351Z ) 2025-12-04T09:28:45.9235456Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9235685Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9235744Z )) 2025-12-04T09:28:45.9235747Z 2025-12-04T09:28:45.9236158Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9236161Z 2025-12-04T09:28:45.9236196Z 2025-12-04T09:28:45.9236335Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9236572Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9236578Z 2025-12-04T09:28:45.9236738Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9237789Z FAILED [0.0603s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9237862Z target: aten.mm.default 2025-12-04T09:28:45.9237932Z args[0]: TensorBox( 2025-12-04T09:28:45.9237996Z ReinterpretView( 2025-12-04T09:28:45.9238058Z StorageBox( 2025-12-04T09:28:45.9238302Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9238357Z ), 2025-12-04T09:28:45.9238507Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9238590Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9238654Z stack_traces = {, 2025-12-04T09:28:45.9238879Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9238956Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9239011Z , 2025-12-04T09:28:45.9239072Z } 2025-12-04T09:28:45.9239128Z ) 2025-12-04T09:28:45.9239185Z ) 2025-12-04T09:28:45.9239263Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9239492Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9239556Z )) 2025-12-04T09:28:45.9239561Z 2025-12-04T09:28:45.9239986Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9239991Z 2025-12-04T09:28:45.9239994Z 2025-12-04T09:28:45.9240120Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9240359Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9240364Z 2025-12-04T09:28:45.9240523Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9241559Z FAILED [0.0586s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9241630Z target: aten.mm.default 2025-12-04T09:28:45.9241691Z args[0]: TensorBox( 2025-12-04T09:28:45.9241761Z ReinterpretView( 2025-12-04T09:28:45.9241871Z StorageBox( 2025-12-04T09:28:45.9242101Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9242163Z ), 2025-12-04T09:28:45.9242509Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9242625Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9242689Z stack_traces = {, 2025-12-04T09:28:45.9242906Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9242983Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9243040Z , 2025-12-04T09:28:45.9243092Z } 2025-12-04T09:28:45.9243151Z ) 2025-12-04T09:28:45.9243204Z ) 2025-12-04T09:28:45.9243274Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9243539Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9243594Z )) 2025-12-04T09:28:45.9243599Z 2025-12-04T09:28:45.9244017Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9244023Z 2025-12-04T09:28:45.9244026Z 2025-12-04T09:28:45.9244157Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9244395Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9244405Z 2025-12-04T09:28:45.9244561Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9245600Z FAILED [0.0578s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9245677Z target: aten.mm.default 2025-12-04T09:28:45.9245738Z args[0]: TensorBox( 2025-12-04T09:28:45.9245801Z ReinterpretView( 2025-12-04T09:28:45.9245867Z StorageBox( 2025-12-04T09:28:45.9246102Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9246161Z ), 2025-12-04T09:28:45.9246306Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9246378Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9246446Z stack_traces = {, 2025-12-04T09:28:45.9246663Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9246736Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9246802Z , 2025-12-04T09:28:45.9246861Z } 2025-12-04T09:28:45.9246915Z ) 2025-12-04T09:28:45.9246984Z ) 2025-12-04T09:28:45.9247061Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9247297Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9247350Z )) 2025-12-04T09:28:45.9247354Z 2025-12-04T09:28:45.9247769Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9247772Z 2025-12-04T09:28:45.9247775Z 2025-12-04T09:28:45.9247909Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9248149Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9248153Z 2025-12-04T09:28:45.9248319Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9249354Z FAILED [0.0589s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9249517Z target: aten.mm.default 2025-12-04T09:28:45.9249581Z args[0]: TensorBox( 2025-12-04T09:28:45.9249678Z ReinterpretView( 2025-12-04T09:28:45.9249743Z StorageBox( 2025-12-04T09:28:45.9249984Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9250039Z ), 2025-12-04T09:28:45.9250190Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9250262Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9250362Z stack_traces = {, 2025-12-04T09:28:45.9250586Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9250662Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9250722Z , 2025-12-04T09:28:45.9250776Z } 2025-12-04T09:28:45.9250830Z ) 2025-12-04T09:28:45.9250889Z ) 2025-12-04T09:28:45.9250959Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9251192Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9251255Z )) 2025-12-04T09:28:45.9251258Z 2025-12-04T09:28:45.9251673Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9251676Z 2025-12-04T09:28:45.9251679Z 2025-12-04T09:28:45.9251813Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9252056Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9252059Z 2025-12-04T09:28:45.9252220Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9253384Z FAILED [0.0603s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9253459Z target: aten.mm.default 2025-12-04T09:28:45.9253527Z args[0]: TensorBox( 2025-12-04T09:28:45.9253593Z ReinterpretView( 2025-12-04T09:28:45.9253652Z StorageBox( 2025-12-04T09:28:45.9253897Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9253955Z ), 2025-12-04T09:28:45.9254108Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9254191Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9254259Z stack_traces = {, 2025-12-04T09:28:45.9254488Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9254566Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9254622Z , 2025-12-04T09:28:45.9254686Z } 2025-12-04T09:28:45.9254741Z ) 2025-12-04T09:28:45.9254797Z ) 2025-12-04T09:28:45.9254877Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9255109Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9255164Z )) 2025-12-04T09:28:45.9255168Z 2025-12-04T09:28:45.9255596Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9255600Z 2025-12-04T09:28:45.9255602Z 2025-12-04T09:28:45.9255738Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9256041Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9256045Z 2025-12-04T09:28:45.9256209Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9257352Z FAILED [0.0603s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9257423Z target: aten.mm.default 2025-12-04T09:28:45.9257488Z args[0]: TensorBox( 2025-12-04T09:28:45.9257558Z ReinterpretView( 2025-12-04T09:28:45.9257651Z StorageBox( 2025-12-04T09:28:45.9257889Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9257952Z ), 2025-12-04T09:28:45.9258103Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9258180Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9258247Z stack_traces = {, 2025-12-04T09:28:45.9258470Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9258551Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9258608Z , 2025-12-04T09:28:45.9258661Z } 2025-12-04T09:28:45.9258721Z ) 2025-12-04T09:28:45.9258774Z ) 2025-12-04T09:28:45.9258850Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9259096Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9259152Z )) 2025-12-04T09:28:45.9259155Z 2025-12-04T09:28:45.9259581Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9259587Z 2025-12-04T09:28:45.9259590Z 2025-12-04T09:28:45.9259729Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9259981Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9259987Z 2025-12-04T09:28:45.9260159Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9261195Z FAILED [0.0607s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9261274Z target: aten.mm.default 2025-12-04T09:28:45.9261337Z args[0]: TensorBox( 2025-12-04T09:28:45.9261404Z ReinterpretView( 2025-12-04T09:28:45.9261470Z StorageBox( 2025-12-04T09:28:45.9261711Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9261775Z ), 2025-12-04T09:28:45.9261928Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9262005Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9262074Z stack_traces = {, 2025-12-04T09:28:45.9262298Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9262375Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9262437Z , 2025-12-04T09:28:45.9262492Z } 2025-12-04T09:28:45.9262550Z ) 2025-12-04T09:28:45.9262610Z ) 2025-12-04T09:28:45.9262684Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9262923Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9263020Z )) 2025-12-04T09:28:45.9263023Z 2025-12-04T09:28:45.9263439Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9263478Z 2025-12-04T09:28:45.9263481Z 2025-12-04T09:28:45.9263660Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9263904Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9263908Z 2025-12-04T09:28:45.9264076Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9265146Z FAILED [0.0602s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9265229Z target: aten.mm.default 2025-12-04T09:28:45.9265292Z args[0]: TensorBox( 2025-12-04T09:28:45.9265360Z ReinterpretView( 2025-12-04T09:28:45.9265425Z StorageBox( 2025-12-04T09:28:45.9265662Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9265719Z ), 2025-12-04T09:28:45.9265874Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9265950Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9266014Z stack_traces = {, 2025-12-04T09:28:45.9266241Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9266318Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9266374Z , 2025-12-04T09:28:45.9266447Z } 2025-12-04T09:28:45.9266505Z ) 2025-12-04T09:28:45.9266564Z ) 2025-12-04T09:28:45.9266636Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9266870Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9266933Z )) 2025-12-04T09:28:45.9266936Z 2025-12-04T09:28:45.9267351Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9267355Z 2025-12-04T09:28:45.9267359Z 2025-12-04T09:28:45.9267496Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9267737Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9267740Z 2025-12-04T09:28:45.9267903Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9268941Z FAILED [0.0599s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9269021Z target: aten.mm.default 2025-12-04T09:28:45.9269088Z args[0]: TensorBox( 2025-12-04T09:28:45.9269163Z ReinterpretView( 2025-12-04T09:28:45.9269224Z StorageBox( 2025-12-04T09:28:45.9269463Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9269519Z ), 2025-12-04T09:28:45.9269665Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9269747Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9269811Z stack_traces = {, 2025-12-04T09:28:45.9270035Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9270155Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9270211Z , 2025-12-04T09:28:45.9270273Z } 2025-12-04T09:28:45.9270363Z ) 2025-12-04T09:28:45.9270422Z ) 2025-12-04T09:28:45.9270502Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9270770Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9270827Z )) 2025-12-04T09:28:45.9270831Z 2025-12-04T09:28:45.9271267Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9271270Z 2025-12-04T09:28:45.9271273Z 2025-12-04T09:28:45.9271436Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9271685Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9271692Z 2025-12-04T09:28:45.9271852Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9272900Z FAILED [0.0606s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9272971Z target: aten.mm.default 2025-12-04T09:28:45.9273036Z args[0]: TensorBox( 2025-12-04T09:28:45.9273105Z ReinterpretView( 2025-12-04T09:28:45.9273163Z StorageBox( 2025-12-04T09:28:45.9273397Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9273459Z ), 2025-12-04T09:28:45.9273606Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9273685Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9273749Z stack_traces = {, 2025-12-04T09:28:45.9273966Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9274047Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9274102Z , 2025-12-04T09:28:45.9274157Z } 2025-12-04T09:28:45.9274223Z ) 2025-12-04T09:28:45.9274278Z ) 2025-12-04T09:28:45.9274350Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9274585Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9274639Z )) 2025-12-04T09:28:45.9274642Z 2025-12-04T09:28:45.9275062Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9275065Z 2025-12-04T09:28:45.9275068Z 2025-12-04T09:28:45.9275202Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9275444Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9275449Z 2025-12-04T09:28:45.9275625Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9276657Z FAILED [0.0601s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9276731Z target: aten.mm.default 2025-12-04T09:28:45.9276794Z args[0]: TensorBox( 2025-12-04T09:28:45.9276858Z ReinterpretView( 2025-12-04T09:28:45.9276926Z StorageBox( 2025-12-04T09:28:45.9277157Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9277260Z ), 2025-12-04T09:28:45.9277405Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9277527Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9277597Z stack_traces = {, 2025-12-04T09:28:45.9277847Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9277924Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9277986Z , 2025-12-04T09:28:45.9278039Z } 2025-12-04T09:28:45.9278103Z ) 2025-12-04T09:28:45.9278169Z ) 2025-12-04T09:28:45.9278239Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9278510Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9278566Z )) 2025-12-04T09:28:45.9278569Z 2025-12-04T09:28:45.9278984Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9278991Z 2025-12-04T09:28:45.9278994Z 2025-12-04T09:28:45.9279133Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9279374Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9279378Z 2025-12-04T09:28:45.9279543Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9280579Z FAILED [0.0604s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9280649Z target: aten.mm.default 2025-12-04T09:28:45.9280719Z args[0]: TensorBox( 2025-12-04T09:28:45.9280784Z ReinterpretView( 2025-12-04T09:28:45.9280860Z StorageBox( 2025-12-04T09:28:45.9281094Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9281153Z ), 2025-12-04T09:28:45.9281310Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9281383Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9281447Z stack_traces = {, 2025-12-04T09:28:45.9281672Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9281746Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9281800Z , 2025-12-04T09:28:45.9281863Z } 2025-12-04T09:28:45.9281934Z ) 2025-12-04T09:28:45.9281995Z ) 2025-12-04T09:28:45.9282069Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9282299Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9282362Z )) 2025-12-04T09:28:45.9282366Z 2025-12-04T09:28:45.9282779Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9282786Z 2025-12-04T09:28:45.9282789Z 2025-12-04T09:28:45.9282924Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9283167Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9283170Z 2025-12-04T09:28:45.9283330Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9284372Z FAILED [0.0600s] inductor/test_kernel_benchmark.py::TestKernelBenchmark::test_mm_triton_kernel_benchmark - torch._inductor.exc.InductorError: LoweringException: NoValidChoicesError: No choices to select. Provided reason: No choices exist for backend. please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice. 2025-12-04T09:28:45.9284493Z target: aten.mm.default 2025-12-04T09:28:45.9284617Z args[0]: TensorBox( 2025-12-04T09:28:45.9284685Z ReinterpretView( 2025-12-04T09:28:45.9284777Z StorageBox( 2025-12-04T09:28:45.9285017Z InputBuffer(name='arg0_1', layout=FixedLayout('cuda:0', torch.float16, size=[2048, 3581], stride=[3581, 1])) 2025-12-04T09:28:45.9285076Z ), 2025-12-04T09:28:45.9285225Z FixedLayout('cuda:0', torch.float16, size=[2048, 1949], stride=[3581, 1]), 2025-12-04T09:28:45.9285306Z origins=OrderedSet([slice_1]), 2025-12-04T09:28:45.9285369Z stack_traces = {, 2025-12-04T09:28:45.9285625Z File "/var/lib/jenkins/workspace/test/inductor/test_kernel_benchmark.py", line 184, in f, 2025-12-04T09:28:45.9285699Z a_1 = torch.narrow(a, 1, 0, K), 2025-12-04T09:28:45.9285758Z , 2025-12-04T09:28:45.9285818Z } 2025-12-04T09:28:45.9285873Z ) 2025-12-04T09:28:45.9285928Z ) 2025-12-04T09:28:45.9286008Z args[1]: TensorBox(StorageBox( 2025-12-04T09:28:45.9286237Z InputBuffer(name='arg1_1', layout=FixedLayout('cuda:0', torch.float16, size=[1949, 2432], stride=[1, 1949])) 2025-12-04T09:28:45.9286295Z )) 2025-12-04T09:28:45.9286300Z 2025-12-04T09:28:45.9286721Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-12-04T09:28:45.9286724Z 2025-12-04T09:28:45.9286727Z 2025-12-04T09:28:45.9286856Z To execute this test, run the following from the base repo dir: 2025-12-04T09:28:45.9287100Z python test/inductor/test_kernel_benchmark.py TestKernelBenchmark.test_mm_triton_kernel_benchmark 2025-12-04T09:28:45.9287103Z 2025-12-04T09:28:45.9287265Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:28:45.9287374Z ======================== 50 failed, 50 skipped in 4.47s ======================== 2025-12-04T09:28:45.9287384Z 2025-12-04T09:28:45.9287739Z FINISHED PRINTING LOG FILE of inductor/test_kernel_benchmark 1/1 (test/test-reports/inductor.test_kernel_benchmark_1.1_4dd3a5ee490d965b_.log) 2025-12-04T09:28:45.9287744Z 2025-12-04T09:28:45.9287973Z Finished inductor/test_kernel_benchmark 1/1 ... [2025-12-04 09:28:45.729844][2161.689384974], took 0.19min 2025-12-04T09:28:45.9288477Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_kernel_benchmark/inductor.test_kernel_benchmark-8244c39f5fb36d0d.xml 2025-12-04T09:28:46.5615217Z Uploading logs for 57116084869 to S3 2025-12-04T09:28:46.6087290Z Uploading artifacts took 0.77 seconds 2025-12-04T09:28:46.6087700Z inductor/test_kernel_benchmark 1/1 failed! 2025-12-04T09:28:49.3554662Z Running inductor/test_torchinductor_opinfo 1/13 ... [2025-12-04 09:28:49.354964][2165.314506768] 2025-12-04T09:28:49.3555435Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:28:49.3557371Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '-m', 'not serial', '--shard-id=1', '--num-shards=13', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:28:49.355380] 2025-12-04T09:29:02.0089989Z 2025-12-04T09:29:02.0091603Z inductor/test_torchinductor_opinfo 1/13 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_1.13_5617d11ad1b4a636_.log 2025-12-04T09:29:02.0125075Z Running 100 items in this shard: test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32 2025-12-04T09:29:02.0156722Z 2025-12-04T09:29:02.0156994Z Finished inductor/test_torchinductor_opinfo 1/13 ... [2025-12-04 09:29:02.009035][2177.968578012], took 0.21min 2025-12-04T09:29:02.0302585Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-b56e595397fe8756.xml 2025-12-04T09:29:05.5564369Z Running inductor/test_torchinductor_opinfo 5/13 ... [2025-12-04 09:29:05.555931][2181.515475428] 2025-12-04T09:29:05.5564883Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:29:05.5566889Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '-m', 'not serial', '--shard-id=5', '--num-shards=13', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:29:05.556385] 2025-12-04T09:29:18.1090867Z 2025-12-04T09:29:18.1092432Z inductor/test_torchinductor_opinfo 5/13 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_5.13_ab6bb32b2389ca26_.log 2025-12-04T09:29:18.1120126Z Running 50 items in this shard: test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_float16 2025-12-04T09:29:18.1145500Z 2025-12-04T09:29:18.1145920Z Finished inductor/test_torchinductor_opinfo 5/13 ... [2025-12-04 09:29:18.108992][2194.06853733], took 0.21min 2025-12-04T09:29:18.1211942Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-0b2c1e92820d7d68.xml 2025-12-04T09:29:21.6856256Z Running inductor/test_torchinductor_opinfo 7/13 ... [2025-12-04 09:29:21.685144][2197.644688009] 2025-12-04T09:29:21.6856908Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:29:21.6859570Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '-m', 'not serial', '--shard-id=7', '--num-shards=13', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:29:21.685611] 2025-12-04T09:29:33.3057742Z 2025-12-04T09:29:33.3058437Z PRINTING LOG FILE of test_decomp 18/22 (test/test-reports/test_decomp_18.22_6d2264971e288acd_.log) 2025-12-04T09:29:33.3059343Z Test results will be stored in test-reports/python-pytest/test_decomp/test_decomp-a49ca6736f66875e.xml 2025-12-04T09:29:33.3060051Z ============================= test session starts ============================== 2025-12-04T09:29:33.3060721Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:29:33.3061251Z cachedir: .pytest_cache 2025-12-04T09:29:33.3061895Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:29:33.3062639Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:29:33.3062954Z configfile: pytest.ini 2025-12-04T09:29:33.3063582Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:29:33.3064172Z collecting ... collected 9010 items 2025-12-04T09:29:33.3064446Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:29:33.3078656Z Running 50 items in this shard: test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 2025-12-04T09:29:33.3091511Z 2025-12-04T09:29:33.3091798Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.5611s] [ 2%] 2025-12-04T09:29:33.3092413Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.1568s] [ 2%] 2025-12-04T09:29:33.3092998Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.2797s] [ 2%] 2025-12-04T09:29:33.3093692Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.2669s] [ 2%] 2025-12-04T09:29:33.3094280Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.2013s] [ 2%] 2025-12-04T09:29:33.3094861Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.3057s] [ 2%] 2025-12-04T09:29:33.3095496Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.3286s] [ 2%] 2025-12-04T09:29:33.3096072Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.1813s] [ 2%] 2025-12-04T09:29:33.3096732Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.1616s] [ 2%] 2025-12-04T09:29:33.3097315Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.2131s] [ 2%] 2025-12-04T09:29:33.3097901Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 FAILED [1.4844s] [ 2%] 2025-12-04T09:29:33.3098508Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.2562s] [ 2%] 2025-12-04T09:29:33.3099103Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.4067s] [ 2%] 2025-12-04T09:29:33.3099685Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.2487s] [ 2%] 2025-12-04T09:29:33.3100257Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.2739s] [ 2%] 2025-12-04T09:29:33.3100836Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.1940s] [ 2%] 2025-12-04T09:29:33.3101418Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.2310s] [ 2%] 2025-12-04T09:29:33.3102000Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.1591s] [ 2%] 2025-12-04T09:29:33.3102576Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.2203s] [ 2%] 2025-12-04T09:29:33.3103157Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.3350s] [ 2%] 2025-12-04T09:29:33.3103743Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.3852s] [ 2%] 2025-12-04T09:29:33.3104327Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.4266s] [ 2%] 2025-12-04T09:29:33.3105244Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.4153s] [ 2%] 2025-12-04T09:29:33.3105845Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.4659s] [ 2%] 2025-12-04T09:29:33.3106431Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.7224s] [ 2%] 2025-12-04T09:29:33.3107015Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.7107s] [ 2%] 2025-12-04T09:29:33.3107601Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.5583s] [ 2%] 2025-12-04T09:29:33.3108189Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.3132s] [ 2%] 2025-12-04T09:29:33.3108774Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.3403s] [ 2%] 2025-12-04T09:29:33.3109357Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.3779s] [ 2%] 2025-12-04T09:29:33.3109941Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.4346s] [ 2%] 2025-12-04T09:29:33.3110526Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.4106s] [ 2%] 2025-12-04T09:29:33.3111108Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.2641s] [ 2%] 2025-12-04T09:29:33.3111694Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.4079s] [ 2%] 2025-12-04T09:29:33.3112272Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.3292s] [ 2%] 2025-12-04T09:29:33.3112985Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.4206s] [ 2%] 2025-12-04T09:29:33.3113574Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.4014s] [ 2%] 2025-12-04T09:29:33.3114297Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.3319s] [ 2%] 2025-12-04T09:29:33.3114882Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.1779s] [ 2%] 2025-12-04T09:29:33.3115491Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 FAILED [1.4553s] [ 2%] 2025-12-04T09:29:33.3116330Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.2052s] [ 2%] 2025-12-04T09:29:33.3117017Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.2088s] [ 2%] 2025-12-04T09:29:33.3117961Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.2661s] [ 2%] 2025-12-04T09:29:33.3118721Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.4101s] [ 2%] 2025-12-04T09:29:33.3119314Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.3178s] [ 2%] 2025-12-04T09:29:33.3119895Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.2687s] [ 2%] 2025-12-04T09:29:33.3120468Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.4501s] [ 2%] 2025-12-04T09:29:33.3121045Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.3490s] [ 2%] 2025-12-04T09:29:33.3121623Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.4028s] [ 2%] 2025-12-04T09:29:33.3122208Z test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 PASSED [7.2887s] [ 2%] 2025-12-04T09:29:33.3122538Z 2025-12-04T09:29:33.3122634Z =================================== FAILURES =================================== 2025-12-04T09:29:33.3123006Z ____ TestDecompCUDA.test_comprehensive_nn_functional_nll_loss_cuda_float64 _____ 2025-12-04T09:29:33.3123345Z Traceback (most recent call last): 2025-12-04T09:29:33.3123841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_device_type.py", line 1151, in test_wrapper 2025-12-04T09:29:33.3124340Z return test(*args, **kwargs) 2025-12-04T09:29:33.3124795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_device_type.py", line 1473, in only_fn 2025-12-04T09:29:33.3125291Z return fn(self, *args, **kwargs) 2025-12-04T09:29:33.3133556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 1647, in wrapper 2025-12-04T09:29:33.3134076Z fn(*args, **kwargs) 2025-12-04T09:29:33.3134553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 2464, in wrapper 2025-12-04T09:29:33.3135042Z fn(*args, **kwargs) 2025-12-04T09:29:33.3135376Z File "/var/lib/jenkins/workspace/test/test_decomp.py", line 602, in test_comprehensive 2025-12-04T09:29:33.3135790Z self.do_cross_ref(device, dtype, op, run_all=True) 2025-12-04T09:29:33.3136290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 1736, in wrapper 2025-12-04T09:29:33.3136746Z fn(*args, **kwargs) 2025-12-04T09:29:33.3137054Z File "/var/lib/jenkins/workspace/test/test_decomp.py", line 989, in do_cross_ref 2025-12-04T09:29:33.3137462Z decomp_out, decomp_vjp_fn = ref_vjp_no_create(fn, *primals) 2025-12-04T09:29:33.3137891Z File "/var/lib/jenkins/workspace/test/test_decomp.py", line 147, in ref_vjp_no_create 2025-12-04T09:29:33.3138246Z result = f(*primals) 2025-12-04T09:29:33.3138648Z File "/var/lib/jenkins/workspace/test/test_decomp.py", line 332, in wrapped 2025-12-04T09:29:33.3138975Z result = f(*_args, **kwargs) 2025-12-04T09:29:33.3139381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 3179, in nll_loss 2025-12-04T09:29:33.3139842Z return torch._C._nn.nll_loss_nd( 2025-12-04T09:29:33.3140267Z File "/var/lib/jenkins/workspace/test/test_decomp.py", line 904, in __torch_dispatch__ 2025-12-04T09:29:33.3140626Z op_assert_equal( 2025-12-04T09:29:33.3140935Z File "/var/lib/jenkins/workspace/test/test_decomp.py", line 301, in op_assert_equal 2025-12-04T09:29:33.3141288Z test_case.assertEqual( 2025-12-04T09:29:33.3141738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:29:33.3142303Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:29:33.3142610Z AssertionError: Scalars are not close! 2025-12-04T09:29:33.3142783Z 2025-12-04T09:29:33.3142862Z Expected 0.0 but got 10.757649417900529. 2025-12-04T09:29:33.3143151Z Absolute difference: 10.757649417900529 (up to 1e-07 allowed) 2025-12-04T09:29:33.3143465Z Relative difference: inf (up to 1e-07 allowed) 2025-12-04T09:29:33.3143716Z nll_loss2d_forward.default 2025-12-04T09:29:33.3143921Z args = (tensor([[[[-4.4896, 3.0513], 2025-12-04T09:29:33.3144126Z [-5.4787, 1.1693]], 2025-12-04T09:29:33.3144256Z 2025-12-04T09:29:33.3144343Z [[-3.6457, -3.2967], 2025-12-04T09:29:33.3144534Z [-4.9368, 6.3443]], 2025-12-04T09:29:33.3144653Z 2025-12-04T09:29:33.3144720Z [[-2.4919, -0.1966], 2025-12-04T09:29:33.3144895Z [ 4.2073, 7.7745]]], 2025-12-04T09:29:33.3145025Z 2025-12-04T09:29:33.3145028Z 2025-12-04T09:29:33.3145089Z [[[ 2.5429, 6.0313], 2025-12-04T09:29:33.3145277Z [ 3.0112, 4.1608]], 2025-12-04T09:29:33.3145395Z 2025-12-04T09:29:33.3145458Z [[ 7.4791, -2.5155], 2025-12-04T09:29:33.3145643Z [-0.7277, 7.4173]], 2025-12-04T09:29:33.3145757Z 2025-12-04T09:29:33.3145830Z [[-7.6495, -8.6749], 2025-12-04T09:29:33.3146067Z [ 7.0393, -4.4170]]]], device='cuda:0', dtype=torch.float64, 2025-12-04T09:29:33.3146358Z requires_grad=True), tensor([[[1, 0], 2025-12-04T09:29:33.3146584Z [1, 0]], 2025-12-04T09:29:33.3146681Z 2025-12-04T09:29:33.3146748Z [[0, 1], 2025-12-04T09:29:33.3147053Z [0, 0]]], device='cuda:0'), tensor([-2.3961, -3.4373, -4.4814], device='cuda:0', dtype=torch.float64), 0, -100) 2025-12-04T09:29:33.3147396Z kwargs = {} 2025-12-04T09:29:33.3147486Z 2025-12-04T09:29:33.3147655Z The above exception was the direct cause of the following exception: 2025-12-04T09:29:33.3147885Z 2025-12-04T09:29:33.3147973Z Traceback (most recent call last): 2025-12-04T09:29:33.3148439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:29:33.3148908Z method(*args, **kwargs) 2025-12-04T09:29:33.3149355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:29:33.3149816Z method(*args, **kwargs) 2025-12-04T09:29:33.3150324Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_device_type.py", line 428, in instantiated_test 2025-12-04T09:29:33.3150857Z result = test(self, **param_kwargs) 2025-12-04T09:29:33.3151325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 1736, in wrapper 2025-12-04T09:29:33.3151767Z fn(*args, **kwargs) 2025-12-04T09:29:33.3152227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_device_type.py", line 1163, in test_wrapper 2025-12-04T09:29:33.3152717Z raise e_tracked from e 2025-12-04T09:29:33.3152914Z Exception: Scalars are not close! 2025-12-04T09:29:33.3153064Z 2025-12-04T09:29:33.3153199Z Expected 0.0 but got 10.757649417900529. 2025-12-04T09:29:33.3153493Z Absolute difference: 10.757649417900529 (up to 1e-07 allowed) 2025-12-04T09:29:33.3153804Z Relative difference: inf (up to 1e-07 allowed) 2025-12-04T09:29:33.3154087Z nll_loss2d_forward.default 2025-12-04T09:29:33.3154285Z args = (tensor([[[[-4.4896, 3.0513], 2025-12-04T09:29:33.3154530Z [-5.4787, 1.1693]], 2025-12-04T09:29:33.3154655Z 2025-12-04T09:29:33.3154719Z [[-3.6457, -3.2967], 2025-12-04T09:29:33.3154904Z [-4.9368, 6.3443]], 2025-12-04T09:29:33.3155019Z 2025-12-04T09:29:33.3155088Z [[-2.4919, -0.1966], 2025-12-04T09:29:33.3155261Z [ 4.2073, 7.7745]]], 2025-12-04T09:29:33.3155398Z 2025-12-04T09:29:33.3155401Z 2025-12-04T09:29:33.3155459Z [[[ 2.5429, 6.0313], 2025-12-04T09:29:33.3155678Z [ 3.0112, 4.1608]], 2025-12-04T09:29:33.3155792Z 2025-12-04T09:29:33.3155860Z [[ 7.4791, -2.5155], 2025-12-04T09:29:33.3156041Z [-0.7277, 7.4173]], 2025-12-04T09:29:33.3156166Z 2025-12-04T09:29:33.3156229Z [[-7.6495, -8.6749], 2025-12-04T09:29:33.3156479Z [ 7.0393, -4.4170]]]], device='cuda:0', dtype=torch.float64, 2025-12-04T09:29:33.3156752Z requires_grad=True), tensor([[[1, 0], 2025-12-04T09:29:33.3156967Z [1, 0]], 2025-12-04T09:29:33.3157063Z 2025-12-04T09:29:33.3157127Z [[0, 1], 2025-12-04T09:29:33.3157425Z [0, 0]]], device='cuda:0'), tensor([-2.3961, -3.4373, -4.4814], device='cuda:0', dtype=torch.float64), 0, -100) 2025-12-04T09:29:33.3157767Z kwargs = {} 2025-12-04T09:29:33.3157861Z 2025-12-04T09:29:33.3158599Z Caused by sample input at index 15: SampleInput(input=Tensor[size=(2, 3, 2, 2), device="cuda:0", dtype=torch.float64], args=TensorList[Tensor[size=(2, 2, 2), device="cuda:0", dtype=torch.int64]], kwargs={'weight': 'Tensor[size=(3,), device="cuda:0", dtype=torch.float64]', 'reduction': "'none'"}, broadcasts_input=False, name='') 2025-12-04T09:29:33.3159405Z 2025-12-04T09:29:33.3159548Z To execute this test, run the following from the base repo dir: 2025-12-04T09:29:33.3160123Z PYTORCH_OPINFO_SAMPLE_INPUT_INDEX=15 python test/test_decomp.py TestDecompCUDA.test_comprehensive_nn_functional_nll_loss_cuda_float64 2025-12-04T09:29:33.3160544Z 2025-12-04T09:29:33.3160712Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:29:33.3161171Z ____ TestDecompCUDA.test_comprehensive_nn_functional_nll_loss_cuda_float64 _____ 2025-12-04T09:29:33.3161525Z Traceback (most recent call last): 2025-12-04T09:29:33.3162017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_device_type.py", line 1151, in test_wrapper 2025-12-04T09:29:33.3162512Z return test(*args, **kwargs) 2025-12-04T09:29:33.3162989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_device_type.py", line 1473, in only_fn 2025-12-04T09:29:33.3163468Z return fn(self, *args, **kwargs) 2025-12-04T09:29:33.3163916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 1647, in wrapper 2025-12-04T09:29:33.3164376Z fn(*args, **kwargs) 2025-12-04T09:29:33.3164798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 2464, in wrapper 2025-12-04T09:29:33.3165244Z fn(*args, **kwargs) 2025-12-04T09:29:33.3165559Z File "/var/lib/jenkins/workspace/test/test_decomp.py", line 602, in test_comprehensive 2025-12-04T09:29:33.3165959Z self.do_cross_ref(device, dtype, op, run_all=True) 2025-12-04T09:29:33.3166441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 1736, in wrapper 2025-12-04T09:29:33.3166882Z fn(*args, **kwargs) 2025-12-04T09:29:33.3167180Z File "/var/lib/jenkins/workspace/test/test_decomp.py", line 989, in do_cross_ref 2025-12-04T09:29:33.3167582Z decomp_out, decomp_vjp_fn = ref_vjp_no_create(fn, *primals) 2025-12-04T09:29:33.3167992Z File "/var/lib/jenkins/workspace/test/test_decomp.py", line 147, in ref_vjp_no_create 2025-12-04T09:29:33.3168394Z result = f(*primals) 2025-12-04T09:29:33.3168688Z File "/var/lib/jenkins/workspace/test/test_decomp.py", line 332, in wrapped 2025-12-04T09:29:33.3169048Z result = f(*_args, **kwargs) 2025-12-04T09:29:33.3169465Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/functional.py", line 3179, in nll_loss 2025-12-04T09:29:33.3169875Z return torch._C._nn.nll_loss_nd( 2025-12-04T09:29:33.3170224Z File "/var/lib/jenkins/workspace/test/test_decomp.py", line 904, in __torch_dispatch__ 2025-12-04T09:29:33.3170563Z op_assert_equal( 2025-12-04T09:29:33.3170866Z File "/var/lib/jenkins/workspace/test/test_decomp.py", line 301, in op_assert_equal 2025-12-04T09:29:33.3171216Z test_case.assertEqual( 2025-12-04T09:29:33.3171694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:29:33.3172227Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:29:33.3172519Z AssertionError: Scalars are not close! 2025-12-04T09:29:33.3172671Z 2025-12-04T09:29:33.3172754Z Expected 0.0 but got 10.757649417900529. 2025-12-04T09:29:33.3173037Z Absolute difference: 10.757649417900529 (up to 1e-07 allowed) 2025-12-04T09:29:33.3173433Z Relative difference: inf (up to 1e-07 allowed) 2025-12-04T09:29:33.3173677Z nll_loss2d_forward.default 2025-12-04T09:29:33.3173871Z args = (tensor([[[[-4.4896, 3.0513], 2025-12-04T09:29:33.3174071Z [-5.4787, 1.1693]], 2025-12-04T09:29:33.3174202Z 2025-12-04T09:29:33.3174272Z [[-3.6457, -3.2967], 2025-12-04T09:29:33.3174458Z [-4.9368, 6.3443]], 2025-12-04T09:29:33.3174571Z 2025-12-04T09:29:33.3174632Z [[-2.4919, -0.1966], 2025-12-04T09:29:33.3174822Z [ 4.2073, 7.7745]]], 2025-12-04T09:29:33.3174944Z 2025-12-04T09:29:33.3174947Z 2025-12-04T09:29:33.3175014Z [[[ 2.5429, 6.0313], 2025-12-04T09:29:33.3175193Z [ 3.0112, 4.1608]], 2025-12-04T09:29:33.3175309Z 2025-12-04T09:29:33.3175370Z [[ 7.4791, -2.5155], 2025-12-04T09:29:33.3175557Z [-0.7277, 7.4173]], 2025-12-04T09:29:33.3175678Z 2025-12-04T09:29:33.3175747Z [[-7.6495, -8.6749], 2025-12-04T09:29:33.3175976Z [ 7.0393, -4.4170]]]], device='cuda:0', dtype=torch.float64, 2025-12-04T09:29:33.3176254Z requires_grad=True), tensor([[[1, 0], 2025-12-04T09:29:33.3176470Z [1, 0]], 2025-12-04T09:29:33.3176566Z 2025-12-04T09:29:33.3176632Z [[0, 1], 2025-12-04T09:29:33.3176940Z [0, 0]]], device='cuda:0'), tensor([-2.3961, -3.4373, -4.4814], device='cuda:0', dtype=torch.float64), 0, -100) 2025-12-04T09:29:33.3177282Z kwargs = {} 2025-12-04T09:29:33.3177369Z 2025-12-04T09:29:33.3177530Z The above exception was the direct cause of the following exception: 2025-12-04T09:29:33.3177758Z 2025-12-04T09:29:33.3177836Z Traceback (most recent call last): 2025-12-04T09:29:33.3178299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:29:33.3178754Z method(*args, **kwargs) 2025-12-04T09:29:33.3179183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:29:33.3179628Z method(*args, **kwargs) 2025-12-04T09:29:33.3180101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_device_type.py", line 428, in instantiated_test 2025-12-04T09:29:33.3180614Z result = test(self, **param_kwargs) 2025-12-04T09:29:33.3181061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 1736, in wrapper 2025-12-04T09:29:33.3181513Z fn(*args, **kwargs) 2025-12-04T09:29:33.3181958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_device_type.py", line 1163, in test_wrapper 2025-12-04T09:29:33.3182500Z raise e_tracked from e 2025-12-04T09:29:33.3182692Z Exception: Scalars are not close! 2025-12-04T09:29:33.3182838Z 2025-12-04T09:29:33.3182917Z Expected 0.0 but got 10.757649417900529. 2025-12-04T09:29:33.3183231Z Absolute difference: 10.757649417900529 (up to 1e-07 allowed) 2025-12-04T09:29:33.3183522Z Relative difference: inf (up to 1e-07 allowed) 2025-12-04T09:29:33.3183796Z nll_loss2d_forward.default 2025-12-04T09:29:33.3183989Z args = (tensor([[[[-4.4896, 3.0513], 2025-12-04T09:29:33.3184183Z [-5.4787, 1.1693]], 2025-12-04T09:29:33.3184306Z 2025-12-04T09:29:33.3184368Z [[-3.6457, -3.2967], 2025-12-04T09:29:33.3184545Z [-4.9368, 6.3443]], 2025-12-04T09:29:33.3184659Z 2025-12-04T09:29:33.3184725Z [[-2.4919, -0.1966], 2025-12-04T09:29:33.3184896Z [ 4.2073, 7.7745]]], 2025-12-04T09:29:33.3185058Z 2025-12-04T09:29:33.3185062Z 2025-12-04T09:29:33.3185123Z [[[ 2.5429, 6.0313], 2025-12-04T09:29:33.3185298Z [ 3.0112, 4.1608]], 2025-12-04T09:29:33.3185412Z 2025-12-04T09:29:33.3185473Z [[ 7.4791, -2.5155], 2025-12-04T09:29:33.3185651Z [-0.7277, 7.4173]], 2025-12-04T09:29:33.3185768Z 2025-12-04T09:29:33.3185828Z [[-7.6495, -8.6749], 2025-12-04T09:29:33.3186059Z [ 7.0393, -4.4170]]]], device='cuda:0', dtype=torch.float64, 2025-12-04T09:29:33.3186331Z requires_grad=True), tensor([[[1, 0], 2025-12-04T09:29:33.3186542Z [1, 0]], 2025-12-04T09:29:33.3186636Z 2025-12-04T09:29:33.3186701Z [[0, 1], 2025-12-04T09:29:33.3186989Z [0, 0]]], device='cuda:0'), tensor([-2.3961, -3.4373, -4.4814], device='cuda:0', dtype=torch.float64), 0, -100) 2025-12-04T09:29:33.3187328Z kwargs = {} 2025-12-04T09:29:33.3187415Z 2025-12-04T09:29:33.3188183Z Caused by sample input at index 15: SampleInput(input=Tensor[size=(2, 3, 2, 2), device="cuda:0", dtype=torch.float64], args=TensorList[Tensor[size=(2, 2, 2), device="cuda:0", dtype=torch.int64]], kwargs={'weight': 'Tensor[size=(3,), device="cuda:0", dtype=torch.float64]', 'reduction': "'none'"}, broadcasts_input=False, name='') 2025-12-04T09:29:33.3189002Z 2025-12-04T09:29:33.3189137Z To execute this test, run the following from the base repo dir: 2025-12-04T09:29:33.3189695Z PYTORCH_OPINFO_SAMPLE_INPUT_INDEX=15 python test/test_decomp.py TestDecompCUDA.test_comprehensive_nn_functional_nll_loss_cuda_float64 2025-12-04T09:29:33.3190111Z 2025-12-04T09:29:33.3190290Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:29:33.3190897Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-a49ca6736f66875e.xml - 2025-12-04T09:29:33.3191643Z =========================== short test summary info ============================ 2025-12-04T09:29:33.3192207Z FAILED [1.4844s] test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 - Exception: Scalars are not close! 2025-12-04T09:29:33.3192621Z 2025-12-04T09:29:33.3192707Z Expected 0.0 but got 10.757649417900529. 2025-12-04T09:29:33.3193009Z Absolute difference: 10.757649417900529 (up to 1e-07 allowed) 2025-12-04T09:29:33.3193318Z Relative difference: inf (up to 1e-07 allowed) 2025-12-04T09:29:33.3193569Z nll_loss2d_forward.default 2025-12-04T09:29:33.3193765Z args = (tensor([[[[-4.4896, 3.0513], 2025-12-04T09:29:33.3193973Z [-5.4787, 1.1693]], 2025-12-04T09:29:33.3194098Z 2025-12-04T09:29:33.3194172Z [[-3.6457, -3.2967], 2025-12-04T09:29:33.3194361Z [-4.9368, 6.3443]], 2025-12-04T09:29:33.3194485Z 2025-12-04T09:29:33.3194546Z [[-2.4919, -0.1966], 2025-12-04T09:29:33.3194728Z [ 4.2073, 7.7745]]], 2025-12-04T09:29:33.3194851Z 2025-12-04T09:29:33.3194854Z 2025-12-04T09:29:33.3194921Z [[[ 2.5429, 6.0313], 2025-12-04T09:29:33.3195095Z [ 3.0112, 4.1608]], 2025-12-04T09:29:33.3195219Z 2025-12-04T09:29:33.3195281Z [[ 7.4791, -2.5155], 2025-12-04T09:29:33.3195466Z [-0.7277, 7.4173]], 2025-12-04T09:29:33.3195654Z 2025-12-04T09:29:33.3195738Z [[-7.6495, -8.6749], 2025-12-04T09:29:33.3196096Z [ 7.0393, -4.4170]]]], device='cuda:0', dtype=torch.float64, 2025-12-04T09:29:33.3196381Z requires_grad=True), tensor([[[1, 0], 2025-12-04T09:29:33.3196676Z [1, 0]], 2025-12-04T09:29:33.3196773Z 2025-12-04T09:29:33.3196841Z [[0, 1], 2025-12-04T09:29:33.3197369Z [0, 0]]], device='cuda:0'), tensor([-2.3961, -3.4373, -4.4814], device='cuda:0', dtype=torch.float64), 0, -100) 2025-12-04T09:29:33.3197980Z kwargs = {} 2025-12-04T09:29:33.3198127Z 2025-12-04T09:29:33.3199001Z Caused by sample input at index 15: SampleInput(input=Tensor[size=(2, 3, 2, 2), device="cuda:0", dtype=torch.float64], args=TensorList[Tensor[size=(2, 2, 2), device="cuda:0", dtype=torch.int64]], kwargs={'weight': 'Tensor[size=(3,), device="cuda:0", dtype=torch.float64]', 'reduction': "'none'"}, broadcasts_input=False, name='') 2025-12-04T09:29:33.3199823Z 2025-12-04T09:29:33.3199970Z To execute this test, run the following from the base repo dir: 2025-12-04T09:29:33.3200528Z PYTORCH_OPINFO_SAMPLE_INPUT_INDEX=15 python test/test_decomp.py TestDecompCUDA.test_comprehensive_nn_functional_nll_loss_cuda_float64 2025-12-04T09:29:33.3200955Z 2025-12-04T09:29:33.3201137Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:29:33.3201733Z FAILED [1.4553s] test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 - Exception: Scalars are not close! 2025-12-04T09:29:33.3202286Z 2025-12-04T09:29:33.3202419Z Expected 0.0 but got 10.757649417900529. 2025-12-04T09:29:33.3202811Z Absolute difference: 10.757649417900529 (up to 1e-07 allowed) 2025-12-04T09:29:33.3203118Z Relative difference: inf (up to 1e-07 allowed) 2025-12-04T09:29:33.3203374Z nll_loss2d_forward.default 2025-12-04T09:29:33.3203572Z args = (tensor([[[[-4.4896, 3.0513], 2025-12-04T09:29:33.3203778Z [-5.4787, 1.1693]], 2025-12-04T09:29:33.3203900Z 2025-12-04T09:29:33.3203972Z [[-3.6457, -3.2967], 2025-12-04T09:29:33.3204146Z [-4.9368, 6.3443]], 2025-12-04T09:29:33.3204274Z 2025-12-04T09:29:33.3204336Z [[-2.4919, -0.1966], 2025-12-04T09:29:33.3204705Z [ 4.2073, 7.7745]]], 2025-12-04T09:29:33.3204834Z 2025-12-04T09:29:33.3204837Z 2025-12-04T09:29:33.3204902Z [[[ 2.5429, 6.0313], 2025-12-04T09:29:33.3205076Z [ 3.0112, 4.1608]], 2025-12-04T09:29:33.3205206Z 2025-12-04T09:29:33.3205270Z [[ 7.4791, -2.5155], 2025-12-04T09:29:33.3205451Z [-0.7277, 7.4173]], 2025-12-04T09:29:33.3205563Z 2025-12-04T09:29:33.3205623Z [[-7.6495, -8.6749], 2025-12-04T09:29:33.3205856Z [ 7.0393, -4.4170]]]], device='cuda:0', dtype=torch.float64, 2025-12-04T09:29:33.3206130Z requires_grad=True), tensor([[[1, 0], 2025-12-04T09:29:33.3206340Z [1, 0]], 2025-12-04T09:29:33.3206439Z 2025-12-04T09:29:33.3206504Z [[0, 1], 2025-12-04T09:29:33.3206806Z [0, 0]]], device='cuda:0'), tensor([-2.3961, -3.4373, -4.4814], device='cuda:0', dtype=torch.float64), 0, -100) 2025-12-04T09:29:33.3207149Z kwargs = {} 2025-12-04T09:29:33.3207236Z 2025-12-04T09:29:33.3207967Z Caused by sample input at index 15: SampleInput(input=Tensor[size=(2, 3, 2, 2), device="cuda:0", dtype=torch.float64], args=TensorList[Tensor[size=(2, 2, 2), device="cuda:0", dtype=torch.int64]], kwargs={'weight': 'Tensor[size=(3,), device="cuda:0", dtype=torch.float64]', 'reduction': "'none'"}, broadcasts_input=False, name='') 2025-12-04T09:29:33.3208774Z 2025-12-04T09:29:33.3208920Z To execute this test, run the following from the base repo dir: 2025-12-04T09:29:33.3209489Z PYTORCH_OPINFO_SAMPLE_INPUT_INDEX=15 python test/test_decomp.py TestDecompCUDA.test_comprehensive_nn_functional_nll_loss_cuda_float64 2025-12-04T09:29:33.3209915Z 2025-12-04T09:29:33.3210084Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:29:33.3210451Z =================== 2 failed, 48 passed in 366.29s (0:06:06) =================== 2025-12-04T09:29:33.3210775Z 2025-12-04T09:29:33.3211024Z FINISHED PRINTING LOG FILE of test_decomp 18/22 (test/test-reports/test_decomp_18.22_6d2264971e288acd_.log) 2025-12-04T09:29:33.3211349Z 2025-12-04T09:29:33.3211594Z Finished test_decomp 18/22 ... [2025-12-04 09:29:33.305644][2209.265189964], took 6.21min 2025-12-04T09:29:33.3212295Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-a49ca6736f66875e.xml 2025-12-04T09:29:33.4155723Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-05457fa7ee6b6307.xml 2025-12-04T09:29:34.2866833Z Uploading logs for 57116084869 to S3 2025-12-04T09:29:34.3446153Z Uploading artifacts took 0.89 seconds 2025-12-04T09:29:34.3446741Z test_decomp 18/22 failed! 2025-12-04T09:29:34.4958312Z 2025-12-04T09:29:34.4959375Z inductor/test_torchinductor_opinfo 7/13 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_7.13_ec766f1d46ff24f6_.log 2025-12-04T09:29:34.4979850Z Running 50 items in this shard: test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64 2025-12-04T09:29:34.4998495Z 2025-12-04T09:29:34.4998757Z Finished inductor/test_torchinductor_opinfo 7/13 ... [2025-12-04 09:29:34.495750][2210.455292199], took 0.21min 2025-12-04T09:29:37.1125133Z Running inductor/test_torchinductor_opinfo 13/13 ... [2025-12-04 09:29:37.112026][2213.071569909] 2025-12-04T09:29:37.1125791Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:29:37.1127607Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '-m', 'not serial', '--shard-id=13', '--num-shards=13', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:29:37.112437] 2025-12-04T09:29:38.1104351Z Running inductor/test_layout_optim 1/1 ... [2025-12-04 09:29:38.109975][2214.069519064] 2025-12-04T09:29:38.1105372Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:29:38.1107163Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_layout_optim.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:29:38.110375] 2025-12-04T09:29:45.4783871Z 2025-12-04T09:29:45.4784920Z inductor/test_layout_optim 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_layout_optim_1.1_05e4855def927acc_.log 2025-12-04T09:29:45.4785975Z Running 0 items in this shard: 2025-12-04T09:29:45.4786152Z 2025-12-04T09:29:45.4786447Z Finished inductor/test_layout_optim 1/1 ... [2025-12-04 09:29:45.478229][2221.437775096], took 0.12min 2025-12-04T09:29:45.4913182Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_layout_optim/inductor.test_layout_optim-081f5af62216a830.xml 2025-12-04T09:29:49.0402517Z Running inductor/test_aot_inductor_arrayref 2/2 ... [2025-12-04 09:29:49.039784][2224.999326156] 2025-12-04T09:29:49.0403234Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:29:49.0409031Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_aot_inductor_arrayref.py', '-m', 'not serial', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:29:49.040543] 2025-12-04T09:29:49.9972336Z 2025-12-04T09:29:49.9973395Z inductor/test_torchinductor_opinfo 13/13 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_13.13_1a71b2cb0cb0ac84_.log 2025-12-04T09:29:49.9974083Z Running 0 items in this shard: 2025-12-04T09:29:49.9974241Z 2025-12-04T09:29:49.9974517Z Finished inductor/test_torchinductor_opinfo 13/13 ... [2025-12-04 09:29:49.997100][2225.956646694], took 0.21min 2025-12-04T09:29:50.0137161Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-6f7410ff779c9f27.xml 2025-12-04T09:29:53.6392619Z Running inductor/test_torchinductor_strided_blocks 1/1 ... [2025-12-04 09:29:53.638775][2229.598318806] 2025-12-04T09:29:53.6393476Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:29:53.6395046Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_strided_blocks.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:29:53.639188] 2025-12-04T09:29:55.9103132Z 2025-12-04T09:29:55.9104417Z inductor/test_aot_inductor_arrayref 2/2 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_aot_inductor_arrayref_2.2_ddb845b598b6b717_.log 2025-12-04T09:29:55.9105648Z Running 0 items in this shard: 2025-12-04T09:29:55.9105951Z 2025-12-04T09:29:55.9106388Z Finished inductor/test_aot_inductor_arrayref 2/2 ... [2025-12-04 09:29:55.910178][2231.869724733], took 0.11min 2025-12-04T09:29:55.9231420Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor_arrayref/inductor.test_aot_inductor_arrayref-51f2220a835161fe.xml 2025-12-04T09:29:59.4886486Z Running test_custom_ops 1/1 ... [2025-12-04 09:29:59.488214][2235.44775842] 2025-12-04T09:29:59.4886962Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:29:59.4889976Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_custom_ops.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:29:59.488666] 2025-12-04T09:30:04.4740650Z 2025-12-04T09:30:04.4741716Z test_custom_ops 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_custom_ops_1.1_80f9a34920161ad5_.log 2025-12-04T09:30:04.4758290Z Running 50 items in this shard: test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python, test/test_custom_ops.py::TestCustomOp::test_defined_in_python 2025-12-04T09:30:04.4773300Z 2025-12-04T09:30:04.4773656Z Finished test_custom_ops 1/1 ... [2025-12-04 09:30:04.474067][2240.433609385], took 0.08min 2025-12-04T09:30:04.4873983Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_custom_ops/test_custom_ops-c87b73c0dc85593b.xml 2025-12-04T09:30:08.1566397Z Running test_content_store 1/1 ... [2025-12-04 09:30:08.156223][2244.115766402] 2025-12-04T09:30:08.1566983Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:30:08.1569300Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_content_store.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:30:08.156649] 2025-12-04T09:30:11.7572371Z 2025-12-04T09:30:11.7573430Z test_content_store 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_content_store_1.1_8b370b99c7663fbf_.log 2025-12-04T09:30:11.7574263Z Running 0 items in this shard: 2025-12-04T09:30:11.7574487Z 2025-12-04T09:30:11.7574744Z Finished test_content_store 1/1 ... [2025-12-04 09:30:11.757169][2247.716713281], took 0.06min 2025-12-04T09:30:11.7707698Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_content_store/test_content_store-a61adca807bbc325.xml 2025-12-04T09:30:15.4595695Z Running inductor/test_flex_decoding 1/3 ... [2025-12-04 09:30:15.459149][2251.41869287] 2025-12-04T09:30:15.4596606Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:30:15.4598478Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_flex_decoding.py', '-m', 'not serial', '--shard-id=1', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:30:15.459567] 2025-12-04T09:30:19.8380793Z 2025-12-04T09:30:19.8381966Z inductor/test_flex_decoding 1/3 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_flex_decoding_1.3_d0db0cd9a9640154_.log 2025-12-04T09:30:19.8382628Z Running 0 items in this shard: 2025-12-04T09:30:19.8382778Z 2025-12-04T09:30:19.8383011Z Finished inductor/test_flex_decoding 1/3 ... [2025-12-04 09:30:19.837985][2255.797530398], took 0.07min 2025-12-04T09:30:19.8592331Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_flex_decoding/inductor.test_flex_decoding-59c9d120af700297.xml 2025-12-04T09:30:23.4145443Z Running inductor/test_flex_decoding 3/3 ... [2025-12-04 09:30:23.414132][2259.373675544] 2025-12-04T09:30:23.4146158Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:30:23.4147628Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_flex_decoding.py', '-m', 'not serial', '--shard-id=3', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:30:23.414496] 2025-12-04T09:30:27.7116181Z 2025-12-04T09:30:27.7117139Z inductor/test_flex_decoding 3/3 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_flex_decoding_3.3_cda88ebe36ff4298_.log 2025-12-04T09:30:27.7117932Z Running 0 items in this shard: 2025-12-04T09:30:27.7118103Z 2025-12-04T09:30:27.7118413Z Finished inductor/test_flex_decoding 3/3 ... [2025-12-04 09:30:27.711452][2263.670997823], took 0.07min 2025-12-04T09:30:27.7251381Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_flex_decoding/inductor.test_flex_decoding-4bc3b9a2985b9ca4.xml 2025-12-04T09:30:31.3561937Z Running inductor/test_deterministic 3/3 ... [2025-12-04 09:30:31.355718][2267.315259821] 2025-12-04T09:30:31.3562572Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:30:31.3565163Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_deterministic.py', '-m', 'not serial', '--shard-id=3', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:30:31.356166] 2025-12-04T09:30:33.9027014Z 2025-12-04T09:30:33.9027901Z inductor/test_torchinductor_strided_blocks 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_strided_blocks_1.1_e98c1bd43c2c27cb_.log 2025-12-04T09:30:33.9107982Z Running 200 items in this shard: test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestCPU::test_2d_reductions_mixed_indexing_reduction_op0_cpu, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonBlockPointerTestGPU::test_broadcast_with_singleton_dims_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda, test/inductor/test_torchinductor_strided_blocks.py::TritonTensorDescriptorTestCUDA::test_2d_reductions_mixed_indexing_reduction_op0_cuda 2025-12-04T09:30:33.9183113Z 2025-12-04T09:30:33.9183417Z Finished inductor/test_torchinductor_strided_blocks 1/1 ... [2025-12-04 09:30:33.903190][2269.862735233], took 0.67min 2025-12-04T09:30:33.9212331Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_strided_blocks/inductor.test_torchinductor_strided_blocks-9737d64a17925152.xml 2025-12-04T09:30:37.4401985Z Running inductor/test_b2b_gemm 1/1 ... [2025-12-04 09:30:37.439716][2273.39925998] 2025-12-04T09:30:37.4402507Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:30:37.4404726Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_b2b_gemm.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:30:37.440157] 2025-12-04T09:30:37.6333878Z 2025-12-04T09:30:37.6334786Z inductor/test_deterministic 3/3 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_deterministic_3.3_49cc6b9d77892b3f_.log 2025-12-04T09:30:37.6335629Z Running 0 items in this shard: 2025-12-04T09:30:37.6335928Z 2025-12-04T09:30:37.6336445Z Finished inductor/test_deterministic 3/3 ... [2025-12-04 09:30:37.633247][2273.592792712], took 0.10min 2025-12-04T09:30:37.6475345Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_deterministic/inductor.test_deterministic-46ebc55a5876c541.xml 2025-12-04T09:30:41.2029470Z Running export/test_tree_utils 1/1 ... [2025-12-04 09:30:41.202497][2277.162040626] 2025-12-04T09:30:41.2030043Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:30:41.2031973Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'export/test_tree_utils.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:30:41.202913] 2025-12-04T09:30:43.5236323Z 2025-12-04T09:30:43.5237592Z inductor/test_b2b_gemm 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_b2b_gemm_1.1_e64cb8635d438c4d_.log 2025-12-04T09:30:43.5238265Z Running 0 items in this shard: 2025-12-04T09:30:43.5238414Z 2025-12-04T09:30:43.5238649Z Finished inductor/test_b2b_gemm 1/1 ... [2025-12-04 09:30:43.523484][2279.483030002], took 0.10min 2025-12-04T09:30:43.5378315Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_b2b_gemm/inductor.test_b2b_gemm-f3a614691e2e3a64.xml 2025-12-04T09:30:44.4592304Z 2025-12-04T09:30:44.4593406Z export/test_tree_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_tree_utils_1.1_a150d77e4196bf39_.log 2025-12-04T09:30:44.4594714Z Running 0 items in this shard: 2025-12-04T09:30:44.4594903Z 2025-12-04T09:30:44.4595164Z Finished export/test_tree_utils 1/1 ... [2025-12-04 09:30:44.459102][2280.418648492], took 0.05min 2025-12-04T09:30:44.4856358Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/export.test_tree_utils/export.test_tree_utils-6c4b1c3428f21437.xml 2025-12-04T09:30:47.1473953Z Running inductor/test_triton_wrapper 1/1 ... [2025-12-04 09:30:47.146952][2283.106496316] 2025-12-04T09:30:47.1474781Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:30:47.1476552Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_triton_wrapper.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:30:47.147361] 2025-12-04T09:30:48.0767191Z Running inductor/test_static_cuda_launcher 1/1 ... [2025-12-04 09:30:48.076220][2284.035763991] 2025-12-04T09:30:48.0767767Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:30:48.0769365Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_static_cuda_launcher.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:30:48.076645] 2025-12-04T09:30:53.1357108Z 2025-12-04T09:30:53.1358205Z inductor/test_triton_wrapper 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_triton_wrapper_1.1_d176b02f5f219154_.log 2025-12-04T09:30:53.1359429Z Running 0 items in this shard: 2025-12-04T09:30:53.1359721Z 2025-12-04T09:30:53.1360186Z Finished inductor/test_triton_wrapper 1/1 ... [2025-12-04 09:30:53.135644][2289.095186484], took 0.10min 2025-12-04T09:30:53.1505779Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_triton_wrapper/inductor.test_triton_wrapper-24ddba239b87c5fd.xml 2025-12-04T09:30:54.7316532Z 2025-12-04T09:30:54.7317839Z inductor/test_static_cuda_launcher 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_static_cuda_launcher_1.1_9b61891d2c79064c_.log 2025-12-04T09:30:54.7318715Z Running 0 items in this shard: 2025-12-04T09:30:54.7318895Z 2025-12-04T09:30:54.7319208Z Finished inductor/test_static_cuda_launcher 1/1 ... [2025-12-04 09:30:54.731483][2290.691028823], took 0.11min 2025-12-04T09:30:54.7468904Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_static_cuda_launcher/inductor.test_static_cuda_launcher-add7e1c387913324.xml 2025-12-04T09:30:56.8316653Z Running export/test_dynamic_shapes 1/1 ... [2025-12-04 09:30:56.831191][2292.790733528] 2025-12-04T09:30:56.8317191Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:30:56.8319914Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'export/test_dynamic_shapes.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:30:56.831672] 2025-12-04T09:30:58.3324899Z Running dynamo/test_sdpa 1/1 ... [2025-12-04 09:30:58.332057][2294.291600452] 2025-12-04T09:30:58.3325591Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:30:58.3327782Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_sdpa.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:30:58.332463] 2025-12-04T09:31:00.0326526Z 2025-12-04T09:31:00.0327462Z export/test_dynamic_shapes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_dynamic_shapes_1.1_0acf0e3fcb724802_.log 2025-12-04T09:31:00.0328206Z Running 0 items in this shard: 2025-12-04T09:31:00.0328628Z 2025-12-04T09:31:00.0328908Z Finished export/test_dynamic_shapes 1/1 ... [2025-12-04 09:31:00.032494][2295.992041622], took 0.05min 2025-12-04T09:31:00.0533419Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/export.test_dynamic_shapes/export.test_dynamic_shapes-80cb687c234e2e79.xml 2025-12-04T09:31:01.9559591Z 2025-12-04T09:31:01.9560947Z dynamo/test_sdpa 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_sdpa_1.1_a67282d1f1c8872a_.log 2025-12-04T09:31:01.9561805Z Running 0 items in this shard: 2025-12-04T09:31:01.9562090Z 2025-12-04T09:31:01.9562473Z Finished dynamo/test_sdpa 1/1 ... [2025-12-04 09:31:01.955871][2297.915416321], took 0.06min 2025-12-04T09:31:01.9722411Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_sdpa/dynamo.test_sdpa-49a3e779aba88f83.xml 2025-12-04T09:31:03.6653098Z Running inductor/test_aot_inductor_package 1/1 ... [2025-12-04 09:31:03.664805][2299.624349651] 2025-12-04T09:31:03.6654147Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:03.6656149Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_aot_inductor_package.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:03.665227] 2025-12-04T09:31:05.5006073Z Running inductor/test_compiled_optimizers 3/3 ... [2025-12-04 09:31:05.500089][2301.459633448] 2025-12-04T09:31:05.5006628Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:05.5008230Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_compiled_optimizers.py', '-m', 'not serial', '--shard-id=3', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:05.500514] 2025-12-04T09:31:09.7096810Z 2025-12-04T09:31:09.7097963Z inductor/test_aot_inductor_package 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_aot_inductor_package_1.1_e6c86135fe5c29f7_.log 2025-12-04T09:31:09.7098823Z Running 0 items in this shard: 2025-12-04T09:31:09.7099009Z 2025-12-04T09:31:09.7099332Z Finished inductor/test_aot_inductor_package 1/1 ... [2025-12-04 09:31:09.709532][2305.669075012], took 0.10min 2025-12-04T09:31:09.7253617Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor_package/inductor.test_aot_inductor_package-b829514935b5c71f.xml 2025-12-04T09:31:13.3893802Z Running inductor/test_aot_inductor_utils 1/1 ... [2025-12-04 09:31:13.388921][2309.34846326] 2025-12-04T09:31:13.3894370Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:13.3896611Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_aot_inductor_utils.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:13.389355] 2025-12-04T09:31:14.0148397Z 2025-12-04T09:31:14.0149367Z inductor/test_compiled_optimizers 3/3 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_compiled_optimizers_3.3_a39d9905fe0a7e96_.log 2025-12-04T09:31:14.0150202Z Running 0 items in this shard: 2025-12-04T09:31:14.0150376Z 2025-12-04T09:31:14.0150703Z Finished inductor/test_compiled_optimizers 3/3 ... [2025-12-04 09:31:14.014644][2309.97419064], took 0.14min 2025-12-04T09:31:14.0401000Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_compiled_optimizers/inductor.test_compiled_optimizers-bccb353119aa7e57.xml 2025-12-04T09:31:17.4940364Z Running inductor/test_control_flow 3/5 ... [2025-12-04 09:31:17.493603][2313.453146866] 2025-12-04T09:31:17.4941063Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:17.4942961Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_control_flow.py', '-m', 'not serial', '--shard-id=3', '--num-shards=5', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:17.494014] 2025-12-04T09:31:19.4010416Z 2025-12-04T09:31:19.4011439Z inductor/test_aot_inductor_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_aot_inductor_utils_1.1_e3aa7bd30346a26c_.log 2025-12-04T09:31:19.4012259Z Running 0 items in this shard: 2025-12-04T09:31:19.4012435Z 2025-12-04T09:31:19.4013027Z Finished inductor/test_aot_inductor_utils 1/1 ... [2025-12-04 09:31:19.400878][2315.360422306], took 0.10min 2025-12-04T09:31:19.4267814Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor_utils/inductor.test_aot_inductor_utils-d291c39c695faa5d.xml 2025-12-04T09:31:23.0278958Z Running test_mkl_verbose 1/1 ... [2025-12-04 09:31:23.027436][2318.98698074] 2025-12-04T09:31:23.0279497Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:23.0282439Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_mkl_verbose.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:23.027889] 2025-12-04T09:31:24.7613037Z 2025-12-04T09:31:24.7614278Z inductor/test_control_flow 3/5 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_control_flow_3.5_00fe48f3237e41c2_.log 2025-12-04T09:31:24.7615289Z Running 0 items in this shard: 2025-12-04T09:31:24.7615435Z 2025-12-04T09:31:24.7615665Z Finished inductor/test_control_flow 3/5 ... [2025-12-04 09:31:24.761200][2320.720742136], took 0.12min 2025-12-04T09:31:24.7865373Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_control_flow/inductor.test_control_flow-11a5ec05df23ef24.xml 2025-12-04T09:31:26.3197316Z 2025-12-04T09:31:26.3198409Z test_mkl_verbose 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_mkl_verbose_1.1_3691733155bb2907_.log 2025-12-04T09:31:26.3199090Z Running 0 items in this shard: 2025-12-04T09:31:26.3199278Z 2025-12-04T09:31:26.3199534Z Finished test_mkl_verbose 1/1 ... [2025-12-04 09:31:26.319628][2322.279174313], took 0.05min 2025-12-04T09:31:26.3422447Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_mkl_verbose/test_mkl_verbose-f26bddd178ca728a.xml 2025-12-04T09:31:28.3646604Z Running test_comparison_utils 1/1 ... [2025-12-04 09:31:28.364163][2324.323707909] 2025-12-04T09:31:28.3647084Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:28.3649131Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_comparison_utils.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:28.364624] 2025-12-04T09:31:29.8722269Z Running functorch/test_ac_logging 1/1 ... [2025-12-04 09:31:29.871771][2325.831315986] 2025-12-04T09:31:29.8722769Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:29.8724952Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_ac_logging.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:29.872190] 2025-12-04T09:31:31.6160553Z 2025-12-04T09:31:31.6161557Z test_comparison_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_comparison_utils_1.1_6dc4c17d20e9f5c0_.log 2025-12-04T09:31:31.6162624Z Running 0 items in this shard: 2025-12-04T09:31:31.6162810Z 2025-12-04T09:31:31.6163238Z Finished test_comparison_utils 1/1 ... [2025-12-04 09:31:31.615898][2327.575444698], took 0.05min 2025-12-04T09:31:31.6425656Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_comparison_utils/test_comparison_utils-d3a1a704cb9d5665.xml 2025-12-04T09:31:33.1158051Z 2025-12-04T09:31:33.1158972Z functorch/test_ac_logging 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_ac_logging_1.1_6ca2b350550f1b46_.log 2025-12-04T09:31:33.1159760Z Running 0 items in this shard: 2025-12-04T09:31:33.1159957Z 2025-12-04T09:31:33.1160526Z Finished functorch/test_ac_logging 1/1 ... [2025-12-04 09:31:33.115569][2329.075115731], took 0.05min 2025-12-04T09:31:33.1329775Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_ac_logging/functorch.test_ac_logging-161e2c94aebfad09.xml 2025-12-04T09:31:35.2089497Z Running test_mkldnn_verbose 1/1 ... [2025-12-04 09:31:35.208498][2331.168042398] 2025-12-04T09:31:35.2090162Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:35.2092532Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_mkldnn_verbose.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:35.208915] 2025-12-04T09:31:36.6606012Z Running test_utils_config_module 1/1 ... [2025-12-04 09:31:36.660110][2332.61965384] 2025-12-04T09:31:36.6606488Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:36.6609041Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_utils_config_module.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:36.660591] 2025-12-04T09:31:38.4889015Z 2025-12-04T09:31:38.4889771Z test_mkldnn_verbose 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_mkldnn_verbose_1.1_c89e68a75cbfee9d_.log 2025-12-04T09:31:38.4890408Z Running 0 items in this shard: 2025-12-04T09:31:38.4890555Z 2025-12-04T09:31:38.4890764Z Finished test_mkldnn_verbose 1/1 ... [2025-12-04 09:31:38.488779][2334.448324277], took 0.05min 2025-12-04T09:31:38.5065886Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_mkldnn_verbose/test_mkldnn_verbose-43202a21dc76fbd5.xml 2025-12-04T09:31:39.9432607Z 2025-12-04T09:31:39.9433585Z test_utils_config_module 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_utils_config_module_1.1_6e4b2011ed24ec12_.log 2025-12-04T09:31:39.9434331Z Running 0 items in this shard: 2025-12-04T09:31:39.9434509Z 2025-12-04T09:31:39.9434777Z Finished test_utils_config_module 1/1 ... [2025-12-04 09:31:39.943114][2335.902661275], took 0.05min 2025-12-04T09:31:39.9604999Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_utils_config_module/test_utils_config_module-262aead0bb49178e.xml 2025-12-04T09:31:41.9877179Z Running test_hop_infra 1/1 ... [2025-12-04 09:31:41.987227][2337.946768478] 2025-12-04T09:31:41.9877782Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:41.9880011Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_hop_infra.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:41.987668] 2025-12-04T09:31:43.4844222Z Running test_appending_byte_serializer 1/1 ... [2025-12-04 09:31:43.483910][2339.443453482] 2025-12-04T09:31:43.4845187Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:43.4846788Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_appending_byte_serializer.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:43.484321] 2025-12-04T09:31:46.0727693Z 2025-12-04T09:31:46.0728423Z test_hop_infra 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_hop_infra_1.1_3ccb8c957d7bb463_.log 2025-12-04T09:31:46.0728991Z Running 0 items in this shard: 2025-12-04T09:31:46.0729132Z 2025-12-04T09:31:46.0729616Z Finished test_hop_infra 1/1 ... [2025-12-04 09:31:46.072648][2342.032191638], took 0.07min 2025-12-04T09:31:46.0991064Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_hop_infra/test_hop_infra-d73563c119a036ea.xml 2025-12-04T09:31:46.7711744Z 2025-12-04T09:31:46.7712915Z test_appending_byte_serializer 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_appending_byte_serializer_1.1_38e7f5cc083c9bad_.log 2025-12-04T09:31:46.7713805Z Running 0 items in this shard: 2025-12-04T09:31:46.7713984Z 2025-12-04T09:31:46.7714301Z Finished test_appending_byte_serializer 1/1 ... [2025-12-04 09:31:46.770999][2342.730543337], took 0.05min 2025-12-04T09:31:46.8266914Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_appending_byte_serializer/test_appending_byte_serializer-4e76a1b178e5d348.xml 2025-12-04T09:31:49.6419566Z Running test_license 1/1 ... [2025-12-04 09:31:49.641517][2345.60106103] 2025-12-04T09:31:49.6420047Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:49.6422554Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_license.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:49.641948] 2025-12-04T09:31:50.3412107Z Running test_ao_sparsity 1/1 ... [2025-12-04 09:31:50.340743][2346.300286701] 2025-12-04T09:31:50.3412606Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:50.3414945Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ao_sparsity.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:50.341167] 2025-12-04T09:31:52.8314769Z 2025-12-04T09:31:52.8315607Z test_license 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_license_1.1_b93bab9da6159df7_.log 2025-12-04T09:31:52.8316264Z Running 0 items in this shard: 2025-12-04T09:31:52.8316455Z 2025-12-04T09:31:52.8316680Z Finished test_license 1/1 ... [2025-12-04 09:31:52.831392][2348.790937313], took 0.05min 2025-12-04T09:31:52.8506637Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_license/test_license-1d2b2367c347e96e.xml 2025-12-04T09:31:54.1590282Z 2025-12-04T09:31:54.1591387Z test_ao_sparsity 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_ao_sparsity_1.1_baacf12029e08a3c_.log 2025-12-04T09:31:54.1591986Z Running 0 items in this shard: 2025-12-04T09:31:54.1592153Z 2025-12-04T09:31:54.1592357Z Finished test_ao_sparsity 1/1 ... [2025-12-04 09:31:54.158871][2350.118414491], took 0.06min 2025-12-04T09:31:54.1783689Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ao_sparsity/test_ao_sparsity-08192502e661e963.xml 2025-12-04T09:31:56.5066932Z Running test_autoload 1/1 ... [2025-12-04 09:31:56.506257][2352.465800067] 2025-12-04T09:31:56.5067472Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:56.5069896Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_autoload.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:56.506667] 2025-12-04T09:31:57.6959466Z Running nn/attention/test_open_registry 1/1 ... [2025-12-04 09:31:57.695474][2353.65501722] 2025-12-04T09:31:57.6960193Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:57.6962388Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/attention/test_open_registry.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:57.695870] 2025-12-04T09:31:59.7848163Z 2025-12-04T09:31:59.7849215Z test_autoload 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_autoload_1.1_bf261471dad7ddf1_.log 2025-12-04T09:31:59.7850094Z Running 0 items in this shard: 2025-12-04T09:31:59.7850277Z 2025-12-04T09:31:59.7850621Z Finished test_autoload 1/1 ... [2025-12-04 09:31:59.784720][2355.744265047], took 0.05min 2025-12-04T09:31:59.8036112Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_autoload/test_autoload-f8c034a1e389653e.xml 2025-12-04T09:32:00.8939534Z 2025-12-04T09:32:00.8940436Z nn/attention/test_open_registry 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.attention.test_open_registry_1.1_c0a1d8e6882b9f0d_.log 2025-12-04T09:32:00.8941143Z Running 0 items in this shard: 2025-12-04T09:32:00.8941342Z 2025-12-04T09:32:00.8941642Z Finished nn/attention/test_open_registry 1/1 ... [2025-12-04 09:32:00.893823][2356.853369139], took 0.05min 2025-12-04T09:32:00.9127598Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.attention.test_open_registry/nn.attention.test_open_registry-7b14a30ba5e47451.xml 2025-12-04T09:32:03.4629019Z Running test_as_strided 1/1 ... [2025-12-04 09:32:03.462457][2359.422000863] 2025-12-04T09:32:03.4629528Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:32:03.4631616Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_as_strided.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:32:03.462864] 2025-12-04T09:32:04.5512816Z Running test_foreach 1/1 ... [2025-12-04 09:32:04.550829][2360.510372378] 2025-12-04T09:32:04.5513390Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:32:04.5515565Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_foreach.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:32:04.551249] 2025-12-04T09:32:06.6035690Z 2025-12-04T09:32:06.6036752Z test_as_strided 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_as_strided_1.1_5667cb6d1805af19_.log 2025-12-04T09:32:06.6037556Z Running 0 items in this shard: 2025-12-04T09:32:06.6037708Z 2025-12-04T09:32:06.6037890Z Finished test_as_strided 1/1 ... [2025-12-04 09:32:06.603439][2362.562984847], took 0.05min 2025-12-04T09:32:06.6228747Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_as_strided/test_as_strided-902df6e2e575efe4.xml 2025-12-04T09:32:10.1937455Z Running xpu/test_gemm 1/1 ... [2025-12-04 09:32:10.193244][2366.152787509] 2025-12-04T09:32:10.1938080Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:32:10.1939485Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'xpu/test_gemm.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:32:10.193656] 2025-12-04T09:32:14.0145436Z 2025-12-04T09:32:14.0146726Z xpu/test_gemm 1/1 was successful, full logs can be found in artifacts with path test/test-reports/xpu.test_gemm_1.1_5c80f370f4543dcf_.log 2025-12-04T09:32:14.0147563Z Running 0 items in this shard: 2025-12-04T09:32:14.0147793Z 2025-12-04T09:32:14.0148085Z Finished xpu/test_gemm 1/1 ... [2025-12-04 09:32:14.014438][2369.97398174], took 0.06min 2025-12-04T09:32:14.0341829Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/xpu.test_gemm/xpu.test_gemm-3a4dcb0c73b9942b.xml 2025-12-04T09:32:17.5729157Z Running higher_order_ops/test_print 1/1 ... [2025-12-04 09:32:17.572375][2373.531917469] 2025-12-04T09:32:17.5729799Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:32:17.5731053Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'higher_order_ops/test_print.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:32:17.572809] 2025-12-04T09:32:20.9639573Z 2025-12-04T09:32:20.9640524Z higher_order_ops/test_print 1/1 was successful, full logs can be found in artifacts with path test/test-reports/higher_order_ops.test_print_1.1_56d44039c67711d6_.log 2025-12-04T09:32:20.9641355Z Running 0 items in this shard: 2025-12-04T09:32:20.9641539Z 2025-12-04T09:32:20.9641828Z Finished higher_order_ops/test_print 1/1 ... [2025-12-04 09:32:20.963901][2376.923447358], took 0.06min 2025-12-04T09:32:20.9833603Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/higher_order_ops.test_print/higher_order_ops.test_print-8bf101c57cf0f824.xml 2025-12-04T09:32:22.7050066Z 2025-12-04T09:32:22.7050862Z test_foreach 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_foreach_1.1_216741b645f3a84c_.log 2025-12-04T09:32:22.7164699Z Running 350 items in this shard: test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16 2025-12-04T09:32:22.7271373Z 2025-12-04T09:32:22.7271612Z Finished test_foreach 1/1 ... [2025-12-04 09:32:22.705574][2378.665119794], took 0.30min 2025-12-04T09:32:22.7317762Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_foreach/test_foreach-014b7fdc7e3e1cf2.xml 2025-12-04T09:32:24.6316186Z Running test_per_overload_api 1/1 ... [2025-12-04 09:32:24.631138][2380.590681908] 2025-12-04T09:32:24.6316724Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:32:24.6319202Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_per_overload_api.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:32:24.631592] 2025-12-04T09:32:26.3146819Z Running torch_np/numpy_tests/core/test_einsum 1/1 ... [2025-12-04 09:32:26.314218][2382.273761703] 2025-12-04T09:32:26.3147375Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:32:26.3150011Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_einsum.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:32:26.314685] 2025-12-04T09:32:27.8167309Z 2025-12-04T09:32:27.8168447Z test_per_overload_api 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_per_overload_api_1.1_d71ebb1277eaeb76_.log 2025-12-04T09:32:27.8169079Z Running 0 items in this shard: 2025-12-04T09:32:27.8169219Z 2025-12-04T09:32:27.8169421Z Finished test_per_overload_api 1/1 ... [2025-12-04 09:32:27.816640][2383.776184412], took 0.05min 2025-12-04T09:32:27.8369668Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_per_overload_api/test_per_overload_api-7ad42b77bba6a3a2.xml 2025-12-04T09:32:29.5622668Z 2025-12-04T09:32:29.5623907Z torch_np/numpy_tests/core/test_einsum 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_einsum_1.1_55ea97b8b3e11159_.log 2025-12-04T09:32:29.5625242Z Running 0 items in this shard: 2025-12-04T09:32:29.5625539Z 2025-12-04T09:32:29.5626126Z Finished torch_np/numpy_tests/core/test_einsum 1/1 ... [2025-12-04 09:32:29.562135][2385.521680523], took 0.05min 2025-12-04T09:32:29.5858287Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.core.test_einsum/torch_np.numpy_tests.core.test_einsum-7f41ab7e4542d2ff.xml 2025-12-04T09:32:31.5087585Z Running test_out_dtype_op 1/1 ... [2025-12-04 09:32:31.508280][2387.467822068] 2025-12-04T09:32:31.5088358Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:32:31.5090381Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_out_dtype_op.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:32:31.508751] 2025-12-04T09:32:33.1456016Z Running torch_np/test_ufuncs_basic 1/1 ... [2025-12-04 09:32:33.145134][2389.104677683] 2025-12-04T09:32:33.1456523Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:32:33.1458575Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/test_ufuncs_basic.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:32:33.145543] 2025-12-04T09:32:35.3882738Z 2025-12-04T09:32:35.3883741Z test_out_dtype_op 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_out_dtype_op_1.1_0c60bd107333eda1_.log 2025-12-04T09:32:35.3884337Z Running 0 items in this shard: 2025-12-04T09:32:35.3884495Z 2025-12-04T09:32:35.3884691Z Finished test_out_dtype_op 1/1 ... [2025-12-04 09:32:35.388113][2391.347655276], took 0.06min 2025-12-04T09:32:35.4091704Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_out_dtype_op/test_out_dtype_op-00e91616e649df2c.xml 2025-12-04T09:32:36.9136588Z 2025-12-04T09:32:36.9137579Z torch_np/test_ufuncs_basic 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.test_ufuncs_basic_1.1_cd7199ad4e0a518a_.log 2025-12-04T09:32:36.9138474Z Running 0 items in this shard: 2025-12-04T09:32:36.9138768Z 2025-12-04T09:32:36.9139102Z Finished torch_np/test_ufuncs_basic 1/1 ... [2025-12-04 09:32:36.913565][2392.873109987], took 0.06min 2025-12-04T09:32:36.9345386Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.test_ufuncs_basic/torch_np.test_ufuncs_basic-c8fa7ec0bb170cfd.xml 2025-12-04T09:32:38.9581173Z Running lazy/test_step_closures 1/1 ... [2025-12-04 09:32:38.957629][2394.917173346] 2025-12-04T09:32:38.9581698Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:32:38.9583402Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'lazy/test_step_closures.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:32:38.958031] 2025-12-04T09:32:40.4488135Z Running functorch/dim/test_getsetitem 1/1 ... [2025-12-04 09:32:40.448334][2396.407877625] 2025-12-04T09:32:40.4488885Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:32:40.4490351Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/dim/test_getsetitem.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:32:40.448715] 2025-12-04T09:32:42.2467420Z 2025-12-04T09:32:42.2468592Z lazy/test_step_closures 1/1 was successful, full logs can be found in artifacts with path test/test-reports/lazy.test_step_closures_1.1_9ba2d3a408de6eb8_.log 2025-12-04T09:32:42.2469761Z Running 0 items in this shard: 2025-12-04T09:32:42.2470022Z 2025-12-04T09:32:42.2470409Z Finished lazy/test_step_closures 1/1 ... [2025-12-04 09:32:42.246646][2398.206191336], took 0.05min 2025-12-04T09:32:42.2679543Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/lazy.test_step_closures/lazy.test_step_closures-ed58f93fb889064c.xml 2025-12-04T09:32:43.7549661Z 2025-12-04T09:32:43.7550839Z functorch/dim/test_getsetitem 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.dim.test_getsetitem_1.1_189c9e5b965e450c_.log 2025-12-04T09:32:43.7552110Z Running 0 items in this shard: 2025-12-04T09:32:43.7552352Z 2025-12-04T09:32:43.7552744Z Finished functorch/dim/test_getsetitem 1/1 ... [2025-12-04 09:32:43.754777][2399.7143245], took 0.06min 2025-12-04T09:32:43.7763928Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.dim.test_getsetitem/functorch.dim.test_getsetitem-674883abaf9ddbc7.xml 2025-12-04T09:32:45.8771819Z Running test_ops_fwd_gradients 1/3 ... [2025-12-04 09:32:45.876717][2401.836259223] 2025-12-04T09:32:45.8772385Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:32:45.8775133Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops_fwd_gradients.py', '-m', 'not serial', '--shard-id=1', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:32:45.877127] 2025-12-04T09:32:47.3550674Z Running test_ops_fwd_gradients 2/3 ... [2025-12-04 09:32:47.354618][2403.314162668] 2025-12-04T09:32:47.3551190Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:32:47.3554041Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops_fwd_gradients.py', '-m', 'not serial', '--shard-id=2', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:32:47.355080] 2025-12-04T09:32:55.1783893Z 2025-12-04T09:32:55.1784930Z test_ops_fwd_gradients 1/3 was successful, full logs can be found in artifacts with path test/test-reports/test_ops_fwd_gradients_1.3_86a0d9916377e0c1_.log 2025-12-04T09:32:55.1800043Z Running 50 items in this shard: test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_float64 2025-12-04T09:32:55.1814761Z 2025-12-04T09:32:55.1814991Z Finished test_ops_fwd_gradients 1/3 ... [2025-12-04 09:32:55.178453][2411.137995847], took 0.16min 2025-12-04T09:32:55.1999188Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-5472f3d31fbed190.xml 2025-12-04T09:32:56.6987783Z 2025-12-04T09:32:56.6988921Z test_ops_fwd_gradients 2/3 was successful, full logs can be found in artifacts with path test/test-reports/test_ops_fwd_gradients_2.3_24e5b94522df8eef_.log 2025-12-04T09:32:56.7005610Z Running 50 items in this shard: test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCUDA::test_forward_mode_AD_cholesky_solve_cuda_complex128 2025-12-04T09:32:56.7020561Z 2025-12-04T09:32:56.7020838Z Finished test_ops_fwd_gradients 2/3 ... [2025-12-04 09:32:56.698749][2412.658294811], took 0.16min 2025-12-04T09:32:56.7328839Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-8f921b290e79acad.xml 2025-12-04T09:32:58.8385114Z Running test_meta 4/4 ... [2025-12-04 09:32:58.838104][2414.797647227] 2025-12-04T09:32:58.8385554Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:32:58.8387646Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_meta.py', '-m', 'not serial', '--shard-id=4', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:32:58.838505] 2025-12-04T09:33:00.3273448Z Running test_nestedtensor 3/4 ... [2025-12-04 09:33:00.326867][2416.286410111] 2025-12-04T09:33:00.3274097Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:33:00.3276380Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_nestedtensor.py', '-m', 'not serial', '--shard-id=3', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:33:00.327342] 2025-12-04T09:33:07.5005507Z 2025-12-04T09:33:07.5006672Z test_nestedtensor 3/4 was successful, full logs can be found in artifacts with path test/test-reports/test_nestedtensor_3.4_b0204203c06b9fb4_.log 2025-12-04T09:33:07.5007551Z Running 0 items in this shard: 2025-12-04T09:33:07.5007767Z 2025-12-04T09:33:07.5008019Z Finished test_nestedtensor 3/4 ... [2025-12-04 09:33:07.500432][2423.459978586], took 0.12min 2025-12-04T09:33:07.5223759Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-2defe60eca797d43.xml 2025-12-04T09:33:10.9843595Z Running test_nestedtensor 4/4 ... [2025-12-04 09:33:10.983901][2426.943442528] 2025-12-04T09:33:10.9844261Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:33:10.9848945Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_nestedtensor.py', '-m', 'not serial', '--shard-id=4', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:33:10.984505] 2025-12-04T09:33:18.1315099Z 2025-12-04T09:33:18.1316057Z test_nestedtensor 4/4 was successful, full logs can be found in artifacts with path test/test-reports/test_nestedtensor_4.4_f06c8dae8e05ce7f_.log 2025-12-04T09:33:18.1316972Z Running 0 items in this shard: 2025-12-04T09:33:18.1317218Z 2025-12-04T09:33:18.1317579Z Finished test_nestedtensor 4/4 ... [2025-12-04 09:33:18.131354][2434.090898307], took 0.12min 2025-12-04T09:33:18.1533709Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-f3843a4cb154c125.xml 2025-12-04T09:33:21.6872395Z Running test_modules 2/2 ... [2025-12-04 09:33:21.686820][2437.646363974] 2025-12-04T09:33:21.6873019Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:33:21.6875179Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_modules.py', '-m', 'not serial', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:33:21.687231] 2025-12-04T09:33:31.5756354Z 2025-12-04T09:33:31.5757391Z test_modules 2/2 was successful, full logs can be found in artifacts with path test/test-reports/test_modules_2.2_5235b4f398c2f08b_.log 2025-12-04T09:33:31.5772652Z Running 50 items in this shard: test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32 2025-12-04T09:33:31.5785672Z 2025-12-04T09:33:31.5785871Z Finished test_modules 2/2 ... [2025-12-04 09:33:31.575632][2447.535178392], took 0.16min 2025-12-04T09:33:31.5981938Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_modules/test_modules-e4b5d71a8ab0cbdf.xml 2025-12-04T09:33:35.0436336Z Running test_tensorboard 1/1 ... [2025-12-04 09:33:35.043194][2451.002738887] 2025-12-04T09:33:35.0436947Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:33:35.0438990Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_tensorboard.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:33:35.043614] 2025-12-04T09:33:38.5975929Z 2025-12-04T09:33:38.5977124Z test_tensorboard 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_tensorboard_1.1_9b09def11169383b_.log 2025-12-04T09:33:38.5977914Z Running 0 items in this shard: 2025-12-04T09:33:38.5978099Z 2025-12-04T09:33:38.5978360Z Finished test_tensorboard 1/1 ... [2025-12-04 09:33:38.597493][2454.55703551], took 0.06min 2025-12-04T09:33:38.6325535Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_tensorboard/test_tensorboard-81b720b0e25291b8.xml 2025-12-04T09:33:42.0396674Z Running torch_np/numpy_tests/core/test_indexing 1/1 ... [2025-12-04 09:33:42.039186][2457.998729313] 2025-12-04T09:33:42.0397456Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:33:42.0399443Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_indexing.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:33:42.039642] 2025-12-04T09:33:45.2588388Z 2025-12-04T09:33:45.2589546Z torch_np/numpy_tests/core/test_indexing 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_indexing_1.1_dec9824650356edd_.log 2025-12-04T09:33:45.2591045Z Running 0 items in this shard: 2025-12-04T09:33:45.2591346Z 2025-12-04T09:33:45.2591914Z Finished torch_np/numpy_tests/core/test_indexing 1/1 ... [2025-12-04 09:33:45.258743][2461.218285922], took 0.05min 2025-12-04T09:33:45.2818518Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.core.test_indexing/torch_np.numpy_tests.core.test_indexing-1a695318c456148e.xml 2025-12-04T09:33:48.7305615Z Running test_futures 1/1 ... [2025-12-04 09:33:48.730050][2464.68959343] 2025-12-04T09:33:48.7306329Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:33:48.7308412Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_futures.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:33:48.730497] 2025-12-04T09:33:51.9097793Z 2025-12-04T09:33:51.9099215Z test_futures 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_futures_1.1_f17fce10eca39347_.log 2025-12-04T09:33:51.9100291Z Running 0 items in this shard: 2025-12-04T09:33:51.9100568Z 2025-12-04T09:33:51.9100944Z Finished test_futures 1/1 ... [2025-12-04 09:33:51.909675][2467.869216852], took 0.05min 2025-12-04T09:33:51.9391152Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_futures/test_futures-eaeb3a3297f2467a.xml 2025-12-04T09:33:55.4545661Z Running nn/test_dropout 1/1 ... [2025-12-04 09:33:55.454108][2471.413652298] 2025-12-04T09:33:55.4546347Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:33:55.4548293Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/test_dropout.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:33:55.454530] 2025-12-04T09:33:59.0823650Z 2025-12-04T09:33:59.0824810Z nn/test_dropout 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.test_dropout_1.1_5b2ab44291c4879d_.log 2025-12-04T09:33:59.0825792Z Running 0 items in this shard: 2025-12-04T09:33:59.0825999Z 2025-12-04T09:33:59.0826209Z Finished nn/test_dropout 1/1 ... [2025-12-04 09:33:59.082276][2475.04181746], took 0.06min 2025-12-04T09:33:59.1041074Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_meta/test_meta-95507d9e0beebf05.xml 2025-12-04T09:33:59.1760971Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.test_dropout/nn.test_dropout-d1d94a3c4456c153.xml 2025-12-04T09:34:01.0512946Z 2025-12-04T09:34:01.0513851Z test_meta 4/4 was successful, full logs can be found in artifacts with path test/test-reports/test_meta_4.4_2bd0d20115e5ff88_.log 2025-12-04T09:34:01.0514533Z Running 0 items in this shard: 2025-12-04T09:34:01.0514709Z 2025-12-04T09:34:01.0514940Z Finished test_meta 4/4 ... [2025-12-04 09:34:01.051168][2477.010715077], took 1.04min 2025-12-04T09:34:02.7755202Z Running functorch/dim/test_split 1/1 ... [2025-12-04 09:34:02.775046][2478.734589633] 2025-12-04T09:34:02.7755718Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:02.7758115Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/dim/test_split.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:02.775491] 2025-12-04T09:34:04.6118743Z Running torch_np/numpy_tests/lib/test_type_check 1/1 ... [2025-12-04 09:34:04.611436][2480.570979386] 2025-12-04T09:34:04.6119205Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:04.6121065Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/lib/test_type_check.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:04.611824] 2025-12-04T09:34:05.9997027Z 2025-12-04T09:34:05.9998160Z functorch/dim/test_split 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.dim.test_split_1.1_14b064ffd62da833_.log 2025-12-04T09:34:05.9999043Z Running 0 items in this shard: 2025-12-04T09:34:05.9999313Z 2025-12-04T09:34:05.9999797Z Finished functorch/dim/test_split 1/1 ... [2025-12-04 09:34:05.999559][2481.959101928], took 0.05min 2025-12-04T09:34:06.0229979Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.dim.test_split/functorch.dim.test_split-ac52710017bc0ec1.xml 2025-12-04T09:34:07.9649885Z 2025-12-04T09:34:07.9651267Z torch_np/numpy_tests/lib/test_type_check 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.lib.test_type_check_1.1_772dbb7849e86426_.log 2025-12-04T09:34:07.9652466Z Running 0 items in this shard: 2025-12-04T09:34:07.9652740Z 2025-12-04T09:34:07.9653347Z Finished torch_np/numpy_tests/lib/test_type_check 1/1 ... [2025-12-04 09:34:07.964858][2483.924405897], took 0.06min 2025-12-04T09:34:07.9885599Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.lib.test_type_check/torch_np.numpy_tests.lib.test_type_check-b366c3bac00b0d6b.xml 2025-12-04T09:34:09.7041173Z Running cpp_extensions/test_libtorch_agnostic 1/1 ... [2025-12-04 09:34:09.703641][2485.663184718] 2025-12-04T09:34:09.7041935Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:09.7043866Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'cpp_extensions/test_libtorch_agnostic.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:09.704054] 2025-12-04T09:34:11.5009383Z Running profiler/test_execution_trace 1/1 ... [2025-12-04 09:34:11.500433][2487.459976994] 2025-12-04T09:34:11.5010144Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:11.5011888Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'profiler/test_execution_trace.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:11.500858] 2025-12-04T09:34:15.0665414Z 2025-12-04T09:34:15.0666640Z profiler/test_execution_trace 1/1 was successful, full logs can be found in artifacts with path test/test-reports/profiler.test_execution_trace_1.1_9cc4d20396616452_.log 2025-12-04T09:34:15.0667749Z Running 0 items in this shard: 2025-12-04T09:34:15.0667986Z 2025-12-04T09:34:15.0668407Z Finished profiler/test_execution_trace 1/1 ... [2025-12-04 09:34:15.066457][2491.026002623], took 0.06min 2025-12-04T09:34:15.1228413Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/profiler.test_execution_trace/profiler.test_execution_trace-2d1f3ef34837b171.xml 2025-12-04T09:34:18.6307692Z Running test_jit 1/1 ... [2025-12-04 09:34:18.630329][2494.589872683] 2025-12-04T09:34:18.6308297Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:18.6310475Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_jit.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:18.630759] 2025-12-04T09:34:27.9751962Z 2025-12-04T09:34:27.9752901Z test_jit 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_jit_1.1_335f426c034a81ab_.log 2025-12-04T09:34:27.9869272Z Running 450 items in this shard: test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_subclasses 2025-12-04T09:34:27.9983546Z 2025-12-04T09:34:27.9983849Z Finished test_jit 1/1 ... [2025-12-04 09:34:27.975696][2503.935238209], took 0.16min 2025-12-04T09:34:27.9998323Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_jit/test_jit-4dc671ed81b7895e.xml 2025-12-04T09:34:29.5606424Z 2025-12-04T09:34:29.5607499Z cpp_extensions/test_libtorch_agnostic 1/1 was successful, full logs can be found in artifacts with path test/test-reports/cpp_extensions.test_libtorch_agnostic_1.1_75653c2be51b730e_.log 2025-12-04T09:34:29.5624823Z Running 50 items in this shard: test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda 2025-12-04T09:34:29.5639830Z 2025-12-04T09:34:29.5640107Z Finished cpp_extensions/test_libtorch_agnostic 1/1 ... [2025-12-04 09:34:29.560775][2505.520321447], took 0.33min 2025-12-04T09:34:29.5851411Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/cpp_extensions.test_libtorch_agnostic/cpp_extensions.test_libtorch_agnostic-8a9c37e2e1da6c28.xml 2025-12-04T09:34:31.6371806Z Running test_datapipe 1/1 ... [2025-12-04 09:34:31.636736][2507.596280556] 2025-12-04T09:34:31.6372252Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:31.6375368Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_datapipe.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:31.637187] 2025-12-04T09:34:33.2049864Z Running test_numba_integration 1/1 ... [2025-12-04 09:34:33.204545][2509.16408818] 2025-12-04T09:34:33.2050360Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:33.2052884Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_numba_integration.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:33.204977] 2025-12-04T09:34:35.0801414Z 2025-12-04T09:34:35.0802336Z test_datapipe 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_datapipe_1.1_8084ffe5984ad57f_.log 2025-12-04T09:34:35.0803166Z Running 0 items in this shard: 2025-12-04T09:34:35.0803356Z 2025-12-04T09:34:35.0803602Z Finished test_datapipe 1/1 ... [2025-12-04 09:34:35.080062][2511.039607825], took 0.06min 2025-12-04T09:34:35.1049871Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_datapipe/test_datapipe-82b13967e74e19a2.xml 2025-12-04T09:34:36.7789696Z 2025-12-04T09:34:36.7790740Z test_numba_integration 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_numba_integration_1.1_22b9b4055cfb8752_.log 2025-12-04T09:34:36.7791572Z Running 0 items in this shard: 2025-12-04T09:34:36.7791866Z 2025-12-04T09:34:36.7792303Z Finished test_numba_integration 1/1 ... [2025-12-04 09:34:36.778848][2512.738391859], took 0.06min 2025-12-04T09:34:36.8041146Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_numba_integration/test_numba_integration-6bf5efdb972a9a8f.xml 2025-12-04T09:34:38.6923963Z Running test_functional_optim 1/1 ... [2025-12-04 09:34:38.691930][2514.651472732] 2025-12-04T09:34:38.6924813Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:38.6926732Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_functional_optim.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:38.692349] 2025-12-04T09:34:40.3300337Z Running test_maskedtensor 1/1 ... [2025-12-04 09:34:40.329583][2516.289127143] 2025-12-04T09:34:40.3300977Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:40.3302692Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_maskedtensor.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:40.329974] 2025-12-04T09:34:42.1308642Z 2025-12-04T09:34:42.1310140Z test_functional_optim 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_functional_optim_1.1_13d5a8983a5027f5_.log 2025-12-04T09:34:42.1311043Z Running 0 items in this shard: 2025-12-04T09:34:42.1311293Z 2025-12-04T09:34:42.1311642Z Finished test_functional_optim 1/1 ... [2025-12-04 09:34:42.130734][2518.090280409], took 0.06min 2025-12-04T09:34:42.1564921Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_functional_optim/test_functional_optim-c14b0e455295dcf4.xml 2025-12-04T09:34:45.6474907Z Running benchmark_utils/test_benchmark_utils 1/1 ... [2025-12-04 09:34:45.647065][2521.606608181] 2025-12-04T09:34:45.6475482Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:45.6478361Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'benchmark_utils/test_benchmark_utils.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:45.647514] 2025-12-04T09:34:46.4819632Z 2025-12-04T09:34:46.4820529Z test_maskedtensor 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_maskedtensor_1.1_86a8621c8ea58e71_.log 2025-12-04T09:34:46.4821239Z Running 0 items in this shard: 2025-12-04T09:34:46.4821409Z 2025-12-04T09:34:46.4821715Z Finished test_maskedtensor 1/1 ... [2025-12-04 09:34:46.481807][2522.441352725], took 0.10min 2025-12-04T09:34:46.5074597Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_maskedtensor/test_maskedtensor-d6bf47e2c29f29e6.xml 2025-12-04T09:34:48.9423850Z 2025-12-04T09:34:48.9424917Z benchmark_utils/test_benchmark_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/benchmark_utils.test_benchmark_utils_1.1_6343f18ab9650ba6_.log 2025-12-04T09:34:48.9425747Z Running 0 items in this shard: 2025-12-04T09:34:48.9425917Z 2025-12-04T09:34:48.9426185Z Finished benchmark_utils/test_benchmark_utils 1/1 ... [2025-12-04 09:34:48.942291][2524.901834261], took 0.05min 2025-12-04T09:34:48.9673686Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/benchmark_utils.test_benchmark_utils/benchmark_utils.test_benchmark_utils-f962c5399d90fb14.xml 2025-12-04T09:34:50.1811010Z Running torch_np/numpy_tests/core/test_scalarmath 1/1 ... [2025-12-04 09:34:50.180606][2526.140149025] 2025-12-04T09:34:50.1811776Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:50.1819950Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_scalarmath.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:50.181601] 2025-12-04T09:34:52.5592762Z Running test_scaled_matmul_cuda 1/1 ... [2025-12-04 09:34:52.558793][2528.518336524] 2025-12-04T09:34:52.5593255Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:52.5595648Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_scaled_matmul_cuda.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:52.559258] 2025-12-04T09:34:53.7571224Z 2025-12-04T09:34:53.7572263Z torch_np/numpy_tests/core/test_scalarmath 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_scalarmath_1.1_de9f81d561ae89ba_.log 2025-12-04T09:34:53.7573555Z Running 0 items in this shard: 2025-12-04T09:34:53.7573755Z 2025-12-04T09:34:53.7574096Z Finished torch_np/numpy_tests/core/test_scalarmath 1/1 ... [2025-12-04 09:34:53.756969][2529.716514696], took 0.06min 2025-12-04T09:34:53.7825048Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.core.test_scalarmath/torch_np.numpy_tests.core.test_scalarmath-6b29237e24c20b32.xml 2025-12-04T09:34:57.3546296Z Running torch_np/numpy_tests/core/test_shape_base 1/1 ... [2025-12-04 09:34:57.354224][2533.313767338] 2025-12-04T09:34:57.3546745Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:57.3549456Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_shape_base.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:57.354676] 2025-12-04T09:34:57.5407990Z 2025-12-04T09:34:57.5408996Z test_scaled_matmul_cuda 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_scaled_matmul_cuda_1.1_d6e96ad4cd7f3119_.log 2025-12-04T09:34:57.5421915Z Running 50 items in this shard: test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda, test/test_scaled_matmul_cuda.py::TestFP8MatmulCUDA::test_honor_sm_carveout_cuda 2025-12-04T09:34:57.5432962Z 2025-12-04T09:34:57.5433175Z Finished test_scaled_matmul_cuda 1/1 ... [2025-12-04 09:34:57.540698][2533.500241247], took 0.08min 2025-12-04T09:34:57.5664994Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_scaled_matmul_cuda/test_scaled_matmul_cuda-6e739739e0574a47.xml 2025-12-04T09:35:00.8811745Z 2025-12-04T09:35:00.8812792Z torch_np/numpy_tests/core/test_shape_base 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_shape_base_1.1_6b4f426779600108_.log 2025-12-04T09:35:00.8813860Z Running 0 items in this shard: 2025-12-04T09:35:00.8814042Z 2025-12-04T09:35:00.8814386Z Finished torch_np/numpy_tests/core/test_shape_base 1/1 ... [2025-12-04 09:35:00.881083][2536.840628579], took 0.06min 2025-12-04T09:35:00.9071650Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.core.test_shape_base/torch_np.numpy_tests.core.test_shape_base-36c86d2bfdd67ddd.xml 2025-12-04T09:35:01.1689844Z Running test_vulkan 1/1 ... [2025-12-04 09:35:01.168561][2537.128104378] 2025-12-04T09:35:01.1690466Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:35:01.1692549Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_vulkan.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:35:01.168969] 2025-12-04T09:35:04.4002530Z 2025-12-04T09:35:04.4003260Z test_vulkan 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_vulkan_1.1_4680f5183d531af0_.log 2025-12-04T09:35:04.4003826Z Running 0 items in this shard: 2025-12-04T09:35:04.4003967Z 2025-12-04T09:35:04.4004139Z Finished test_vulkan 1/1 ... [2025-12-04 09:35:04.400110][2540.359657119], took 0.05min 2025-12-04T09:35:04.4272405Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_vulkan/test_vulkan-e1a521ce9cf70bfb.xml 2025-12-04T09:35:04.5102011Z Running lazy/test_generator 1/1 ... [2025-12-04 09:35:04.509765][2540.469307701] 2025-12-04T09:35:04.5102634Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:35:04.5105376Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'lazy/test_generator.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:35:04.510185] 2025-12-04T09:35:07.8038840Z 2025-12-04T09:35:07.8040160Z lazy/test_generator 1/1 was successful, full logs can be found in artifacts with path test/test-reports/lazy.test_generator_1.1_647ef07d1063ea00_.log 2025-12-04T09:35:07.8041021Z Running 0 items in this shard: 2025-12-04T09:35:07.8041513Z 2025-12-04T09:35:07.8041861Z Finished lazy/test_generator 1/1 ... [2025-12-04 09:35:07.803829][2543.763371081], took 0.05min 2025-12-04T09:35:07.8303531Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/lazy.test_generator/lazy.test_generator-5262beb36f04167a.xml 2025-12-04T09:35:08.0632152Z Running torch_np/numpy_tests/linalg/test_linalg 1/1 ... [2025-12-04 09:35:08.062716][2544.02225966] 2025-12-04T09:35:08.0632926Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:35:08.0635139Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/linalg/test_linalg.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:35:08.063144] 2025-12-04T09:35:11.3339468Z Running torch_np/numpy_tests/core/test_dtype 1/1 ... [2025-12-04 09:35:11.333453][2547.29299605] 2025-12-04T09:35:11.3340241Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:35:11.3342187Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_dtype.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:35:11.333858] 2025-12-04T09:35:11.8768837Z 2025-12-04T09:35:11.8770014Z torch_np/numpy_tests/linalg/test_linalg 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.linalg.test_linalg_1.1_5ada73b70ce3b922_.log 2025-12-04T09:35:11.8771211Z Running 0 items in this shard: 2025-12-04T09:35:11.8771460Z 2025-12-04T09:35:11.8771940Z Finished torch_np/numpy_tests/linalg/test_linalg 1/1 ... [2025-12-04 09:35:11.876782][2547.836327493], took 0.06min 2025-12-04T09:35:11.9301148Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.linalg.test_linalg/torch_np.numpy_tests.linalg.test_linalg-fa5a137aca79e3bc.xml 2025-12-04T09:35:14.7298373Z 2025-12-04T09:35:14.7299293Z torch_np/numpy_tests/core/test_dtype 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_dtype_1.1_c041cb87250c2eda_.log 2025-12-04T09:35:14.7300175Z Running 0 items in this shard: 2025-12-04T09:35:14.7300410Z 2025-12-04T09:35:14.7300786Z Finished torch_np/numpy_tests/core/test_dtype 1/1 ... [2025-12-04 09:35:14.729803][2550.689347242], took 0.06min 2025-12-04T09:35:14.7571277Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.core.test_dtype/torch_np.numpy_tests.core.test_dtype-6960413498a3b4a7.xml 2025-12-04T09:35:15.4031895Z Running lazy/test_debug_util 1/1 ... [2025-12-04 09:35:15.402686][2551.362227431] 2025-12-04T09:35:15.4032512Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:35:15.4035456Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'lazy/test_debug_util.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:35:15.403236] 2025-12-04T09:35:18.3440450Z Running nn/test_load_state_dict 1/1 ... [2025-12-04 09:35:18.343561][2554.303104657] 2025-12-04T09:35:18.3440930Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:35:18.3443013Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/test_load_state_dict.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:35:18.344003] 2025-12-04T09:35:18.6872519Z 2025-12-04T09:35:18.6873711Z lazy/test_debug_util 1/1 was successful, full logs can be found in artifacts with path test/test-reports/lazy.test_debug_util_1.1_5381255485755d96_.log 2025-12-04T09:35:18.6875014Z Running 0 items in this shard: 2025-12-04T09:35:18.6875202Z 2025-12-04T09:35:18.6875461Z Finished lazy/test_debug_util 1/1 ... [2025-12-04 09:35:18.687161][2554.646705157], took 0.05min 2025-12-04T09:35:18.7146217Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/lazy.test_debug_util/lazy.test_debug_util-24c3e47e3d2b7323.xml 2025-12-04T09:35:21.8916769Z 2025-12-04T09:35:21.8918007Z nn/test_load_state_dict 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.test_load_state_dict_1.1_5a4907470fe975cf_.log 2025-12-04T09:35:21.8918770Z Running 0 items in this shard: 2025-12-04T09:35:21.8918965Z 2025-12-04T09:35:21.8919232Z Finished nn/test_load_state_dict 1/1 ... [2025-12-04 09:35:21.891516][2557.851063015], took 0.06min 2025-12-04T09:35:21.9189364Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.test_load_state_dict/nn.test_load_state_dict-b877f1ca494eb282.xml 2025-12-04T09:35:22.2666096Z Running test_shape_ops 1/1 ... [2025-12-04 09:35:22.266124][2558.22566741] 2025-12-04T09:35:22.2666531Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:35:22.2669629Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_shape_ops.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:35:22.266610] 2025-12-04T09:35:25.4962046Z Running nn/test_module_hooks 1/1 ... [2025-12-04 09:35:25.495746][2561.45528932] 2025-12-04T09:35:25.4962517Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:35:25.4965074Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/test_module_hooks.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:35:25.496193] 2025-12-04T09:35:25.9910066Z 2025-12-04T09:35:25.9911063Z test_shape_ops 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_shape_ops_1.1_aa3f7cc31cd86f20_.log 2025-12-04T09:35:25.9917311Z Running 0 items in this shard: 2025-12-04T09:35:25.9917567Z 2025-12-04T09:35:25.9917891Z Finished test_shape_ops 1/1 ... [2025-12-04 09:35:25.990748][2561.950294086], took 0.06min 2025-12-04T09:35:26.0313525Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_shape_ops/test_shape_ops-feb6212001ee076f.xml 2025-12-04T09:35:29.1396586Z 2025-12-04T09:35:29.1397630Z nn/test_module_hooks 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.test_module_hooks_1.1_af6a85c4218349c4_.log 2025-12-04T09:35:29.1398506Z Running 0 items in this shard: 2025-12-04T09:35:29.1398718Z 2025-12-04T09:35:29.1399002Z Finished nn/test_module_hooks 1/1 ... [2025-12-04 09:35:29.139567][2565.099110621], took 0.06min 2025-12-04T09:35:29.1680084Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.test_module_hooks/nn.test_module_hooks-8ff29e8b05880501.xml 2025-12-04T09:35:29.5943223Z Running torch_np/numpy_tests/lib/test_twodim_base 1/1 ... [2025-12-04 09:35:29.593848][2565.553391934] 2025-12-04T09:35:29.5943771Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:35:29.5946479Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/lib/test_twodim_base.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:35:29.594306] 2025-12-04T09:35:32.7500669Z Running profiler/test_memory_profiler 1/1 ... [2025-12-04 09:35:32.749606][2568.709149447] 2025-12-04T09:35:32.7501579Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:35:32.7503757Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'profiler/test_memory_profiler.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:35:32.750050] 2025-12-04T09:35:32.8912721Z 2025-12-04T09:35:32.8914327Z torch_np/numpy_tests/lib/test_twodim_base 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.lib.test_twodim_base_1.1_96deb5da66224442_.log 2025-12-04T09:35:32.8915301Z Running 0 items in this shard: 2025-12-04T09:35:32.8915479Z 2025-12-04T09:35:32.8915827Z Finished torch_np/numpy_tests/lib/test_twodim_base 1/1 ... [2025-12-04 09:35:32.891164][2568.850706217], took 0.05min 2025-12-04T09:35:32.9194981Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.lib.test_twodim_base/torch_np.numpy_tests.lib.test_twodim_base-fd2b4839449b1b20.xml 2025-12-04T09:35:36.4540873Z 2025-12-04T09:35:36.4541915Z profiler/test_memory_profiler 1/1 was successful, full logs can be found in artifacts with path test/test-reports/profiler.test_memory_profiler_1.1_d170501bd0fd49c4_.log 2025-12-04T09:35:36.4560294Z Running 50 items in this shard: test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none 2025-12-04T09:35:36.4576411Z 2025-12-04T09:35:36.4576655Z Finished profiler/test_memory_profiler 1/1 ... [2025-12-04 09:35:36.454176][2572.413721295], took 0.06min 2025-12-04T09:35:36.4826656Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/profiler.test_memory_profiler/profiler.test_memory_profiler-89519a5c938e684c.xml 2025-12-04T09:35:36.4963720Z Running test_jit_llga_fuser 1/1 ... [2025-12-04 09:35:36.496050][2572.455594938] 2025-12-04T09:35:36.4964166Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:35:36.4967471Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_jit_llga_fuser.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:35:36.496466] 2025-12-04T09:35:40.0900362Z Running test_serialization 1/1 ... [2025-12-04 09:35:40.089559][2576.049100762] 2025-12-04T09:35:40.0901030Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:35:40.0903011Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_serialization.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:35:40.089994] 2025-12-04T09:35:40.4012433Z 2025-12-04T09:35:40.4013335Z test_jit_llga_fuser 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_jit_llga_fuser_1.1_490d6d2a811ac6f0_.log 2025-12-04T09:35:40.4014042Z Running 0 items in this shard: 2025-12-04T09:35:40.4014491Z 2025-12-04T09:35:40.4014782Z Finished test_jit_llga_fuser 1/1 ... [2025-12-04 09:35:40.401083][2576.360625945], took 0.07min 2025-12-04T09:35:40.4293497Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_jit_llga_fuser/test_jit_llga_fuser-d79703873568858a.xml 2025-12-04T09:35:43.9693543Z Running test_sparse_csr 1/3 ... [2025-12-04 09:35:43.968847][2579.928388979] 2025-12-04T09:35:43.9694035Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:35:43.9696221Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_sparse_csr.py', '-m', 'not serial', '--shard-id=1', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:35:43.969293] 2025-12-04T09:35:44.0843784Z 2025-12-04T09:35:44.0844871Z test_serialization 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_serialization_1.1_30115278e988556a_.log 2025-12-04T09:35:44.0845758Z Running 0 items in this shard: 2025-12-04T09:35:44.0846042Z 2025-12-04T09:35:44.0846488Z Finished test_serialization 1/1 ... [2025-12-04 09:35:44.084270][2580.04381336], took 0.07min 2025-12-04T09:35:44.1131954Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_serialization/test_serialization-d1a17d69bf98ee36.xml 2025-12-04T09:35:47.7465879Z Running functorch/test_ops 6/9 ... [2025-12-04 09:35:47.746091][2583.705634682] 2025-12-04T09:35:47.7466386Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:35:47.7468413Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_ops.py', '-m', 'not serial', '--shard-id=6', '--num-shards=9', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:35:47.746515] 2025-12-04T09:35:55.9915282Z 2025-12-04T09:35:55.9916176Z test_sparse_csr 1/3 was successful, full logs can be found in artifacts with path test/test-reports/test_sparse_csr_1.3_999a94d874ab59b6_.log 2025-12-04T09:35:55.9928869Z Running 50 items in this shard: test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32, test/test_sparse_csr.py::TestSparseCSRCUDA::test_sparse_add_cuda_float32 2025-12-04T09:35:55.9939072Z 2025-12-04T09:35:55.9939262Z Finished test_sparse_csr 1/3 ... [2025-12-04 09:35:55.991505][2591.951050305], took 0.20min 2025-12-04T09:35:56.0203746Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_sparse_csr/test_sparse_csr-68a0a0e21687dd8c.xml 2025-12-04T09:35:59.5159131Z Running functorch/test_ops 8/9 ... [2025-12-04 09:35:59.515419][2595.474962626] 2025-12-04T09:35:59.5159882Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:35:59.5161380Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_ops.py', '-m', 'not serial', '--shard-id=8', '--num-shards=9', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:35:59.515806] 2025-12-04T09:36:06.7973961Z 2025-12-04T09:36:06.7975293Z functorch/test_ops 6/9 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_ops_6.9_7aa027fe1b42b488_.log 2025-12-04T09:36:06.7976016Z Running 0 items in this shard: 2025-12-04T09:36:06.7976221Z 2025-12-04T09:36:06.7976504Z Finished functorch/test_ops 6/9 ... [2025-12-04 09:36:06.797286][2602.756830885], took 0.32min 2025-12-04T09:36:06.8264039Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-63eca64add6314a4.xml 2025-12-04T09:36:10.4462153Z Running test_mkldnn 1/1 ... [2025-12-04 09:36:10.445734][2606.405278229] 2025-12-04T09:36:10.4462970Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:36:10.4465005Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_mkldnn.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:36:10.446157] 2025-12-04T09:36:13.9722624Z 2025-12-04T09:36:13.9723696Z test_mkldnn 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_mkldnn_1.1_cf74075bbf0a4f72_.log 2025-12-04T09:36:13.9724872Z Running 0 items in this shard: 2025-12-04T09:36:13.9725103Z 2025-12-04T09:36:13.9725331Z Finished test_mkldnn 1/1 ... [2025-12-04 09:36:13.972152][2609.931693061], took 0.06min 2025-12-04T09:36:14.0015080Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_mkldnn/test_mkldnn-3fd399b85d51ccfc.xml 2025-12-04T09:36:18.7568070Z 2025-12-04T09:36:18.7569302Z functorch/test_ops 8/9 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_ops_8.9_bb4596b740c061de_.log 2025-12-04T09:36:18.7570342Z Running 0 items in this shard: 2025-12-04T09:36:18.7570519Z 2025-12-04T09:36:18.7570786Z Finished functorch/test_ops 8/9 ... [2025-12-04 09:36:18.756649][2614.716194528], took 0.32min 2025-12-04T09:36:18.7868274Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-4a923aaed07d8de4.xml 2025-12-04T10:09:28.9169576Z 2025-12-04T10:09:28.9170708Z dynamo/test_fx_graph_runnable 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_fx_graph_runnable_1.1_3eb668d8af36329e_.log 2025-12-04T10:09:28.9260156Z Running 350 items in this shard: test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_all_reduce_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_basic_tensor_add, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_add_dynamic, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_broadcast_collective, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_expression, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_dynamic_shapes_run, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing, test/dynamo/test_fx_graph_runnable.py::FxGraphRunnableTest::test_toy_model_batch_processing 2025-12-04T10:09:28.9344811Z 2025-12-04T10:09:28.9345060Z Finished dynamo/test_fx_graph_runnable 1/1 ... [2025-12-04 10:09:28.917477][4604.877019593], took 48.00min 2025-12-04T10:09:28.9345867Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_fx_graph_runnable/dynamo.test_fx_graph_runnable-71d7146dfdd25cde.xml 2025-12-04T10:09:29.9330559Z Uploading artifacts took 0.81 seconds 2025-12-04T10:09:32.7923881Z Running test batch 'tests to run' cost 3756.4 seconds 2025-12-04T10:09:32.7936816Z Emitting td_test_failure_stats_v2 2025-12-04T10:09:32.7941002Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842972_5448358ad0f911f0a89e0242ac110002 2025-12-04T10:09:32.8845607Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842972_5448358ad0f911f0a89e0242ac110002 2025-12-04T10:09:32.8856678Z Emitting td_test_failure_stats_v2 2025-12-04T10:09:32.8858592Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842972_545637a2d0f911f0a89e0242ac110002 2025-12-04T10:09:32.9184505Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842972_545637a2d0f911f0a89e0242ac110002 2025-12-04T10:09:32.9195802Z Emitting td_test_failure_stats_v2 2025-12-04T10:09:32.9197106Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842972_545b63c6d0f911f0a89e0242ac110002 2025-12-04T10:09:32.9589493Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842972_545b63c6d0f911f0a89e0242ac110002 2025-12-04T10:09:32.9601404Z Emitting td_test_failure_stats_v2 2025-12-04T10:09:32.9602874Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842972_54619534d0f911f0a89e0242ac110002 2025-12-04T10:09:32.9932008Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842972_54619534d0f911f0a89e0242ac110002 2025-12-04T10:09:32.9942784Z Emitting td_test_failure_stats_v2 2025-12-04T10:09:32.9944258Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842972_5466cb08d0f911f0a89e0242ac110002 2025-12-04T10:09:33.0279559Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842972_5466cb08d0f911f0a89e0242ac110002 2025-12-04T10:09:33.0291507Z Emitting td_test_failure_stats_v2 2025-12-04T10:09:33.0292994Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842973_546c1d56d0f911f0a89e0242ac110002 2025-12-04T10:09:33.0622081Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842973_546c1d56d0f911f0a89e0242ac110002 2025-12-04T10:09:33.0632964Z Emitting td_test_failure_stats_v2 2025-12-04T10:09:33.0634451Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842973_54715320d0f911f0a89e0242ac110002 2025-12-04T10:09:33.1002803Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842973_54715320d0f911f0a89e0242ac110002 2025-12-04T10:09:33.1003947Z test_cuda 1/1 failed! 2025-12-04T10:09:33.1004233Z inductor/test_extension_backend 1/1 failed! 2025-12-04T10:09:33.1004743Z test_cuda 1/1 failed! 2025-12-04T10:09:33.1004973Z test_ops 8/9 failed! 2025-12-04T10:09:33.1005218Z functorch/test_dims 1/1 failed! 2025-12-04T10:09:33.1005496Z inductor/test_kernel_benchmark 1/1 failed! 2025-12-04T10:09:33.1005776Z test_decomp 18/22 failed! 2025-12-04T10:09:33.7017528Z 2025-12-04T10:09:33.7017981Z real 62m41.748s 2025-12-04T10:09:33.7018252Z user 92m44.936s 2025-12-04T10:09:33.7018459Z sys 28m29.833s 2025-12-04T10:09:33.7018665Z + assert_git_not_dirty 2025-12-04T10:09:33.7018963Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 != *rocm* ]] 2025-12-04T10:09:33.7019352Z + [[ linux-jammy-cuda12.8-py3.10-gcc11 != *xla* ]] 2025-12-04T10:09:33.7024015Z ++ git status --porcelain 2025-12-04T10:09:33.7024522Z ++ grep -v '?? third_party' 2025-12-04T10:09:36.9240649Z ++ true 2025-12-04T10:09:36.9241795Z + git_status= 2025-12-04T10:09:36.9242066Z + [[ -n '' ]] 2025-12-04T10:09:36.9243850Z + sccache_epilogue 2025-12-04T10:09:36.9244154Z + echo '::group::Sccache Compilation Log' 2025-12-04T10:09:36.9245076Z ##[group]Sccache Compilation Log 2025-12-04T10:09:36.9245416Z + echo '=================== sccache compilation log ===================' 2025-12-04T10:09:36.9245927Z =================== sccache compilation log =================== 2025-12-04T10:09:36.9246506Z + python /var/lib/jenkins/workspace/.ci/pytorch/print_sccache_log.py /var/lib/jenkins/sccache_error.log 2025-12-04T10:09:36.9368851Z + echo '=========== If your build fails, please take a look at the log above for possible reasons ===========' 2025-12-04T10:09:36.9369616Z =========== If your build fails, please take a look at the log above for possible reasons =========== 2025-12-04T10:09:36.9370080Z + sccache --show-stats 2025-12-04T10:09:36.9397559Z Compile requests 3103 2025-12-04T10:09:36.9397900Z Compile requests executed 105 2025-12-04T10:09:36.9398208Z Cache hits 58 2025-12-04T10:09:36.9398482Z Cache hits (C/C++) 58 2025-12-04T10:09:36.9398753Z Cache misses 47 2025-12-04T10:09:36.9399009Z Cache misses (C/C++) 47 2025-12-04T10:09:36.9399279Z Cache hits rate 55.24 % 2025-12-04T10:09:36.9399567Z Cache hits rate (C/C++) 55.24 % 2025-12-04T10:09:36.9399832Z Cache timeouts 0 2025-12-04T10:09:36.9400222Z Cache read errors 0 2025-12-04T10:09:36.9400490Z Forced recaches 0 2025-12-04T10:09:36.9400739Z Cache write errors 0 2025-12-04T10:09:36.9400993Z Cache errors 0 2025-12-04T10:09:36.9401255Z Compilations 47 2025-12-04T10:09:36.9401475Z Compilation failures 0 2025-12-04T10:09:36.9401695Z Non-cacheable compilations 0 2025-12-04T10:09:36.9401918Z Non-cacheable calls 352 2025-12-04T10:09:36.9402134Z Non-compilation calls 2646 2025-12-04T10:09:36.9402345Z Unsupported compiler calls 0 2025-12-04T10:09:36.9402582Z Average cache write 0.047 s 2025-12-04T10:09:36.9402820Z Average compiler 11.827 s 2025-12-04T10:09:36.9403036Z Average cache read hit 0.049 s 2025-12-04T10:09:36.9403265Z Failed distributed compilations 0 2025-12-04T10:09:36.9403418Z 2025-12-04T10:09:36.9403495Z Non-cacheable reasons: 2025-12-04T10:09:36.9403673Z -E 352 2025-12-04T10:09:36.9403832Z 2025-12-04T10:09:36.9404001Z Cache location s3, name: ossci-compiler-cache-circleci-v2, prefix: / 2025-12-04T10:09:36.9404336Z Version (client) 0.10.0 2025-12-04T10:09:36.9404831Z + sccache --stop-server 2025-12-04T10:09:36.9422479Z Stopping sccache server... 2025-12-04T10:09:36.9424816Z Compile requests 3103 2025-12-04T10:09:36.9425144Z Compile requests executed 105 2025-12-04T10:09:36.9425411Z Cache hits 58 2025-12-04T10:09:36.9425670Z Cache hits (C/C++) 58 2025-12-04T10:09:36.9425927Z Cache misses 47 2025-12-04T10:09:36.9426181Z Cache misses (C/C++) 47 2025-12-04T10:09:36.9426446Z Cache hits rate 55.24 % 2025-12-04T10:09:36.9426725Z Cache hits rate (C/C++) 55.24 % 2025-12-04T10:09:36.9426990Z Cache timeouts 0 2025-12-04T10:09:36.9427241Z Cache read errors 0 2025-12-04T10:09:36.9427506Z Forced recaches 0 2025-12-04T10:09:36.9427771Z Cache write errors 0 2025-12-04T10:09:36.9428145Z Cache errors 0 2025-12-04T10:09:36.9428627Z Compilations 47 2025-12-04T10:09:36.9429019Z Compilation failures 0 2025-12-04T10:09:36.9429455Z Non-cacheable compilations 0 2025-12-04T10:09:36.9429722Z Non-cacheable calls 352 2025-12-04T10:09:36.9429991Z Non-compilation calls 2646 2025-12-04T10:09:36.9430263Z Unsupported compiler calls 0 2025-12-04T10:09:36.9430626Z Average cache write 0.047 s 2025-12-04T10:09:36.9430985Z Average compiler 11.827 s 2025-12-04T10:09:36.9431278Z Average cache read hit 0.049 s 2025-12-04T10:09:36.9431533Z Failed distributed compilations 0 2025-12-04T10:09:36.9431690Z 2025-12-04T10:09:36.9431760Z Non-cacheable reasons: 2025-12-04T10:09:36.9431944Z -E 352 2025-12-04T10:09:36.9432087Z 2025-12-04T10:09:36.9432263Z Cache location s3, name: ossci-compiler-cache-circleci-v2, prefix: / 2025-12-04T10:09:36.9432639Z Version (client) 0.10.0 2025-12-04T10:09:36.9432878Z + echo ::endgroup:: 2025-12-04T10:09:36.9433209Z ##[endgroup] 2025-12-04T10:09:36.9433374Z + cleanup_workspace 2025-12-04T10:09:36.9433717Z + echo 'sudo may print the following warning message that can be ignored. The chown command will still run.' 2025-12-04T10:09:36.9434265Z sudo may print the following warning message that can be ignored. The chown command will still run. 2025-12-04T10:09:36.9434724Z + echo ' sudo: setrlimit(RLIMIT_STACK): Operation not permitted' 2025-12-04T10:09:36.9435053Z sudo: setrlimit(RLIMIT_STACK): Operation not permitted 2025-12-04T10:09:36.9435438Z + echo 'For more details refer to https://github.com/sudo-project/sudo/issues/42' 2025-12-04T10:09:36.9435877Z For more details refer to https://github.com/sudo-project/sudo/issues/42 2025-12-04T10:09:36.9436218Z + sudo chown -R 1000 /var/lib/jenkins/workspace 2025-12-04T10:09:37.8919898Z ##[group]Run pytorch/test-infra/.github/actions/upload-benchmark-results@main 2025-12-04T10:09:37.8920269Z with: 2025-12-04T10:09:37.8920462Z benchmark-results-dir: test/test-reports 2025-12-04T10:09:37.8920693Z dry-run: false 2025-12-04T10:09:37.8920867Z schema-version: v3 2025-12-04T10:09:37.8921235Z github-token: *** 2025-12-04T10:09:37.8921409Z env: 2025-12-04T10:09:37.8921556Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:09:37.8921750Z HAS_NVIDIA_GPU: true 2025-12-04T10:09:37.8921985Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:09:37.8922379Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:09:37.8922722Z ##[endgroup] 2025-12-04T10:09:37.8937748Z ##[group]Run set -eux 2025-12-04T10:09:37.8937964Z set -eux 2025-12-04T10:09:37.8938122Z  2025-12-04T10:09:37.8938271Z if [[ -n "" ]]; then 2025-12-04T10:09:37.8938464Z  source "" 2025-12-04T10:09:37.8938662Z fi 2025-12-04T10:09:37.8938925Z python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-12-04T10:09:37.8939237Z  2025-12-04T10:09:37.8939393Z DEVICE_NAME="" 2025-12-04T10:09:37.8939582Z DEVICE_TYPE="" 2025-12-04T10:09:37.8939761Z  2025-12-04T10:09:37.8939939Z if command -v nvidia-smi; then 2025-12-04T10:09:37.8940264Z  # NB: I'm using PyTorch here to get the device name, however, it needs to 2025-12-04T10:09:37.8940694Z  # install the correct version of PyTorch manually for now. Any PyTorch 2025-12-04T10:09:37.8941077Z  # version is fine, I just use 2.7.1 to satify PYPIDEP linter 2025-12-04T10:09:37.8941381Z  python3 -mpip install torch==2.7.1 2025-12-04T10:09:37.8941621Z elif command -v rocminfo; then 2025-12-04T10:09:37.8941926Z  # NB: Installing torch on ROCm runner with pip here causes CI to fail 2025-12-04T10:09:37.8942330Z  # with a memoryview is too large error only on MI300 runners. Is pip 2025-12-04T10:09:37.8942721Z  # version on ROCm runner there too old? As a workaround, let's use the 2025-12-04T10:09:37.8943061Z  # GPU device name coming from rocminfo instead 2025-12-04T10:09:37.8943419Z  DEVICE_NAME=rocm 2025-12-04T10:09:37.8943757Z  DEVICE_TYPE=$(rocminfo | grep "Marketing Name" | tail -n1 | awk -F':' '{print $2}' | xargs) 2025-12-04T10:09:37.8944171Z fi 2025-12-04T10:09:37.8944316Z  2025-12-04T10:09:37.8944501Z echo "DEVICE_NAME=$DEVICE_NAME" >> $GITHUB_ENV 2025-12-04T10:09:37.8944784Z echo "DEVICE_TYPE=$DEVICE_TYPE" >> $GITHUB_ENV 2025-12-04T10:09:37.8956145Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:09:37.8956426Z env: 2025-12-04T10:09:37.8956586Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:09:37.8956774Z HAS_NVIDIA_GPU: true 2025-12-04T10:09:37.8957014Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:09:37.8957420Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:09:37.8957757Z ##[endgroup] 2025-12-04T10:09:37.8993150Z + [[ -n '' ]] 2025-12-04T10:09:37.8993511Z + python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-12-04T10:09:38.1136592Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T10:09:39.1746282Z Collecting boto3==1.35.33 2025-12-04T10:09:39.1923004Z Downloading boto3-1.35.33-py3-none-any.whl (139 kB) 2025-12-04T10:09:39.4856855Z Collecting psutil==7.0.0 2025-12-04T10:09:39.4899452Z Downloading psutil-7.0.0-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (277 kB) 2025-12-04T10:09:39.5228894Z Collecting pynvml==12.0.0 2025-12-04T10:09:39.5264791Z Downloading pynvml-12.0.0-py3-none-any.whl (26 kB) 2025-12-04T10:09:39.5338765Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.33) (0.10.0) 2025-12-04T10:09:39.5710085Z Collecting s3transfer<0.11.0,>=0.10.0 2025-12-04T10:09:39.5747888Z Downloading s3transfer-0.10.4-py3-none-any.whl (83 kB) 2025-12-04T10:09:40.7213915Z Collecting botocore<1.36.0,>=1.35.33 2025-12-04T10:09:40.7253450Z Downloading botocore-1.35.99-py3-none-any.whl (13.3 MB) 2025-12-04T10:09:40.9011311Z Collecting nvidia-ml-py<13.0.0a0,>=12.0.0 2025-12-04T10:09:40.9047281Z Downloading nvidia_ml_py-12.575.51-py3-none-any.whl (47 kB) 2025-12-04T10:09:40.9137569Z Requirement already satisfied: python-dateutil<3.0.0,>=2.1 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (2.8.1) 2025-12-04T10:09:40.9147119Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.25.10) 2025-12-04T10:09:41.1320674Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil<3.0.0,>=2.1->botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.15.0) 2025-12-04T10:09:41.2458755Z Installing collected packages: botocore, s3transfer, nvidia-ml-py, pynvml, psutil, boto3 2025-12-04T10:09:41.7638253Z Attempting uninstall: nvidia-ml-py 2025-12-04T10:09:41.7640011Z Found existing installation: nvidia-ml-py 11.525.84 2025-12-04T10:09:41.7653632Z Uninstalling nvidia-ml-py-11.525.84: 2025-12-04T10:09:41.7845618Z Successfully uninstalled nvidia-ml-py-11.525.84 2025-12-04T10:09:41.8386262Z Attempting uninstall: psutil 2025-12-04T10:09:41.8388173Z Found existing installation: psutil 5.9.8 2025-12-04T10:09:41.8468264Z Uninstalling psutil-5.9.8: 2025-12-04T10:09:41.8475862Z Successfully uninstalled psutil-5.9.8 2025-12-04T10:09:41.9993001Z Successfully installed boto3-1.35.33 botocore-1.35.99 nvidia-ml-py-12.575.51 psutil-7.0.0 pynvml-12.0.0 s3transfer-0.10.4 2025-12-04T10:09:42.0881048Z + DEVICE_NAME= 2025-12-04T10:09:42.0881324Z + DEVICE_TYPE= 2025-12-04T10:09:42.0881551Z + command -v nvidia-smi 2025-12-04T10:09:42.0881857Z + python3 -mpip install torch==2.7.1 2025-12-04T10:09:42.0882159Z /usr/bin/nvidia-smi 2025-12-04T10:09:42.3007135Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T10:09:42.5479642Z Collecting torch==2.7.1 2025-12-04T10:09:42.5663236Z Downloading torch-2.7.1-cp39-cp39-manylinux_2_28_x86_64.whl (821.1 MB) 2025-12-04T10:09:54.6663371Z Collecting networkx 2025-12-04T10:09:54.6698631Z Downloading networkx-3.2.1-py3-none-any.whl (1.6 MB) 2025-12-04T10:09:54.7192477Z Collecting nvidia-curand-cu12==10.3.7.77 2025-12-04T10:09:54.7228976Z Downloading nvidia_curand_cu12-10.3.7.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (56.3 MB) 2025-12-04T10:09:55.2698397Z Requirement already satisfied: jinja2 in /usr/lib/python3.9/site-packages (from torch==2.7.1) (2.11.3) 2025-12-04T10:09:55.2969598Z Collecting nvidia-cuda-cupti-cu12==12.6.80 2025-12-04T10:09:55.3029459Z Downloading nvidia_cuda_cupti_cu12-12.6.80-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (8.9 MB) 2025-12-04T10:09:55.4113026Z Collecting nvidia-cublas-cu12==12.6.4.1 2025-12-04T10:09:55.4174906Z Downloading nvidia_cublas_cu12-12.6.4.1-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (393.1 MB) 2025-12-04T10:10:00.7980650Z Collecting nvidia-nccl-cu12==2.26.2 2025-12-04T10:10:00.8012873Z Downloading nvidia_nccl_cu12-2.26.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (201.3 MB) 2025-12-04T10:10:03.0722412Z Collecting triton==3.3.1 2025-12-04T10:10:03.0780701Z Downloading triton-3.3.1-cp39-cp39-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl (155.6 MB) 2025-12-04T10:10:04.5510330Z Collecting nvidia-cuda-nvrtc-cu12==12.6.77 2025-12-04T10:10:04.5583373Z Downloading nvidia_cuda_nvrtc_cu12-12.6.77-py3-none-manylinux2014_x86_64.whl (23.7 MB) 2025-12-04T10:10:04.8383676Z Collecting nvidia-nvtx-cu12==12.6.77 2025-12-04T10:10:04.8415204Z Downloading nvidia_nvtx_cu12-12.6.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (89 kB) 2025-12-04T10:10:04.8978733Z Collecting nvidia-cudnn-cu12==9.5.1.17 2025-12-04T10:10:04.9051804Z Downloading nvidia_cudnn_cu12-9.5.1.17-py3-none-manylinux_2_28_x86_64.whl (571.0 MB) 2025-12-04T10:10:12.9349912Z Requirement already satisfied: typing-extensions>=4.10.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from torch==2.7.1) (4.15.0) 2025-12-04T10:10:12.9668886Z Collecting nvidia-cusolver-cu12==11.7.1.2 2025-12-04T10:10:12.9734186Z Downloading nvidia_cusolver_cu12-11.7.1.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (158.2 MB) 2025-12-04T10:10:14.6016944Z Collecting nvidia-cusparse-cu12==12.5.4.2 2025-12-04T10:10:14.6116424Z Downloading nvidia_cusparse_cu12-12.5.4.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (216.6 MB) 2025-12-04T10:10:17.0891880Z Collecting filelock 2025-12-04T10:10:17.0929673Z Downloading filelock-3.19.1-py3-none-any.whl (15 kB) 2025-12-04T10:10:17.1109513Z Collecting nvidia-cusparselt-cu12==0.6.3 2025-12-04T10:10:17.1145270Z Downloading nvidia_cusparselt_cu12-0.6.3-py3-none-manylinux2014_x86_64.whl (156.8 MB) 2025-12-04T10:10:18.6023424Z Collecting nvidia-cufile-cu12==1.11.1.6 2025-12-04T10:10:18.6053084Z Downloading nvidia_cufile_cu12-1.11.1.6-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (1.1 MB) 2025-12-04T10:10:18.6554252Z Collecting sympy>=1.13.3 2025-12-04T10:10:18.6586949Z Downloading sympy-1.14.0-py3-none-any.whl (6.3 MB) 2025-12-04T10:10:18.7437050Z Collecting nvidia-cufft-cu12==11.3.0.4 2025-12-04T10:10:18.7469448Z Downloading nvidia_cufft_cu12-11.3.0.4-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (200.2 MB) 2025-12-04T10:10:20.9449196Z Collecting nvidia-cuda-runtime-cu12==12.6.77 2025-12-04T10:10:20.9512291Z Downloading nvidia_cuda_runtime_cu12-12.6.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (897 kB) 2025-12-04T10:10:21.0322450Z Collecting fsspec 2025-12-04T10:10:21.0355104Z Downloading fsspec-2025.10.0-py3-none-any.whl (200 kB) 2025-12-04T10:10:21.0658072Z Collecting nvidia-nvjitlink-cu12==12.6.85 2025-12-04T10:10:21.0719308Z Downloading nvidia_nvjitlink_cu12-12.6.85-py3-none-manylinux2010_x86_64.manylinux_2_12_x86_64.whl (19.7 MB) 2025-12-04T10:10:21.2561338Z Requirement already satisfied: setuptools>=40.8.0 in /usr/lib/python3.9/site-packages (from triton==3.3.1->torch==2.7.1) (59.6.0) 2025-12-04T10:10:21.2821213Z Collecting mpmath<1.4,>=1.1.0 2025-12-04T10:10:21.2859876Z Downloading mpmath-1.3.0-py3-none-any.whl (536 kB) 2025-12-04T10:10:21.3650773Z Requirement already satisfied: MarkupSafe>=0.23 in /usr/lib64/python3.9/site-packages (from jinja2->torch==2.7.1) (1.1.1) 2025-12-04T10:10:21.6640623Z Installing collected packages: nvidia-nvjitlink-cu12, nvidia-cusparse-cu12, nvidia-cublas-cu12, mpmath, triton, sympy, nvidia-nvtx-cu12, nvidia-nccl-cu12, nvidia-cusparselt-cu12, nvidia-cusolver-cu12, nvidia-curand-cu12, nvidia-cufile-cu12, nvidia-cufft-cu12, nvidia-cudnn-cu12, nvidia-cuda-runtime-cu12, nvidia-cuda-nvrtc-cu12, nvidia-cuda-cupti-cu12, networkx, fsspec, filelock, torch 2025-12-04T10:10:29.3326216Z WARNING: The scripts proton and proton-viewer are installed in '/home/ec2-user/.local/bin' which is not on PATH. 2025-12-04T10:10:29.3327083Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-12-04T10:10:32.7777404Z WARNING: The script isympy is installed in '/home/ec2-user/.local/bin' which is not on PATH. 2025-12-04T10:10:32.7778203Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-12-04T10:10:59.0328588Z WARNING: The scripts torchfrtrace and torchrun are installed in '/home/ec2-user/.local/bin' which is not on PATH. 2025-12-04T10:10:59.0329450Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-12-04T10:10:59.2122882Z Successfully installed filelock-3.19.1 fsspec-2025.10.0 mpmath-1.3.0 networkx-3.2.1 nvidia-cublas-cu12-12.6.4.1 nvidia-cuda-cupti-cu12-12.6.80 nvidia-cuda-nvrtc-cu12-12.6.77 nvidia-cuda-runtime-cu12-12.6.77 nvidia-cudnn-cu12-9.5.1.17 nvidia-cufft-cu12-11.3.0.4 nvidia-cufile-cu12-1.11.1.6 nvidia-curand-cu12-10.3.7.77 nvidia-cusolver-cu12-11.7.1.2 nvidia-cusparse-cu12-12.5.4.2 nvidia-cusparselt-cu12-0.6.3 nvidia-nccl-cu12-2.26.2 nvidia-nvjitlink-cu12-12.6.85 nvidia-nvtx-cu12-12.6.77 sympy-1.14.0 torch-2.7.1 triton-3.3.1 2025-12-04T10:10:59.7202195Z + echo DEVICE_NAME= 2025-12-04T10:10:59.7214087Z + echo DEVICE_TYPE= 2025-12-04T10:10:59.7244895Z ##[group]Run set -eux 2025-12-04T10:10:59.7245101Z set -eux 2025-12-04T10:10:59.7245263Z  2025-12-04T10:10:59.7245441Z if [[ -z "${GITHUB_TOKEN}" ]]; then 2025-12-04T10:10:59.7245690Z  echo "Missing github-token input" 2025-12-04T10:10:59.7245913Z  exit 1 2025-12-04T10:10:59.7246078Z fi 2025-12-04T10:10:59.7253641Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:10:59.7253924Z env: 2025-12-04T10:10:59.7254085Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:10:59.7254280Z HAS_NVIDIA_GPU: true 2025-12-04T10:10:59.7254508Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:10:59.7254899Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:10:59.7255274Z DEVICE_NAME: 2025-12-04T10:10:59.7255436Z DEVICE_TYPE: 2025-12-04T10:10:59.7255782Z GITHUB_TOKEN: *** 2025-12-04T10:10:59.7255968Z ##[endgroup] 2025-12-04T10:10:59.7286332Z + [[ -z *** ]] 2025-12-04T10:10:59.7364681Z ##[group]Run pytorch/test-infra/.github/actions/get-workflow-job-id@main 2025-12-04T10:10:59.7365012Z with: 2025-12-04T10:10:59.7365306Z github-token: *** 2025-12-04T10:10:59.7365488Z env: 2025-12-04T10:10:59.7365673Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:10:59.7365881Z HAS_NVIDIA_GPU: true 2025-12-04T10:10:59.7366118Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:10:59.7366529Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:10:59.7366903Z DEVICE_NAME: 2025-12-04T10:10:59.7367075Z DEVICE_TYPE: 2025-12-04T10:10:59.7367247Z ##[endgroup] 2025-12-04T10:10:59.7456023Z ##[group]Run set -eux 2025-12-04T10:10:59.7456228Z set -eux 2025-12-04T10:10:59.7456397Z  2025-12-04T10:10:59.7456879Z python3 "${GITHUB_ACTION_PATH}/../../scripts/get_workflow_job_id.py" "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-12-04T10:10:59.7463877Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:10:59.7464270Z env: 2025-12-04T10:10:59.7464439Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:10:59.7464645Z HAS_NVIDIA_GPU: true 2025-12-04T10:10:59.7464885Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:10:59.7465282Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:10:59.7465633Z DEVICE_NAME: 2025-12-04T10:10:59.7465799Z DEVICE_TYPE: 2025-12-04T10:10:59.7466093Z GITHUB_TOKEN: *** 2025-12-04T10:10:59.7466275Z ##[endgroup] 2025-12-04T10:10:59.7492022Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/get-workflow-job-id/../../scripts/get_workflow_job_id.py 19922768520 i-077765b42bd7d5290 2025-12-04T10:11:02.0137825Z setting job-id=57116084869 2025-12-04T10:11:02.0138541Z setting job-name=linux-jammy-cuda12.8-py3.10-gcc11 / test (default, 4, 5, lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests) 2025-12-04T10:11:02.0235772Z ##[group]Run set -eux 2025-12-04T10:11:02.0235996Z set -eux 2025-12-04T10:11:02.0236147Z  2025-12-04T10:11:02.0236299Z if [[ -n "" ]]; then 2025-12-04T10:11:02.0236491Z  source "" 2025-12-04T10:11:02.0236670Z fi 2025-12-04T10:11:02.0236815Z  2025-12-04T10:11:02.0237090Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_metadata.py" \ 2025-12-04T10:11:02.0237467Z  --schema-version "${SCHEMA_VERSION}" \ 2025-12-04T10:11:02.0237710Z  --repo "${REPO}" \ 2025-12-04T10:11:02.0238115Z  --head-branch "${HEAD_BRANCH}" \ 2025-12-04T10:11:02.0238383Z  --head-sha "${HEAD_SHA}" \ 2025-12-04T10:11:02.0238620Z  --workflow-id "${WORKFLOW_RUN_ID}" \ 2025-12-04T10:11:02.0238875Z  --run-attempt "${RUN_ATTEMPT}" \ 2025-12-04T10:11:02.0239111Z  --job-id "${JOB_ID}" \ 2025-12-04T10:11:02.0239325Z  --job-name "${JOB_NAME}" 2025-12-04T10:11:02.0246696Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:11:02.0246978Z env: 2025-12-04T10:11:02.0247147Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:02.0247332Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:02.0247557Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:02.0247947Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:02.0248294Z DEVICE_NAME: 2025-12-04T10:11:02.0248458Z DEVICE_TYPE: 2025-12-04T10:11:02.0248617Z SCHEMA_VERSION: v3 2025-12-04T10:11:02.0248804Z REPO: pytorch/pytorch 2025-12-04T10:11:02.0248984Z HEAD_BRANCH: refs/heads/main 2025-12-04T10:11:02.0249214Z HEAD_SHA: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T10:11:02.0249528Z WORKFLOW_RUN_ID: 19922768520 2025-12-04T10:11:02.0249710Z RUN_ATTEMPT: 1 2025-12-04T10:11:02.0249901Z JOB_ID: 57116084869 2025-12-04T10:11:02.0250332Z JOB_NAME: linux-jammy-cuda12.8-py3.10-gcc11 / test (default, 4, 5, lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests) 2025-12-04T10:11:02.0250820Z ##[endgroup] 2025-12-04T10:11:02.0276515Z + [[ -n '' ]] 2025-12-04T10:11:02.0278252Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_metadata.py --schema-version v3 --repo pytorch/pytorch --head-branch refs/heads/main --head-sha ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 --workflow-id 19922768520 --run-attempt 1 --job-id 57116084869 --job-name 'linux-jammy-cuda12.8-py3.10-gcc11 / test (default, 4, 5, lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests)' 2025-12-04T10:11:02.0695737Z ##[group]Run set -eux 2025-12-04T10:11:02.0695929Z set -eux 2025-12-04T10:11:02.0696087Z  2025-12-04T10:11:02.0696235Z if [[ -n "" ]]; then 2025-12-04T10:11:02.0696557Z  source "" 2025-12-04T10:11:02.0696720Z fi 2025-12-04T10:11:02.0696856Z  2025-12-04T10:11:02.0697144Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_runners_info.py" 2025-12-04T10:11:02.0704068Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:11:02.0704350Z env: 2025-12-04T10:11:02.0704736Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:02.0704970Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:02.0705200Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:02.0705585Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:02.0705928Z DEVICE_NAME: 2025-12-04T10:11:02.0706098Z DEVICE_TYPE: 2025-12-04T10:11:02.0706247Z ##[endgroup] 2025-12-04T10:11:02.0730753Z + [[ -n '' ]] 2025-12-04T10:11:02.0731362Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_runners_info.py 2025-12-04T10:11:02.8871566Z /home/ec2-user/.local/lib/python3.9/site-packages/torch/_subclasses/functional_tensor.py:276: UserWarning: Failed to initialize NumPy: No module named 'numpy' (Triggered internally at /pytorch/torch/csrc/utils/tensor_numpy.cpp:81.) 2025-12-04T10:11:02.8872757Z cpu = _conversion_method_template(device=torch.device("cpu")) 2025-12-04T10:11:03.6945270Z ##[group]Run set -eux 2025-12-04T10:11:03.6945478Z set -eux 2025-12-04T10:11:03.6945634Z  2025-12-04T10:11:03.6945815Z # TODO (huydhn): Implement this part 2025-12-04T10:11:03.6946098Z echo "dependencies={}" >> "${GITHUB_OUTPUT}" 2025-12-04T10:11:03.6953892Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:11:03.6954182Z env: 2025-12-04T10:11:03.6954369Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:03.6954703Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:03.6954988Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:03.6955385Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:03.6955723Z DEVICE_NAME: 2025-12-04T10:11:03.6955884Z DEVICE_TYPE: 2025-12-04T10:11:03.6956049Z ##[endgroup] 2025-12-04T10:11:03.6983074Z + echo 'dependencies={}' 2025-12-04T10:11:03.7044211Z ##[group]Run set -eux 2025-12-04T10:11:03.7044425Z set -eux 2025-12-04T10:11:03.7044601Z  2025-12-04T10:11:03.7044765Z if [[ -n "" ]]; then 2025-12-04T10:11:03.7044970Z  source "" 2025-12-04T10:11:03.7045143Z fi 2025-12-04T10:11:03.7045294Z  2025-12-04T10:11:03.7045488Z if [[ ! -d "${BENCHMARK_RESULTS_DIR}" ]]; then 2025-12-04T10:11:03.7045836Z  echo "${BENCHMARK_RESULTS_DIR} does not exist, skipping" 2025-12-04T10:11:03.7046202Z  # We don't want the job to fail if the directory doesn't exist 2025-12-04T10:11:03.7046486Z  exit 0 2025-12-04T10:11:03.7046648Z fi 2025-12-04T10:11:03.7046806Z  2025-12-04T10:11:03.7046978Z if [[ "${DRY_RUN}" == "true" ]]; then 2025-12-04T10:11:03.7047329Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-12-04T10:11:03.7047741Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-12-04T10:11:03.7048046Z  --metadata "${BENCHMARK_METADATA}" \ 2025-12-04T10:11:03.7048310Z  --runners "${RUNNER_INFO}" \ 2025-12-04T10:11:03.7048569Z  --dependencies "${DEPENDENCIES}" \ 2025-12-04T10:11:03.7048815Z  --dry-run 2025-12-04T10:11:03.7048994Z else 2025-12-04T10:11:03.7049282Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-12-04T10:11:03.7049690Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-12-04T10:11:03.7049991Z  --metadata "${BENCHMARK_METADATA}" \ 2025-12-04T10:11:03.7050245Z  --runners "${RUNNER_INFO}" \ 2025-12-04T10:11:03.7050609Z  --dependencies "${DEPENDENCIES}" 2025-12-04T10:11:03.7050841Z fi 2025-12-04T10:11:03.7057600Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:11:03.7057977Z env: 2025-12-04T10:11:03.7058148Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:03.7058345Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:03.7058582Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:03.7058988Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:03.7059348Z DEVICE_NAME: 2025-12-04T10:11:03.7059649Z DEVICE_TYPE: 2025-12-04T10:11:03.7059851Z BENCHMARK_RESULTS_DIR: test/test-reports 2025-12-04T10:11:03.7060093Z DRY_RUN: false 2025-12-04T10:11:03.7061182Z BENCHMARK_METADATA: {"timestamp": 1764843062, "schema_version": "v3", "name": "linux-jammy-cuda12.8-py3.10-gcc11 / test (default, 4, 5, lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32", "workflow_id": 19922768520, "run_attempt": 1, "job_id": 57116084869} 2025-12-04T10:11:03.7062657Z RUNNER_INFO: [{"cpu_info": "x86_64", "cpu_count": 16, "avail_mem_in_gb": 60, "extra_info": {"hostname": "ip-10-1-22-180.ec2.internal"}, "name": "cuda", "type": "NVIDIA L4", "gpu_count": 1, "avail_gpu_mem_in_gb": 22}] 2025-12-04T10:11:03.7063239Z DEPENDENCIES: {} 2025-12-04T10:11:03.7063419Z ##[endgroup] 2025-12-04T10:11:03.7086907Z + [[ -n '' ]] 2025-12-04T10:11:03.7087163Z + [[ ! -d test/test-reports ]] 2025-12-04T10:11:03.7087440Z + [[ false == \t\r\u\e ]] 2025-12-04T10:11:03.7090428Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/upload_benchmark_results.py --benchmark-results-dir test/test-reports --metadata '{"timestamp": 1764843062, "schema_version": "v3", "name": "linux-jammy-cuda12.8-py3.10-gcc11 / test (default, 4, 5, lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32", "workflow_id": 19922768520, "run_attempt": 1, "job_id": 57116084869}' --runners '[{"cpu_info": "x86_64", "cpu_count": 16, "avail_mem_in_gb": 60, "extra_info": {"hostname": "ip-10-1-22-180.ec2.internal"}, "name": "cuda", "type": "NVIDIA L4", "gpu_count": 1, "avail_gpu_mem_in_gb": 22}]' --dependencies '{}' 2025-12-04T10:11:03.8467951Z /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/upload_benchmark_results.py:236: UserWarning: {'included': [{'test_file': 'inductor/test_aot_inductor'}, {'test_file': 'dynamo/test_dynamic_shapes'}, {'test_file': 'inductor/test_cudagraph_trees'}, {'test_file': 'dynamo/test_repros'}, {'test_file': 'dynamo/test_misc'}, {'test_file': 'inductor/test_flex_attention'}, {'test_file': 'inductor/test_cuda_select_algorithm'}, {'test_file': 'inductor/test_compile_subprocess'}, {'test_file': 'export/test_retraceability'}, {'test_file': 'dynamo/test_model_output'}, {'test_file': 'inductor/test_extension_backend'}, {'test_file': 'inductor/test_cooperative_reductions'}, {'test_file': 'dynamo/test_fx_graph_runnable'}, {'test_file': 'dynamo/test_backends'}, {'test_file': 'dynamo/test_functions'}, {'test_file': 'inductor/test_mix_order_reduction'}, {'test_file': 'inductor/test_padding'}, {'test_file': 'dynamo/test_dicts'}, {'test_file': 'dynamo/test_aot_compile'}, {'test_file': 'dynamo/test_sets'}, {'test_file': 'dynamo/test_callback'}, {'test_file': 'dynamo/test_fake_distributed'}, {'test_file': 'inductor/test_cudagraph_trees_expandable_segments'}, {'test_file': 'dynamo/test_exceptions'}, {'test_file': 'inductor/test_collective_autotuning'}, {'test_file': 'test_cuda'}, {'test_file': 'test_transformers'}, {'test_file': 'higher_order_ops/test_local_map'}, {'test_file': 'test_dataloader'}, {'test_file': 'test_decomp'}, {'test_file': 'test_ci_sanity_check_fail'}, {'test_file': 'test_cuda_expandable_segments'}, {'test_file': 'test_public_bindings'}, {'test_file': 'test_ops'}, {'test_file': 'functorch/test_dims'}, {'test_file': 'test_torchfuzz_repros'}, {'test_file': 'inductor/test_max_autotune'}, {'test_file': 'doctests'}, {'test_file': 'inductor/test_torchinductor'}, {'test_file': 'inductor/test_select_algorithm'}, {'test_file': 'inductor/test_torchinductor_dynamic_shapes'}, {'test_file': 'inductor/test_torchinductor_codegen_dynamic_shapes'}, {'test_file': 'inductor/test_kernel_benchmark'}, {'test_file': 'inductor/test_torchinductor_opinfo'}, {'test_file': 'inductor/test_pattern_matcher'}, {'test_file': 'inductor/test_group_batch_fusion'}, {'test_file': 'inductor/test_cpu_repro'}, {'test_file': 'inductor/test_cuda_repro'}, {'test_file': 'inductor/test_smoke'}, {'test_file': 'dynamo/test_after_aot'}, {'test_file': 'inductor/test_snode_runtime'}, {'test_file': 'inductor/test_minifier'}, {'test_file': 'inductor/test_compiled_autograd'}, {'test_file': 'inductor/test_custom_lowering'}, {'test_file': 'inductor/test_perf'}, {'test_file': 'inductor/test_fused_attention'}, {'test_file': 'dynamo/test_activation_checkpointing'}, {'test_file': 'inductor/test_binary_folding'}, {'test_file': 'inductor/test_mkldnn_pattern_matcher'}, {'test_file': 'dynamo/test_logging'}, {'test_file': 'inductor/test_inductor_freezing'}, {'test_file': 'inductor/test_layout_optim'}, {'test_file': 'dynamo/test_unspec'}, {'test_file': 'dynamo/test_higher_order_ops'}, {'test_file': 'inductor/test_mmdecomp'}, {'test_file': 'dynamo/test_ctx_manager'}, {'test_file': 'dynamo/test_exc'}, {'test_file': 'inductor/test_cpu_select_algorithm'}, {'test_file': 'inductor/test_aot_inductor_arrayref'}, {'test_file': 'inductor/test_cpu_cpp_wrapper'}, {'test_file': 'inductor/test_cutlass_backend'}, {'test_file': 'inductor/test_triton_cpu_backend'}, {'test_file': 'inductor/test_torchinductor_strided_blocks'}, {'test_file': 'test_custom_ops'}, {'test_file': 'test_content_store'}, {'test_file': 'inductor/test_halide'}, {'test_file': 'inductor/test_flex_decoding'}, {'test_file': 'inductor/test_deterministic'}, {'test_file': 'inductor/test_multi_kernel'}, {'test_file': 'inductor/test_analysis'}, {'test_file': 'inductor/test_pad_mm'}, {'test_file': 'inductor/test_triton_syntax'}, {'test_file': 'inductor/test_triton_extension_backend'}, {'test_file': 'test_sparse_semi_structured'}, {'test_file': 'inductor/test_op_completeness'}, {'test_file': 'inductor/test_subgraph_choice'}, {'test_file': 'inductor/test_b2b_gemm'}, {'test_file': 'inductor/test_triton_heuristics'}, {'test_file': 'inductor/test_cutedsl_grouped_mm'}, {'test_file': 'inductor/test_cpp_wrapper_hipify'}, {'test_file': 'inductor/test_ck_backend'}, {'test_file': 'inductor/test_inductor_utils'}, {'test_file': 'inductor/test_template_heuristics_registry'}, {'test_file': 'inductor/test_async_compile'}, {'test_file': 'inductor/test_gpu_cpp_wrapper'}, {'test_file': 'export/test_export_training_ir_to_run_decomp'}, {'test_file': 'dynamo/test_deque_reconstruct'}, {'test_file': 'inductor/test_utils'}, {'test_file': 'inductor/test_indexing'}, {'test_file': 'inductor/test_inductor_annotations'}, {'test_file': 'inductor/test_compile_worker'}, {'test_file': 'dynamo/test_einops'}, {'test_file': 'inductor/test_external_callables'}, {'test_file': 'test_testing'}, {'test_file': 'dynamo/test_fx_passes_pre_grad'}, {'test_file': 'inductor/test_fp8'}, {'test_file': 'inductor/test_autoheuristic'}, {'test_file': 'export/test_strict_export_v2'}, {'test_file': 'inductor/test_flex_flash'}, {'test_file': 'inductor/test_segmented_tree'}, {'test_file': 'inductor/test_kernel_optimization'}, {'test_file': 'inductor/test_metrics'}, {'test_file': 'export/test_unflatten_training_ir'}, {'test_file': 'inductor/test_triton_kernels'}, {'test_file': 'inductor/test_fx_fusion'}, {'test_file': 'inductor/test_native_matmul'}, {'test_file': 'inductor/test_xpu_basic'}, {'test_file': 'dynamo/test_inline_and_install'}, {'test_file': 'export/test_functionalized_assertions'}, {'test_file': 'inductor/test_loop_ordering'}, {'test_file': 'inductor/test_selective_lowering'}, {'test_file': 'dynamo/test_base_output'}, {'test_file': 'inductor/test_lookup_table'}, {'test_file': 'export/test_serialize'}, {'test_file': 'inductor/test_cutedsl_template'}, {'test_file': 'inductor/test_benchmark_fusion'}, {'test_file': 'inductor/test_inductor_scheduler'}, {'test_file': 'inductor/test_move_constructors_to_gpu'}, {'test_file': 'export/test_export_strict'}, {'test_file': 'dynamo/test_modules'}, {'test_file': 'export/test_serdes'}, {'test_file': 'dynamo/test_regional_inductor'}, {'test_file': 'inductor/test_remote_cache'}, {'test_file': 'inductor/test_coordinate_descent_tuner'}, {'test_file': 'inductor/test_inplace_padding'}, {'test_file': 'inductor/test_cudacodecache'}, {'test_file': 'inductor/test_minifier_utils'}, {'test_file': 'inductor/test_debug_trace'}, {'test_file': 'dynamo/test_recompiles'}, {'test_file': 'inductor/test_foreach'}, {'test_file': 'export/test_tree_utils'}, {'test_file': 'inductor/test_triton_wrapper'}, {'test_file': 'inductor/test_static_cuda_launcher'}, {'test_file': 'export/test_dynamic_shapes'}, {'test_file': 'dynamo/test_sdpa'}, {'test_file': 'inductor/test_aot_inductor_package'}, {'test_file': 'dynamo/test_utils'}, {'test_file': 'inductor/test_provenance_tracing'}, {'test_file': 'inductor/test_combo_kernels'}, {'test_file': 'inductor/test_codegen_triton'}, {'test_file': 'dynamo/test_frame_init'}, {'test_file': 'inductor/test_device_assert'}, {'test_file': 'dynamo/test_skip_non_tensor'}, {'test_file': 'dynamo/test_skip_guard_eval_unsafe'}, {'test_file': 'inductor/test_decompose_mem_bound_mm'}, {'test_file': 'dynamo/test_interop'}, {'test_file': 'functorch/test_eager_transforms'}, {'test_file': 'inductor/test_control_deps'}, {'test_file': 'inductor/test_benchmarking'}, {'test_file': 'inductor/test_helion_kernels'}, {'test_file': 'inductor/test_quantization'}, {'test_file': 'inductor/test_op_dtype_prop'}, {'test_file': 'inductor/test_best_config'}, {'test_file': 'export/test_tools'}, {'test_file': 'inductor/test_compiled_optimizers'}, {'test_file': 'dynamo/test_buffers_override'}, {'test_file': 'inductor/test_inplacing_pass'}, {'test_file': 'inductor/test_aot_inductor_custom_ops'}, {'test_file': 'inductor/test_split_cat_fx_passes'}, {'test_file': 'inductor/test_profiler'}, {'test_file': 'inductor/test_memory_planning'}, {'test_file': 'inductor/test_mem_estimation'}, {'test_file': 'dynamo/test_view'}, {'test_file': 'inductor/test_cutlass_evt'}, {'test_file': 'dynamo/test_reconstruct'}, {'test_file': 'dynamo/test_aot_autograd'}, {'test_file': 'export/test_cpp_serdes'}, {'test_file': 'inductor/test_cache'}, {'test_file': 'inductor/test_block_analysis'}, {'test_file': 'inductor/test_online_softmax'}, {'test_file': 'dynamo/test_subgraphs'}, {'test_file': 'dynamo/test_pre_dispatch'}, {'test_file': 'inductor/test_custom_post_grad_passes'}, {'test_file': 'dynamo/test_fx_annotate'}, {'test_file': 'dynamo/test_pgo'}, {'test_file': 'dynamo/test_config'}, {'test_file': 'dynamo/test_metrics_context'}, {'test_file': 'export/test_package'}, {'test_file': 'export/test_export_opinfo'}, {'test_file': 'dynamo/test_nops'}, {'test_file': 'inductor/test_graph_transform_observer'}, {'test_file': 'inductor/test_memory'}, {'test_file': 'inductor/test_aot_inductor_utils'}, {'test_file': 'export/test_db'}, {'test_file': 'dynamo/test_export_mutations'}, {'test_file': 'inductor/test_config'}, {'test_file': 'inductor/test_dependencies'}, {'test_file': 'inductor/test_fuzzer'}, {'test_file': 'dynamo/test_global'}, {'test_file': 'inductor/test_control_flow'}, {'test_file': 'dynamo/test_graph_region_tracker'}, {'test_file': 'dynamo/test_unittest'}, {'test_file': 'dynamo/test_streams'}, {'test_file': 'inductor/test_compile'}, {'test_file': 'inductor/test_unbacked_symints'}, {'test_file': 'inductor/test_scatter_optimization'}, {'test_file': 'inductor/test_ordered_set'}, {'test_file': 'inductor/test_pallas'}, {'test_file': 'dynamo/test_install_free_tensors'}, {'test_file': 'inductor/test_torchinductor_codegen_config_overrides'}, {'test_file': 'export/test_passes'}, {'test_file': 'dynamo/test_autograd_function'}, {'test_file': 'inductor/test_codecache'}, {'test_file': 'dynamo/test_cudagraphs'}, {'test_file': 'inductor/test_alignment'}, {'test_file': 'dynamo/test_profiler'}, {'test_file': 'dynamo/test_guard_serialization'}, {'test_file': 'dynamo/test_compile'}, {'test_file': 'dynamo/test_nested_graph_breaks'}, {'test_file': 'inductor/test_needs_exact_strides'}, {'test_file': 'inductor/test_auto_functionalize'}, {'test_file': 'inductor/test_split_cat_fx_aten_passes'}, {'test_file': 'inductor/test_minifier_isolate'}, {'test_file': 'dynamo/test_list'}, {'test_file': 'dynamo/test_resume'}, {'test_file': 'inductor/test_augmented_graph_helper'}, {'test_file': 'dynamo/test_deviceguard'}, {'test_file': 'dynamo/test_sources'}, {'test_file': 'dynamo/test_backward_higher_order_ops'}, {'test_file': 'dynamo/test_modes'}, {'test_file': 'dynamo/test_optimizers'}, {'test_file': 'export/test_torchbind'}, {'test_file': 'inductor/test_custom_partitioner_fn'}, {'test_file': 'dynamo/test_debug_utils'}, {'test_file': 'dynamo/test_base_hop'}, {'test_file': 'dynamo/test_export'}, {'test_file': 'dynamo/test_package'}, {'test_file': 'inductor/test_efficient_conv_bn_eval'}, {'test_file': 'inductor/test_torchbind'}, {'test_file': 'dynamo/test_python_dispatcher'}, {'test_file': 'export/test_swap'}, {'test_file': 'export/test_unflatten'}, {'test_file': 'dynamo/test_verify_correctness'}, {'test_file': 'dynamo/test_wrap_inductor_compiled_regions'}, {'test_file': 'inductor/test_fxir_backend'}, {'test_file': 'dynamo/test_cudagraphs_expandable_segments'}, {'test_file': 'inductor/test_caching'}, {'test_file': 'dynamo/test_aot_autograd_cache'}, {'test_file': 'dynamo/test_flat_apply'}, {'test_file': 'dynamo/test_input_attr_tracking'}, {'test_file': 'dynamo/test_compiler_bisector'}, {'test_file': 'dynamo/test_graph_deduplication'}, {'test_file': 'inductor/test_distributed_patterns'}, {'test_file': 'dynamo/test_structured_trace'}, {'test_file': 'dynamo/test_error_messages'}, {'test_file': 'dynamo/test_bytecode_utils'}, {'test_file': 'inductor/test_mps_basic'}, {'test_file': 'export/test_nativert'}, {'test_file': 'inductor/test_custom_op_autotune'}, {'test_file': 'export/test_hop'}, {'test_file': 'dynamo/test_tree_map'}, {'test_file': 'dynamo/test_minifier'}, {'test_file': 'dynamo/test_guard_manager'}, {'test_file': 'export/test_schema'}, {'test_file': 'dynamo/test_torchrec'}, {'test_file': 'export/test_pass_infra'}, {'test_file': 'dynamo/test_recompile_ux'}, {'test_file': 'export/test_experimental'}, {'test_file': 'export/test_converter'}, {'test_file': 'export/test_export'}, {'test_file': 'test_model_exports_to_core_aten'}, {'test_file': 'dynamo/test_precompile_context'}, {'test_file': 'dynamo/test_trace_rules'}, {'test_file': 'export/test_upgrader'}, {'test_file': 'dynamo/test_hooks'}, {'test_file': 'dynamo/test_reorder_logs'}, {'test_file': 'dynamo/test_subclasses'}, {'test_file': 'dynamo/test_generator'}, {'test_file': 'export/test_lift_unlift'}, {'test_file': 'dynamo/test_decorators'}, {'test_file': 'export/test_verifier'}, {'test_file': 'export/test_sparse'}, {'test_file': 'dynamo/test_python_autograd'}, {'test_file': 'export/test_draft_export'}, {'test_file': 'dynamo/test_comptime'}, {'test_file': 'test_sort_and_select'}, {'test_file': 'test_privateuseone_python_backend'}, {'test_file': 'functorch/test_rearrange'}, {'test_file': 'functorch/test_parsing'}, {'test_file': 'test_varlen_attention'}, {'test_file': 'test_package'}, {'test_file': 'profiler/test_profiler'}, {'test_file': 'test_mkl_verbose'}, {'test_file': 'test_comparison_utils'}, {'test_file': 'functorch/test_ac_logging'}, {'test_file': 'test_mkldnn_verbose'}, {'test_file': 'test_cpp_api_parity'}, {'test_file': 'test_utils_config_module'}, {'test_file': 'test_hop_infra'}, {'test_file': 'test_appending_byte_serializer'}, {'test_file': 'test_license'}, {'test_file': 'test_ao_sparsity'}, {'test_file': 'test_autoload'}, {'test_file': 'nn/attention/test_open_registry'}, {'test_file': 'xpu/test_fusion'}, {'test_file': 'test_as_strided'}, {'test_file': 'test_foreach'}, {'test_file': 'test_proxy_tensor'}, {'test_file': 'torch_np/test_binary_ufuncs'}, {'test_file': 'torch_np/test_unary_ufuncs'}, {'test_file': 'test_utils_filelock'}, {'test_file': 'test_extension_utils'}, {'test_file': 'test_rename_privateuse1_to_existing_device'}, {'test_file': 'nn/attention/test_fa4'}, {'test_file': 'typing/test_python_operators'}, {'test_file': 'test_functionalization'}, {'test_file': 'test_matmul_cuda'}, {'test_file': 'torch_np/test_dtype'}, {'test_file': 'test_file_check'}, {'test_file': 'profiler/test_kineto'}, {'test_file': 'test_flop_counter'}, {'test_file': 'backends/xeon/test_launch'}, {'test_file': 'test_show_pickle'}, {'test_file': 'test_openmp'}, {'test_file': 'test_expanded_weights'}, {'test_file': 'test_module_tracker'}, {'test_file': 'test_autograd'}, {'test_file': 'torch_np/numpy_tests/core/test_scalarinherit'}, {'test_file': 'test_tensorexpr_pybind'}, {'test_file': 'test_fx_experimental'}, {'test_file': 'functorch/test_ac_knapsack'}, {'test_file': 'torch_np/test_nep50_examples'}, {'test_file': 'test_torch'}, {'test_file': 'xpu/test_gemm'}, {'test_file': 'test_sparse'}, {'test_file': 'test_fx_passes'}, {'test_file': 'functorch/test_logging'}, {'test_file': 'test_namedtensor'}, {'test_file': 'test_tensorexpr'}, {'test_file': 'functorch/test_minifier'}, {'test_file': 'higher_order_ops/test_invoke_quant'}, {'test_file': 'torch_np/test_basic'}, {'test_file': 'test_jiterator'}, {'test_file': 'test_native_functions'}, {'test_file': 'test_typing'}, {'test_file': 'higher_order_ops/test_with_effects'}, {'test_file': 'higher_order_ops/test_invoke_subgraph'}, {'test_file': 'test_weak'}, {'test_file': 'test_complex'}, {'test_file': 'test_optim'}, {'test_file': 'lazy/test_functionalization'}, {'test_file': 'torch_np/test_random'}, {'test_file': 'nn/test_multihead_attention'}, {'test_file': 'test_legacy_vmap'}, {'test_file': 'lazy/test_bindings'}, {'test_file': 'xpu/test_conv'}, {'test_file': 'test_utils'}, {'test_file': 'test_pytree'}, {'test_file': 'test_namedtuple_return_api'}, {'test_file': 'profiler/test_record_function'}, {'test_file': 'test_compile_benchmark_util'}, {'test_file': 'test_set_default_mobile_cpu_allocator'}, {'test_file': 'test_fake_tensor'}, {'test_file': 'test_stateless'}, {'test_file': 'functorch/test_ac'}, {'test_file': 'test_binary_ufuncs'}, {'test_file': 'higher_order_ops/test_print'}, {'test_file': 'test_per_overload_api'}, {'test_file': 'torch_np/numpy_tests/core/test_einsum'}, {'test_file': 'test_multiprocessing'}, {'test_file': 'test_out_dtype_op'}, {'test_file': 'torch_np/test_ufuncs_basic'}, {'test_file': 'lazy/test_step_closures'}, {'test_file': 'functorch/dim/test_getsetitem'}, {'test_file': 'test_ops_fwd_gradients'}, {'test_file': 'test_meta'}, {'test_file': 'test_ops_jit'}, {'test_file': 'test_fx'}, {'test_file': 'test_numpy_interop'}, {'test_file': 'profiler/test_cpp_thread'}, {'test_file': 'test_ops_gradients'}, {'test_file': 'test_hub'}, {'test_file': 'test_segment_reductions'}, {'test_file': 'test_opaque_obj_v2'}, {'test_file': 'test_autograd_fallback'}, {'test_file': 'test_type_hints'}, {'test_file': 'functorch/test_aot_joint_with_descriptors'}, {'test_file': 'test_nestedtensor'}, {'test_file': 'test_linalg'}, {'test_file': 'test_functionalization_of_rng_ops'}, {'test_file': 'test_fx_reinplace_pass'}, {'test_file': 'functorch/test_control_flow'}, {'test_file': 'test_modules'}, {'test_file': 'nn/test_packed_sequence'}, {'test_file': 'test_numa_binding'}, {'test_file': 'test_pruning_op'}, {'test_file': 'test_jit_fuser_te'}, {'test_file': 'test_autocast'}, {'test_file': 'test_logging'}, {'test_file': 'test_python_dispatch'}, {'test_file': 'nn/test_lazy_modules'}, {'test_file': 'nn/test_pruning'}, {'test_file': 'test_monitor'}, {'test_file': 'test_cuda_sanitizer'}, {'test_file': 'test_bundled_inputs'}, {'test_file': 'torch_np/numpy_tests/core/test_numeric'}, {'test_file': 'torch_np/numpy_tests/core/test_multiarray'}, {'test_file': 'test_itt'}, {'test_file': 'torch_np/numpy_tests/lib/test_function_base'}, {'test_file': 'test_masked'}, {'test_file': 'test_sympy_utils'}, {'test_file': 'test_jit_disabled'}, {'test_file': 'test_subclass'}, {'test_file': 'test_import_stats'}, {'test_file': 'functorch/test_vmap_registrations'}, {'test_file': 'nn/test_parametrization'}, {'test_file': 'test_mkldnn_fusion'}, {'test_file': 'test_cpp_extensions_mtia_backend'}, {'test_file': 'lazy/test_ts_opinfo'}, {'test_file': 'test_dynamic_shapes'}, {'test_file': 'complex_tensor/test_complex_tensor'}, {'test_file': 'optim/test_lrscheduler'}, {'test_file': 'optim/test_swa_utils'}, {'test_file': 'cpp_extensions/python_agnostic_extension/test/test_python_agnostic'}, {'test_file': 'functorch/test_memory_efficient_fusion'}, {'test_file': 'torch_np/numpy_tests/lib/test_histograms'}, {'test_file': 'torch_np/test_indexing'}, {'test_file': 'test_schema_check'}, {'test_file': 'test_tensorboard'}, {'test_file': 'torch_np/numpy_tests/core/test_indexing'}, {'test_file': 'test_futures'}, {'test_file': 'test_tensor_creation_ops'}, {'test_file': 'nn/test_dropout'}, {'test_file': 'functorch/dim/test_split'}, {'test_file': 'torch_np/numpy_tests/lib/test_type_check'}, {'test_file': 'cpp_extensions/test_libtorch_agnostic'}, {'test_file': 'test_cpp_extensions_stream_and_event'}, {'test_file': 'profiler/test_execution_trace'}, {'test_file': 'test_jit'}, {'test_file': 'test_dispatch'}, {'test_file': 'test_datapipe'}, {'test_file': 'test_numba_integration'}, {'test_file': 'test_functional_optim'}, {'test_file': 'test_maskedtensor'}, {'test_file': 'benchmark_utils/test_benchmark_utils'}, {'test_file': 'torch_np/numpy_tests/core/test_scalarmath'}, {'test_file': 'test_scaled_matmul_cuda'}, {'test_file': 'torch_np/numpy_tests/core/test_shape_base'}, {'test_file': 'test_vulkan'}, {'test_file': 'lazy/test_generator'}, {'test_file': 'nn/test_convolution'}, {'test_file': 'torch_np/numpy_tests/linalg/test_linalg'}, {'test_file': 'torch_np/numpy_tests/core/test_dtype'}, {'test_file': 'lazy/test_debug_util'}, {'test_file': 'nn/test_load_state_dict'}, {'test_file': 'test_shape_ops'}, {'test_file': 'nn/test_module_hooks'}, {'test_file': 'torch_np/numpy_tests/lib/test_twodim_base'}, {'test_file': 'profiler/test_memory_profiler'}, {'test_file': 'test_jit_llga_fuser'}, {'test_file': 'test_serialization'}, {'test_file': 'test_sparse_csr'}, {'test_file': 'optim/test_optim'}, {'test_file': 'test_indexing'}, {'test_file': 'torch_np/numpy_tests/fft/test_pocketfft'}, {'test_file': 'functorch/test_ops'}, {'test_file': 'torch_np/numpy_tests/lib/test_shape_base_'}, {'test_file': 'test_cpp_extensions_jit'}, {'test_file': 'torch_np/numpy_tests/core/test_getlimits'}, {'test_file': 'torch_np/test_ndarray_methods'}, {'test_file': 'test_view_ops'}, {'test_file': 'test_type_info'}, {'test_file': 'functorch/test_aotdispatch'}, {'test_file': 'test_nn'}, {'test_file': 'torch_np/numpy_tests/core/test_dlpack'}, {'test_file': 'test_multiprocessing_spawn'}, {'test_file': 'test_scatter_gather_ops'}, {'test_file': 'test_cuda_multigpu'}, {'test_file': 'test_mkldnn'}, {'test_file': 'functorch/test_vmap'}, {'test_file': 'torch_np/numpy_tests/lib/test_index_tricks'}, {'test_file': 'test_jit_autocast'}, {'test_file': 'nn/test_pooling'}, {'test_file': 'nn/test_embedding'}, {'test_file': 'test_unary_ufuncs'}, {'test_file': 'test_xnnpack_integration'}, {'test_file': 'test_cuda_trace'}, {'test_file': 'test_native_mha'}, {'test_file': 'torch_np/numpy_tests/core/test_numerictypes'}, {'test_file': 'test_cuda_nvml_based_avail'}, {'test_file': 'test_function_schema'}, {'test_file': 'test_accelerator'}, {'test_file': 'nn/test_init'}, {'test_file': 'torch_np/numpy_tests/core/test_scalar_methods'}, {'test_file': 'torch_np/numpy_tests/fft/test_helper'}, {'test_file': 'test_mobile_optimizer'}, {'test_file': 'test_overrides'}, {'test_file': 'torch_np/test_function_base'}, {'test_file': 'test_type_promotion'}, {'test_file': 'torch_np/test_scalars_0D_arrays'}, {'test_file': 'test_cuda_primary_ctx'}, {'test_file': 'profiler/test_profiler_tree'}, {'test_file': 'torch_np/numpy_tests/lib/test_arraysetops'}, {'test_file': 'test_dlpack'}, {'test_file': 'profiler/test_torch_tidy'}, {'test_file': 'lazy/test_reuse_ir'}, {'test_file': 'test_functional_autograd_benchmark'}, {'test_file': 'test_reductions'}, {'test_file': 'torch_np/test_reductions'}, {'test_file': 'torch_np/numpy_tests/core/test_scalar_ctors'}, {'test_file': 'torch_np/numpy_tests/lib/test_arraypad'}, {'test_file': 'test_prims'}, {'test_file': 'test_spectral_ops'}, {'test_file': 'profiler/test_python_tracer'}, {'test_file': 'cpp_extensions/libtorch_agnostic_2_10_extension/test_version_compatibility'}, {'test_file': 'distributions/test_distributions'}, {'test_file': 'test_autoload_disable'}, {'test_file': 'test_autoload_enable'}, {'test_file': 'test_cpp_extensions_aot_ninja'}, {'test_file': 'test_cpp_extensions_aot_no_ninja'}], 'excluded': []} from test/test-reports/td_exclusions-cf4f9bb2408eb653a222.json is not a benchmark record, skipping 2025-12-04T10:11:03.8519966Z warn(f"{result} from {filepath} is not a benchmark record, skipping") 2025-12-04T10:11:03.8607195Z ##[group]Run cat test/**/*_toprint.log || true 2025-12-04T10:11:03.8607549Z cat test/**/*_toprint.log || true 2025-12-04T10:11:03.8615022Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:11:03.8615315Z env: 2025-12-04T10:11:03.8615476Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:03.8615677Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:03.8615894Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:03.8616283Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:03.8616625Z DEVICE_NAME: 2025-12-04T10:11:03.8616785Z DEVICE_TYPE: 2025-12-04T10:11:03.8616942Z ##[endgroup] 2025-12-04T10:11:03.8711332Z cat: 'test/**/*_toprint.log': No such file or directory 2025-12-04T10:11:03.9210808Z ##[group]Run kill "$MONITOR_SCRIPT_PID" 2025-12-04T10:11:03.9211191Z kill "$MONITOR_SCRIPT_PID" 2025-12-04T10:11:03.9218151Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:11:03.9218425Z env: 2025-12-04T10:11:03.9218586Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:03.9218796Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:03.9219021Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:03.9219423Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:03.9219780Z DEVICE_NAME: 2025-12-04T10:11:03.9219944Z DEVICE_TYPE: 2025-12-04T10:11:03.9220104Z MONITOR_SCRIPT_PID: 58913 2025-12-04T10:11:03.9220296Z ##[endgroup] 2025-12-04T10:11:03.9244873Z /home/ec2-user/actions-runner/_work/_temp/0dcd14ac-6080-4ff4-b64f-2e3ab6805ce6.sh: line 1: kill: (58913) - No such process 2025-12-04T10:11:03.9255100Z ##[error]Process completed with exit code 1. 2025-12-04T10:11:03.9366718Z Prepare all required actions 2025-12-04T10:11:03.9367074Z Getting action download info 2025-12-04T10:11:04.1409347Z Download action repository 'seemethere/upload-artifact-s3@v5' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-12-04T10:11:04.8479906Z Download action repository 'actions/upload-artifact@v4' (SHA:ea165f8d65b6e75b540449e92b4886f43607fa02) 2025-12-04T10:11:06.7993260Z ##[group]Run ./.github/actions/upload-test-artifacts 2025-12-04T10:11:06.7993532Z with: 2025-12-04T10:11:06.7993847Z file-suffix: test-default-4-5-lf.linux.g6.4xlarge.experimental.nvidia.gpu_57116084869 2025-12-04T10:11:06.7994246Z s3-bucket: gha-artifacts 2025-12-04T10:11:06.7994440Z env: 2025-12-04T10:11:06.7994596Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:06.7994793Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:06.7995030Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:06.7995424Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:06.7995823Z DEVICE_NAME: 2025-12-04T10:11:06.7995992Z DEVICE_TYPE: 2025-12-04T10:11:06.7996156Z ##[endgroup] 2025-12-04T10:11:06.8094188Z ##[group]Run # Remove any previous test jsons if they exist 2025-12-04T10:11:06.8094523Z # Remove any previous test jsons if they exist 2025-12-04T10:11:06.8094794Z rm -f test-jsons-*.zip 2025-12-04T10:11:06.8095222Z zip -r "test-jsons-${FILE_SUFFIX}.zip" test/test-reports -i '*.json' 2025-12-04T10:11:06.8103101Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:11:06.8103541Z env: 2025-12-04T10:11:06.8103725Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:06.8103948Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:06.8104184Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:06.8104866Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:06.8105229Z DEVICE_NAME: 2025-12-04T10:11:06.8105383Z DEVICE_TYPE: 2025-12-04T10:11:06.8105708Z FILE_SUFFIX: test-default-4-5-lf.linux.g6.4xlarge.experimental.nvidia.gpu_57116084869 2025-12-04T10:11:06.8106068Z ##[endgroup] 2025-12-04T10:11:06.9424257Z adding: test/test-reports/td_exclusions-cf4f9bb2408eb653a222.json (deflated 82%) 2025-12-04T10:11:06.9425070Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-78423a5d6c0345b8.json (stored 0%) 2025-12-04T10:11:06.9426031Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-863ed9428b922f00.json (stored 0%) 2025-12-04T10:11:06.9426973Z adding: test/test-reports/python-pytest/dynamo.test_model_output/dynamo.test_model_output-5ac7859c2c204934.json (stored 0%) 2025-12-04T10:11:06.9427856Z adding: test/test-reports/python-pytest/dynamo.test_model_output/dynamo.test_model_output-0d371bc3cd364085.json (deflated 98%) 2025-12-04T10:11:06.9428812Z adding: test/test-reports/python-pytest/inductor.test_extension_backend/inductor.test_extension_backend-4a93539bbcae0552.json (stored 0%) 2025-12-04T10:11:06.9431776Z adding: test/test-reports/python-pytest/inductor.test_extension_backend/inductor.test_extension_backend-184a28db46f4749e.json (deflated 98%) 2025-12-04T10:11:06.9432779Z adding: test/test-reports/python-pytest/inductor.test_cooperative_reductions/inductor.test_cooperative_reductions-fc3209be19432af6.json (stored 0%) 2025-12-04T10:11:06.9433656Z adding: test/test-reports/python-pytest/inductor.test_cooperative_reductions/inductor.test_cooperative_reductions-61458c2618582136.json (stored 0%) 2025-12-04T10:11:06.9434476Z adding: test/test-reports/python-pytest/dynamo.test_fx_graph_runnable/dynamo.test_fx_graph_runnable-ddbc1ae1b738e2f4.json (stored 0%) 2025-12-04T10:11:06.9483492Z adding: test/test-reports/python-pytest/dynamo.test_fx_graph_runnable/dynamo.test_fx_graph_runnable-71d7146dfdd25cde.json (deflated 99%) 2025-12-04T10:11:06.9484752Z adding: test/test-reports/python-pytest/dynamo.test_backends/dynamo.test_backends-8659746a8e039725.json (stored 0%) 2025-12-04T10:11:06.9485651Z adding: test/test-reports/python-pytest/dynamo.test_backends/dynamo.test_backends-2bf853a9ce3082ab.json (stored 0%) 2025-12-04T10:11:06.9486522Z adding: test/test-reports/python-pytest/dynamo.test_functions/dynamo.test_functions-891767433e56c016.json (stored 0%) 2025-12-04T10:11:06.9487359Z adding: test/test-reports/python-pytest/dynamo.test_functions/dynamo.test_functions-74b8016243633845.json (stored 0%) 2025-12-04T10:11:06.9488296Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-adb9352cbf30e883.json (stored 0%) 2025-12-04T10:11:06.9489340Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7d44f20bb1c46c3f.json (deflated 98%) 2025-12-04T10:11:06.9490289Z adding: test/test-reports/python-pytest/inductor.test_padding/inductor.test_padding-2d400f1d2457cdfd.json (stored 0%) 2025-12-04T10:11:06.9491133Z adding: test/test-reports/python-pytest/inductor.test_padding/inductor.test_padding-764f2c6729c80dee.json (stored 0%) 2025-12-04T10:11:06.9491936Z adding: test/test-reports/python-pytest/dynamo.test_dicts/dynamo.test_dicts-4379238afe4ffe3d.json (stored 0%) 2025-12-04T10:11:06.9492716Z adding: test/test-reports/python-pytest/dynamo.test_dicts/dynamo.test_dicts-2ba8c58b73d80421.json (stored 0%) 2025-12-04T10:11:06.9493594Z adding: test/test-reports/python-pytest/dynamo.test_aot_compile/dynamo.test_aot_compile-79eacfeb4f1e860c.json (stored 0%) 2025-12-04T10:11:06.9494289Z adding: test/test-reports/python-pytest/dynamo.test_aot_compile/dynamo.test_aot_compile-603f3f2f0f262772.json (stored 0%) 2025-12-04T10:11:06.9495015Z adding: test/test-reports/python-pytest/dynamo.test_sets/dynamo.test_sets-135ba238b392c31c.json (stored 0%) 2025-12-04T10:11:06.9495624Z adding: test/test-reports/python-pytest/dynamo.test_sets/dynamo.test_sets-03077bc01df815c9.json (stored 0%) 2025-12-04T10:11:06.9496257Z adding: test/test-reports/python-pytest/dynamo.test_callback/dynamo.test_callback-d90a836cf7319d29.json (stored 0%) 2025-12-04T10:11:06.9496922Z adding: test/test-reports/python-pytest/dynamo.test_callback/dynamo.test_callback-9aebdb8d0ae9ab09.json (stored 0%) 2025-12-04T10:11:06.9497757Z adding: test/test-reports/python-pytest/inductor.test_cudagraph_trees_expandable_segments/inductor.test_cudagraph_trees_expandable_segments-87db9d88f872c44a.json (stored 0%) 2025-12-04T10:11:06.9506181Z adding: test/test-reports/python-pytest/inductor.test_cudagraph_trees_expandable_segments/inductor.test_cudagraph_trees_expandable_segments-06ab583be2814889.json (deflated 99%) 2025-12-04T10:11:06.9507321Z adding: test/test-reports/python-pytest/dynamo.test_exceptions/dynamo.test_exceptions-68332230799da90f.json (stored 0%) 2025-12-04T10:11:06.9508201Z adding: test/test-reports/python-pytest/dynamo.test_exceptions/dynamo.test_exceptions-d0179208da9e5b37.json (stored 0%) 2025-12-04T10:11:06.9518141Z adding: test/test-reports/python-pytest/test_cuda/test_cuda-0b3575165da88d35.json (deflated 99%) 2025-12-04T10:11:06.9525848Z adding: test/test-reports/python-pytest/test_cuda/test_cuda-8b00a0d732cb0448.json (deflated 98%) 2025-12-04T10:11:06.9526644Z adding: test/test-reports/python-pytest/test_transformers/test_transformers-0aff76852e040d8e.json (stored 0%) 2025-12-04T10:11:06.9527449Z adding: test/test-reports/python-pytest/test_transformers/test_transformers-d83a44de2640a948.json (stored 0%) 2025-12-04T10:11:06.9528346Z adding: test/test-reports/python-pytest/higher_order_ops.test_local_map/higher_order_ops.test_local_map-c325eb7c48529fe1.json (stored 0%) 2025-12-04T10:11:06.9529304Z adding: test/test-reports/python-pytest/higher_order_ops.test_local_map/higher_order_ops.test_local_map-218c693b3a0f40d8.json (stored 0%) 2025-12-04T10:11:06.9530164Z adding: test/test-reports/python-pytest/test_dataloader/test_dataloader-c3ab56b023ec83c9.json (stored 0%) 2025-12-04T10:11:06.9531122Z adding: test/test-reports/python-pytest/test_dataloader/test_dataloader-ce202fcb501e6304.json (stored 0%) 2025-12-04T10:11:06.9531853Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-1d76070ccdaa412b.json (stored 0%) 2025-12-04T10:11:06.9532482Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-d3940d286152db68.json (stored 0%) 2025-12-04T10:11:06.9533014Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-7a98f01e940d9313.json (stored 0%) 2025-12-04T10:11:06.9533626Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-21da0f73c4549c1e.json (stored 0%) 2025-12-04T10:11:06.9534184Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-ac6cc9ee3eedc42e.json (stored 0%) 2025-12-04T10:11:06.9534728Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-866538190db3f9f9.json (stored 0%) 2025-12-04T10:11:06.9535282Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-5ef04743f41caed7.json (stored 0%) 2025-12-04T10:11:06.9535972Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-26241f1d9fe6a84a.json (stored 0%) 2025-12-04T10:11:06.9536508Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-3733592a1d6e9f3d.json (stored 0%) 2025-12-04T10:11:06.9537055Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-a49ca6736f66875e.json (deflated 93%) 2025-12-04T10:11:06.9537601Z adding: test/test-reports/python-pytest/test_ops/test_ops-6e21a7eebf8ff2d6.json (stored 0%) 2025-12-04T10:11:06.9538196Z adding: test/test-reports/python-pytest/test_ops/test_ops-3e23a1ec37f874be.json (stored 0%) 2025-12-04T10:11:06.9538705Z adding: test/test-reports/python-pytest/test_ops/test_ops-5e4cfed161e0a0a1.json (stored 0%) 2025-12-04T10:11:06.9695638Z adding: test/test-reports/python-pytest/test_ops/test_ops-3bbe4a7adb52bac4.json (deflated 99%) 2025-12-04T10:11:06.9696435Z adding: test/test-reports/python-pytest/functorch.test_dims/functorch.test_dims-46a4cc4d3955dce2.json (stored 0%) 2025-12-04T10:11:06.9707431Z adding: test/test-reports/python-pytest/functorch.test_dims/functorch.test_dims-09de3759aef74478.json (deflated 99%) 2025-12-04T10:11:06.9708341Z adding: test/test-reports/python-pytest/test_torchfuzz_repros/test_torchfuzz_repros-8e06ea0d070c14c7.json (stored 0%) 2025-12-04T10:11:06.9709200Z adding: test/test-reports/python-pytest/test_torchfuzz_repros/test_torchfuzz_repros-184e97a40463bc74.json (stored 0%) 2025-12-04T10:11:06.9710103Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-9fcfc42987163c90.json (stored 0%) 2025-12-04T10:11:06.9711058Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-6af558fb08ec07a5.json (stored 0%) 2025-12-04T10:11:06.9924312Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-dbac887b32dd50ab.json (deflated 98%) 2025-12-04T10:11:07.0569680Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-22efee56a99319ab.json (deflated 98%) 2025-12-04T10:11:07.0570748Z adding: test/test-reports/python-pytest/inductor.test_kernel_benchmark/inductor.test_kernel_benchmark-1b89aea89d8a1e3b.json (stored 0%) 2025-12-04T10:11:07.0597820Z adding: test/test-reports/python-pytest/inductor.test_kernel_benchmark/inductor.test_kernel_benchmark-8244c39f5fb36d0d.json (deflated 99%) 2025-12-04T10:11:07.0598897Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-c8a0664bdf5f16cf.json (stored 0%) 2025-12-04T10:11:07.0599954Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-4be68e197594ecf9.json (stored 0%) 2025-12-04T10:11:07.0600989Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-4814c98fafa7f915.json (stored 0%) 2025-12-04T10:11:07.0602023Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-e6a790fd2716c9ce.json (stored 0%) 2025-12-04T10:11:07.0603212Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-b56e595397fe8756.json (deflated 99%) 2025-12-04T10:11:07.0604080Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-0b2c1e92820d7d68.json (deflated 98%) 2025-12-04T10:11:07.0605128Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-05457fa7ee6b6307.json (deflated 98%) 2025-12-04T10:11:07.0605985Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-6f7410ff779c9f27.json (stored 0%) 2025-12-04T10:11:07.0606780Z adding: test/test-reports/python-pytest/inductor.test_layout_optim/inductor.test_layout_optim-0bd77606ca486567.json (stored 0%) 2025-12-04T10:11:07.0607514Z adding: test/test-reports/python-pytest/inductor.test_layout_optim/inductor.test_layout_optim-081f5af62216a830.json (stored 0%) 2025-12-04T10:11:07.0608293Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor_arrayref/inductor.test_aot_inductor_arrayref-f384b91aa578f585.json (stored 0%) 2025-12-04T10:11:07.0609109Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor_arrayref/inductor.test_aot_inductor_arrayref-51f2220a835161fe.json (stored 0%) 2025-12-04T10:11:07.0610087Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_strided_blocks/inductor.test_torchinductor_strided_blocks-ea8814c1f4889964.json (stored 0%) 2025-12-04T10:11:07.0633043Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_strided_blocks/inductor.test_torchinductor_strided_blocks-9737d64a17925152.json (deflated 99%) 2025-12-04T10:11:07.0634131Z adding: test/test-reports/python-pytest/test_custom_ops/test_custom_ops-efe082e4ab0a44ae.json (stored 0%) 2025-12-04T10:11:07.0634894Z adding: test/test-reports/python-pytest/test_custom_ops/test_custom_ops-c87b73c0dc85593b.json (deflated 98%) 2025-12-04T10:11:07.0635666Z adding: test/test-reports/python-pytest/test_content_store/test_content_store-e26a61a7c77ed768.json (stored 0%) 2025-12-04T10:11:07.0636452Z adding: test/test-reports/python-pytest/test_content_store/test_content_store-a61adca807bbc325.json (stored 0%) 2025-12-04T10:11:07.0637306Z adding: test/test-reports/python-pytest/inductor.test_flex_decoding/inductor.test_flex_decoding-7745b7464b29f22e.json (stored 0%) 2025-12-04T10:11:07.0638260Z adding: test/test-reports/python-pytest/inductor.test_flex_decoding/inductor.test_flex_decoding-ba88f8b8af9ad9df.json (stored 0%) 2025-12-04T10:11:07.0639198Z adding: test/test-reports/python-pytest/inductor.test_flex_decoding/inductor.test_flex_decoding-59c9d120af700297.json (stored 0%) 2025-12-04T10:11:07.0640128Z adding: test/test-reports/python-pytest/inductor.test_flex_decoding/inductor.test_flex_decoding-4bc3b9a2985b9ca4.json (stored 0%) 2025-12-04T10:11:07.0641067Z adding: test/test-reports/python-pytest/inductor.test_deterministic/inductor.test_deterministic-5cb12759f42ba3ba.json (stored 0%) 2025-12-04T10:11:07.0642010Z adding: test/test-reports/python-pytest/inductor.test_deterministic/inductor.test_deterministic-46ebc55a5876c541.json (stored 0%) 2025-12-04T10:11:07.0642867Z adding: test/test-reports/python-pytest/inductor.test_b2b_gemm/inductor.test_b2b_gemm-2ff610447b04c70b.json (stored 0%) 2025-12-04T10:11:07.0643544Z adding: test/test-reports/python-pytest/inductor.test_b2b_gemm/inductor.test_b2b_gemm-f3a614691e2e3a64.json (stored 0%) 2025-12-04T10:11:07.0644220Z adding: test/test-reports/python-pytest/export.test_tree_utils/export.test_tree_utils-05f2e9a7741495f8.json (stored 0%) 2025-12-04T10:11:07.0644890Z adding: test/test-reports/python-pytest/export.test_tree_utils/export.test_tree_utils-6c4b1c3428f21437.json (stored 0%) 2025-12-04T10:11:07.0645595Z adding: test/test-reports/python-pytest/inductor.test_triton_wrapper/inductor.test_triton_wrapper-c2ec4261f482965d.json (stored 0%) 2025-12-04T10:11:07.0646474Z adding: test/test-reports/python-pytest/inductor.test_triton_wrapper/inductor.test_triton_wrapper-24ddba239b87c5fd.json (stored 0%) 2025-12-04T10:11:07.0647269Z adding: test/test-reports/python-pytest/inductor.test_static_cuda_launcher/inductor.test_static_cuda_launcher-49f744a2d7796139.json (stored 0%) 2025-12-04T10:11:07.0648087Z adding: test/test-reports/python-pytest/inductor.test_static_cuda_launcher/inductor.test_static_cuda_launcher-add7e1c387913324.json (stored 0%) 2025-12-04T10:11:07.0648861Z adding: test/test-reports/python-pytest/export.test_dynamic_shapes/export.test_dynamic_shapes-66fe2c3481907496.json (stored 0%) 2025-12-04T10:11:07.0649589Z adding: test/test-reports/python-pytest/export.test_dynamic_shapes/export.test_dynamic_shapes-80cb687c234e2e79.json (stored 0%) 2025-12-04T10:11:07.0650265Z adding: test/test-reports/python-pytest/dynamo.test_sdpa/dynamo.test_sdpa-bdc7e48df97c3ca8.json (stored 0%) 2025-12-04T10:11:07.0650892Z adding: test/test-reports/python-pytest/dynamo.test_sdpa/dynamo.test_sdpa-49a3e779aba88f83.json (stored 0%) 2025-12-04T10:11:07.0651606Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor_package/inductor.test_aot_inductor_package-3ed92bfad98c98b2.json (stored 0%) 2025-12-04T10:11:07.0652432Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor_package/inductor.test_aot_inductor_package-b829514935b5c71f.json (stored 0%) 2025-12-04T10:11:07.0653411Z adding: test/test-reports/python-pytest/inductor.test_compiled_optimizers/inductor.test_compiled_optimizers-fc3fdb63f48e8ded.json (stored 0%) 2025-12-04T10:11:07.0654529Z adding: test/test-reports/python-pytest/inductor.test_compiled_optimizers/inductor.test_compiled_optimizers-bccb353119aa7e57.json (stored 0%) 2025-12-04T10:11:07.0655340Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor_utils/inductor.test_aot_inductor_utils-6fb3d946f8c4ffd3.json (stored 0%) 2025-12-04T10:11:07.0656134Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor_utils/inductor.test_aot_inductor_utils-d291c39c695faa5d.json (stored 0%) 2025-12-04T10:11:07.0656886Z adding: test/test-reports/python-pytest/inductor.test_control_flow/inductor.test_control_flow-8be7394d621c6360.json (stored 0%) 2025-12-04T10:11:07.0657609Z adding: test/test-reports/python-pytest/inductor.test_control_flow/inductor.test_control_flow-11a5ec05df23ef24.json (stored 0%) 2025-12-04T10:11:07.0658273Z adding: test/test-reports/python-pytest/test_mkl_verbose/test_mkl_verbose-13dd2bd43e71ccf7.json (stored 0%) 2025-12-04T10:11:07.0658870Z adding: test/test-reports/python-pytest/test_mkl_verbose/test_mkl_verbose-f26bddd178ca728a.json (stored 0%) 2025-12-04T10:11:07.0659509Z adding: test/test-reports/python-pytest/test_comparison_utils/test_comparison_utils-54963d4d516ccde5.json (stored 0%) 2025-12-04T10:11:07.0660184Z adding: test/test-reports/python-pytest/test_comparison_utils/test_comparison_utils-d3a1a704cb9d5665.json (stored 0%) 2025-12-04T10:11:07.0660885Z adding: test/test-reports/python-pytest/functorch.test_ac_logging/functorch.test_ac_logging-2dd9f98043af7cf7.json (stored 0%) 2025-12-04T10:11:07.0661606Z adding: test/test-reports/python-pytest/functorch.test_ac_logging/functorch.test_ac_logging-161e2c94aebfad09.json (stored 0%) 2025-12-04T10:11:07.0662291Z adding: test/test-reports/python-pytest/test_mkldnn_verbose/test_mkldnn_verbose-d7761a02a830b7c0.json (stored 0%) 2025-12-04T10:11:07.0662932Z adding: test/test-reports/python-pytest/test_mkldnn_verbose/test_mkldnn_verbose-43202a21dc76fbd5.json (stored 0%) 2025-12-04T10:11:07.0663600Z adding: test/test-reports/python-pytest/test_utils_config_module/test_utils_config_module-472c75b33322e8a7.json (stored 0%) 2025-12-04T10:11:07.0664309Z adding: test/test-reports/python-pytest/test_utils_config_module/test_utils_config_module-262aead0bb49178e.json (stored 0%) 2025-12-04T10:11:07.0664940Z adding: test/test-reports/python-pytest/test_hop_infra/test_hop_infra-2ef307420583b1a3.json (stored 0%) 2025-12-04T10:11:07.0665595Z adding: test/test-reports/python-pytest/test_hop_infra/test_hop_infra-d73563c119a036ea.json (stored 0%) 2025-12-04T10:11:07.0666272Z adding: test/test-reports/python-pytest/test_appending_byte_serializer/test_appending_byte_serializer-7b18d6c4f7206cf9.json (stored 0%) 2025-12-04T10:11:07.0667052Z adding: test/test-reports/python-pytest/test_appending_byte_serializer/test_appending_byte_serializer-4e76a1b178e5d348.json (stored 0%) 2025-12-04T10:11:07.0667713Z adding: test/test-reports/python-pytest/test_license/test_license-787c69b7e1011532.json (stored 0%) 2025-12-04T10:11:07.0668273Z adding: test/test-reports/python-pytest/test_license/test_license-1d2b2367c347e96e.json (stored 0%) 2025-12-04T10:11:07.0686351Z adding: test/test-reports/python-pytest/test_ao_sparsity/test_ao_sparsity-ce7eaf33f7dbaec9.json (stored 0%) 2025-12-04T10:11:07.0687518Z adding: test/test-reports/python-pytest/test_ao_sparsity/test_ao_sparsity-08192502e661e963.json (stored 0%) 2025-12-04T10:11:07.0688328Z adding: test/test-reports/python-pytest/test_autoload/test_autoload-e264c515ca0f679c.json (stored 0%) 2025-12-04T10:11:07.0688966Z adding: test/test-reports/python-pytest/test_autoload/test_autoload-f8c034a1e389653e.json (stored 0%) 2025-12-04T10:11:07.0689693Z adding: test/test-reports/python-pytest/nn.attention.test_open_registry/nn.attention.test_open_registry-96e12afd5c2c5834.json (stored 0%) 2025-12-04T10:11:07.0690593Z adding: test/test-reports/python-pytest/nn.attention.test_open_registry/nn.attention.test_open_registry-7b14a30ba5e47451.json (stored 0%) 2025-12-04T10:11:07.0691297Z adding: test/test-reports/python-pytest/test_as_strided/test_as_strided-498f7726c4fd5386.json (stored 0%) 2025-12-04T10:11:07.0691962Z adding: test/test-reports/python-pytest/test_as_strided/test_as_strided-902df6e2e575efe4.json (stored 0%) 2025-12-04T10:11:07.0692583Z adding: test/test-reports/python-pytest/test_foreach/test_foreach-7dea0e9af6e9ef83.json (stored 0%) 2025-12-04T10:11:07.0693387Z adding: test/test-reports/python-pytest/test_foreach/test_foreach-014b7fdc7e3e1cf2.json (deflated 99%) 2025-12-04T10:11:07.0693994Z adding: test/test-reports/python-pytest/xpu.test_gemm/xpu.test_gemm-e12f724e557e6d57.json (stored 0%) 2025-12-04T10:11:07.0694563Z adding: test/test-reports/python-pytest/xpu.test_gemm/xpu.test_gemm-3a4dcb0c73b9942b.json (stored 0%) 2025-12-04T10:11:07.0695250Z adding: test/test-reports/python-pytest/higher_order_ops.test_print/higher_order_ops.test_print-4bc80d64e1280602.json (stored 0%) 2025-12-04T10:11:07.0696020Z adding: test/test-reports/python-pytest/higher_order_ops.test_print/higher_order_ops.test_print-8bf101c57cf0f824.json (stored 0%) 2025-12-04T10:11:07.0696759Z adding: test/test-reports/python-pytest/test_per_overload_api/test_per_overload_api-21ff1c6aab58c94d.json (stored 0%) 2025-12-04T10:11:07.0697428Z adding: test/test-reports/python-pytest/test_per_overload_api/test_per_overload_api-7ad42b77bba6a3a2.json (stored 0%) 2025-12-04T10:11:07.0698198Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_einsum/torch_np.numpy_tests.core.test_einsum-e842edbf10eb231f.json (stored 0%) 2025-12-04T10:11:07.0699059Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_einsum/torch_np.numpy_tests.core.test_einsum-7f41ab7e4542d2ff.json (stored 0%) 2025-12-04T10:11:07.0699808Z adding: test/test-reports/python-pytest/test_out_dtype_op/test_out_dtype_op-000e0deed8d272bf.json (stored 0%) 2025-12-04T10:11:07.0700428Z adding: test/test-reports/python-pytest/test_out_dtype_op/test_out_dtype_op-00e91616e649df2c.json (stored 0%) 2025-12-04T10:11:07.0701167Z adding: test/test-reports/python-pytest/torch_np.test_ufuncs_basic/torch_np.test_ufuncs_basic-f7d6a1f02b6919cc.json (stored 0%) 2025-12-04T10:11:07.0705247Z adding: test/test-reports/python-pytest/torch_np.test_ufuncs_basic/torch_np.test_ufuncs_basic-c8fa7ec0bb170cfd.json (stored 0%) 2025-12-04T10:11:07.0706687Z adding: test/test-reports/python-pytest/lazy.test_step_closures/lazy.test_step_closures-b9a26ea3a1d388cc.json (stored 0%) 2025-12-04T10:11:07.0707999Z adding: test/test-reports/python-pytest/lazy.test_step_closures/lazy.test_step_closures-ed58f93fb889064c.json (stored 0%) 2025-12-04T10:11:07.0709395Z adding: test/test-reports/python-pytest/functorch.dim.test_getsetitem/functorch.dim.test_getsetitem-54dfd23e9b8e2df1.json (stored 0%) 2025-12-04T10:11:07.0710807Z adding: test/test-reports/python-pytest/functorch.dim.test_getsetitem/functorch.dim.test_getsetitem-674883abaf9ddbc7.json (stored 0%) 2025-12-04T10:11:07.0712007Z adding: test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-4bb6f0d5c3842f89.json (stored 0%) 2025-12-04T10:11:07.0713206Z adding: test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-bc1bcd823497d313.json (stored 0%) 2025-12-04T10:11:07.0714301Z adding: test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-5472f3d31fbed190.json (deflated 98%) 2025-12-04T10:11:07.0715421Z adding: test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-8f921b290e79acad.json (deflated 98%) 2025-12-04T10:11:07.0716394Z adding: test/test-reports/python-pytest/test_meta/test_meta-b802ca42a17a2408.json (stored 0%) 2025-12-04T10:11:07.0717238Z adding: test/test-reports/python-pytest/test_meta/test_meta-95507d9e0beebf05.json (stored 0%) 2025-12-04T10:11:07.0718314Z adding: test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-49efc715bdff07e3.json (stored 0%) 2025-12-04T10:11:07.0719338Z adding: test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-66fecba3f80d413a.json (stored 0%) 2025-12-04T10:11:07.0720453Z adding: test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-2defe60eca797d43.json (stored 0%) 2025-12-04T10:11:07.0721449Z adding: test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-f3843a4cb154c125.json (stored 0%) 2025-12-04T10:11:07.0722416Z adding: test/test-reports/python-pytest/test_modules/test_modules-dedf31149329b85c.json (stored 0%) 2025-12-04T10:11:07.0723387Z adding: test/test-reports/python-pytest/test_modules/test_modules-e4b5d71a8ab0cbdf.json (deflated 98%) 2025-12-04T10:11:07.0724345Z adding: test/test-reports/python-pytest/test_tensorboard/test_tensorboard-2f994ca1a93f319e.json (stored 0%) 2025-12-04T10:11:07.0725320Z adding: test/test-reports/python-pytest/test_tensorboard/test_tensorboard-81b720b0e25291b8.json (stored 0%) 2025-12-04T10:11:07.0726527Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_indexing/torch_np.numpy_tests.core.test_indexing-273e2720b4753f9b.json (stored 0%) 2025-12-04T10:11:07.0727921Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_indexing/torch_np.numpy_tests.core.test_indexing-1a695318c456148e.json (stored 0%) 2025-12-04T10:11:07.0729050Z adding: test/test-reports/python-pytest/test_futures/test_futures-e1ab9c636dcc5536.json (stored 0%) 2025-12-04T10:11:07.0729799Z adding: test/test-reports/python-pytest/test_futures/test_futures-eaeb3a3297f2467a.json (stored 0%) 2025-12-04T10:11:07.0730391Z adding: test/test-reports/python-pytest/nn.test_dropout/nn.test_dropout-3a6a2fbf43a8e461.json (stored 0%) 2025-12-04T10:11:07.0730993Z adding: test/test-reports/python-pytest/nn.test_dropout/nn.test_dropout-d1d94a3c4456c153.json (stored 0%) 2025-12-04T10:11:07.0731660Z adding: test/test-reports/python-pytest/functorch.dim.test_split/functorch.dim.test_split-8adecf644c26bf14.json (stored 0%) 2025-12-04T10:11:07.0732388Z adding: test/test-reports/python-pytest/functorch.dim.test_split/functorch.dim.test_split-ac52710017bc0ec1.json (stored 0%) 2025-12-04T10:11:07.0733293Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.lib.test_type_check/torch_np.numpy_tests.lib.test_type_check-8efd39d4632ebedf.json (stored 0%) 2025-12-04T10:11:07.0734164Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.lib.test_type_check/torch_np.numpy_tests.lib.test_type_check-b366c3bac00b0d6b.json (stored 0%) 2025-12-04T10:11:07.0735165Z adding: test/test-reports/python-pytest/cpp_extensions.test_libtorch_agnostic/cpp_extensions.test_libtorch_agnostic-9de1e21a730c2263.json (stored 0%) 2025-12-04T10:11:07.0736064Z adding: test/test-reports/python-pytest/cpp_extensions.test_libtorch_agnostic/cpp_extensions.test_libtorch_agnostic-8a9c37e2e1da6c28.json (deflated 98%) 2025-12-04T10:11:07.0736894Z adding: test/test-reports/python-pytest/profiler.test_execution_trace/profiler.test_execution_trace-852e4822f8230313.json (stored 0%) 2025-12-04T10:11:07.0737670Z adding: test/test-reports/python-pytest/profiler.test_execution_trace/profiler.test_execution_trace-2d1f3ef34837b171.json (stored 0%) 2025-12-04T10:11:07.0738312Z adding: test/test-reports/python-pytest/test_jit/test_jit-711c2ccc279f7210.json (stored 0%) 2025-12-04T10:11:07.0738841Z adding: test/test-reports/python-pytest/test_jit/test_jit-4dc671ed81b7895e.json (deflated 99%) 2025-12-04T10:11:07.0739406Z adding: test/test-reports/python-pytest/test_datapipe/test_datapipe-0e1964a956d0a480.json (stored 0%) 2025-12-04T10:11:07.0739985Z adding: test/test-reports/python-pytest/test_datapipe/test_datapipe-82b13967e74e19a2.json (stored 0%) 2025-12-04T10:11:07.0740616Z adding: test/test-reports/python-pytest/test_numba_integration/test_numba_integration-844fbe0ccaacabee.json (stored 0%) 2025-12-04T10:11:07.0741306Z adding: test/test-reports/python-pytest/test_numba_integration/test_numba_integration-6bf5efdb972a9a8f.json (stored 0%) 2025-12-04T10:11:07.0742025Z adding: test/test-reports/python-pytest/test_functional_optim/test_functional_optim-182d833e6572f213.json (stored 0%) 2025-12-04T10:11:07.0742741Z adding: test/test-reports/python-pytest/test_functional_optim/test_functional_optim-c14b0e455295dcf4.json (stored 0%) 2025-12-04T10:11:07.0743385Z adding: test/test-reports/python-pytest/test_maskedtensor/test_maskedtensor-2e832fb0545eaaab.json (stored 0%) 2025-12-04T10:11:07.0744023Z adding: test/test-reports/python-pytest/test_maskedtensor/test_maskedtensor-d6bf47e2c29f29e6.json (stored 0%) 2025-12-04T10:11:07.0744771Z adding: test/test-reports/python-pytest/benchmark_utils.test_benchmark_utils/benchmark_utils.test_benchmark_utils-af96c8a6347bcdb1.json (stored 0%) 2025-12-04T10:11:07.0745635Z adding: test/test-reports/python-pytest/benchmark_utils.test_benchmark_utils/benchmark_utils.test_benchmark_utils-f962c5399d90fb14.json (stored 0%) 2025-12-04T10:11:07.0746503Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_scalarmath/torch_np.numpy_tests.core.test_scalarmath-b099d6ab8284837f.json (stored 0%) 2025-12-04T10:11:07.0747399Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_scalarmath/torch_np.numpy_tests.core.test_scalarmath-6b29237e24c20b32.json (stored 0%) 2025-12-04T10:11:07.0748189Z adding: test/test-reports/python-pytest/test_scaled_matmul_cuda/test_scaled_matmul_cuda-5972d960cfe7ec4e.json (stored 0%) 2025-12-04T10:11:07.0748879Z adding: test/test-reports/python-pytest/test_scaled_matmul_cuda/test_scaled_matmul_cuda-6e739739e0574a47.json (deflated 98%) 2025-12-04T10:11:07.0749660Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_shape_base/torch_np.numpy_tests.core.test_shape_base-ea32bc17692c1ce8.json (stored 0%) 2025-12-04T10:11:07.0750542Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_shape_base/torch_np.numpy_tests.core.test_shape_base-36c86d2bfdd67ddd.json (stored 0%) 2025-12-04T10:11:07.0751266Z adding: test/test-reports/python-pytest/test_vulkan/test_vulkan-8b9c4d8ede98ee3d.json (stored 0%) 2025-12-04T10:11:07.0751824Z adding: test/test-reports/python-pytest/test_vulkan/test_vulkan-e1a521ce9cf70bfb.json (stored 0%) 2025-12-04T10:11:07.0752434Z adding: test/test-reports/python-pytest/lazy.test_generator/lazy.test_generator-3b115040434e772f.json (stored 0%) 2025-12-04T10:11:07.0753080Z adding: test/test-reports/python-pytest/lazy.test_generator/lazy.test_generator-5262beb36f04167a.json (stored 0%) 2025-12-04T10:11:07.0753924Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.linalg.test_linalg/torch_np.numpy_tests.linalg.test_linalg-40abae860675e681.json (stored 0%) 2025-12-04T10:11:07.0754794Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.linalg.test_linalg/torch_np.numpy_tests.linalg.test_linalg-fa5a137aca79e3bc.json (stored 0%) 2025-12-04T10:11:07.0755636Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_dtype/torch_np.numpy_tests.core.test_dtype-fa2df09a508d13dc.json (stored 0%) 2025-12-04T10:11:07.0756477Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_dtype/torch_np.numpy_tests.core.test_dtype-6960413498a3b4a7.json (stored 0%) 2025-12-04T10:11:07.0757222Z adding: test/test-reports/python-pytest/lazy.test_debug_util/lazy.test_debug_util-4dae4a778d170bc1.json (stored 0%) 2025-12-04T10:11:07.0757889Z adding: test/test-reports/python-pytest/lazy.test_debug_util/lazy.test_debug_util-24c3e47e3d2b7323.json (stored 0%) 2025-12-04T10:11:07.0758548Z adding: test/test-reports/python-pytest/nn.test_load_state_dict/nn.test_load_state_dict-91536496375f24b6.json (stored 0%) 2025-12-04T10:11:07.0759210Z adding: test/test-reports/python-pytest/nn.test_load_state_dict/nn.test_load_state_dict-b877f1ca494eb282.json (stored 0%) 2025-12-04T10:11:07.0759837Z adding: test/test-reports/python-pytest/test_shape_ops/test_shape_ops-012d0fcc882a88fd.json (stored 0%) 2025-12-04T10:11:07.0760466Z adding: test/test-reports/python-pytest/test_shape_ops/test_shape_ops-feb6212001ee076f.json (stored 0%) 2025-12-04T10:11:07.0761072Z adding: test/test-reports/python-pytest/nn.test_module_hooks/nn.test_module_hooks-46ca7ec97ec9318c.json (stored 0%) 2025-12-04T10:11:07.0761752Z adding: test/test-reports/python-pytest/nn.test_module_hooks/nn.test_module_hooks-8ff29e8b05880501.json (stored 0%) 2025-12-04T10:11:07.0762526Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.lib.test_twodim_base/torch_np.numpy_tests.lib.test_twodim_base-2accadefabd28fef.json (stored 0%) 2025-12-04T10:11:07.0763413Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.lib.test_twodim_base/torch_np.numpy_tests.lib.test_twodim_base-fd2b4839449b1b20.json (stored 0%) 2025-12-04T10:11:07.0764242Z adding: test/test-reports/python-pytest/profiler.test_memory_profiler/profiler.test_memory_profiler-d6dc2fc9f3dab1f2.json (stored 0%) 2025-12-04T10:11:07.0765034Z adding: test/test-reports/python-pytest/profiler.test_memory_profiler/profiler.test_memory_profiler-89519a5c938e684c.json (deflated 98%) 2025-12-04T10:11:07.0765752Z adding: test/test-reports/python-pytest/test_jit_llga_fuser/test_jit_llga_fuser-644d80dd48236bd4.json (stored 0%) 2025-12-04T10:11:07.0766408Z adding: test/test-reports/python-pytest/test_jit_llga_fuser/test_jit_llga_fuser-d79703873568858a.json (stored 0%) 2025-12-04T10:11:07.0767047Z adding: test/test-reports/python-pytest/test_serialization/test_serialization-9d6f2802fe66a8b4.json (stored 0%) 2025-12-04T10:11:07.0767830Z adding: test/test-reports/python-pytest/test_serialization/test_serialization-d1a17d69bf98ee36.json (stored 0%) 2025-12-04T10:11:07.0768463Z adding: test/test-reports/python-pytest/test_sparse_csr/test_sparse_csr-1e945673d9d0e64f.json (stored 0%) 2025-12-04T10:11:07.0769063Z adding: test/test-reports/python-pytest/test_sparse_csr/test_sparse_csr-68a0a0e21687dd8c.json (deflated 98%) 2025-12-04T10:11:07.0769688Z adding: test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-258fb62f6ed3ad5d.json (stored 0%) 2025-12-04T10:11:07.0770331Z adding: test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-a4489a5f35bbfe70.json (stored 0%) 2025-12-04T10:11:07.0770971Z adding: test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-63eca64add6314a4.json (stored 0%) 2025-12-04T10:11:07.0771602Z adding: test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-4a923aaed07d8de4.json (stored 0%) 2025-12-04T10:11:07.0772210Z adding: test/test-reports/python-pytest/test_mkldnn/test_mkldnn-96ab41951b468da7.json (stored 0%) 2025-12-04T10:11:07.0772856Z adding: test/test-reports/python-pytest/test_mkldnn/test_mkldnn-3fd399b85d51ccfc.json (stored 0%) 2025-12-04T10:11:07.0813379Z ##[group]Run # Remove any previous test reports if they exist 2025-12-04T10:11:07.0813731Z # Remove any previous test reports if they exist 2025-12-04T10:11:07.0814009Z rm -f test-reports-*.zip 2025-12-04T10:11:07.0814354Z zip -r "test-reports-${FILE_SUFFIX}.zip" test/test-reports -i '*.xml' -i '*.csv' 2025-12-04T10:11:07.0821835Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:11:07.0822104Z env: 2025-12-04T10:11:07.0822253Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:07.0822436Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:07.0822672Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:07.0823061Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:07.0823405Z DEVICE_NAME: 2025-12-04T10:11:07.0823560Z DEVICE_TYPE: 2025-12-04T10:11:07.0823869Z FILE_SUFFIX: test-default-4-5-lf.linux.g6.4xlarge.experimental.nvidia.gpu_57116084869 2025-12-04T10:11:07.0824244Z ##[endgroup] 2025-12-04T10:11:07.0953534Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-78423a5d6c0345b8.xml (deflated 28%) 2025-12-04T10:11:07.0954529Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-863ed9428b922f00.xml (deflated 28%) 2025-12-04T10:11:07.0955573Z adding: test/test-reports/python-pytest/dynamo.test_model_output/dynamo.test_model_output-5ac7859c2c204934.xml (deflated 28%) 2025-12-04T10:11:07.0956537Z adding: test/test-reports/python-pytest/dynamo.test_model_output/dynamo.test_model_output-0d371bc3cd364085.xml (deflated 97%) 2025-12-04T10:11:07.0957494Z adding: test/test-reports/python-pytest/inductor.test_extension_backend/inductor.test_extension_backend-4a93539bbcae0552.xml (deflated 28%) 2025-12-04T10:11:07.0961046Z adding: test/test-reports/python-pytest/inductor.test_extension_backend/inductor.test_extension_backend-184a28db46f4749e.xml (deflated 98%) 2025-12-04T10:11:07.0962120Z adding: test/test-reports/python-pytest/inductor.test_cooperative_reductions/inductor.test_cooperative_reductions-fc3209be19432af6.xml (deflated 28%) 2025-12-04T10:11:07.0963211Z adding: test/test-reports/python-pytest/inductor.test_cooperative_reductions/inductor.test_cooperative_reductions-61458c2618582136.xml (deflated 27%) 2025-12-04T10:11:07.0964286Z adding: test/test-reports/python-pytest/dynamo.test_fx_graph_runnable/dynamo.test_fx_graph_runnable-ddbc1ae1b738e2f4.xml (deflated 28%) 2025-12-04T10:11:07.1010538Z adding: test/test-reports/python-pytest/dynamo.test_fx_graph_runnable/dynamo.test_fx_graph_runnable-71d7146dfdd25cde.xml (deflated 99%) 2025-12-04T10:11:07.1011490Z adding: test/test-reports/python-pytest/dynamo.test_backends/dynamo.test_backends-8659746a8e039725.xml (deflated 28%) 2025-12-04T10:11:07.1012324Z adding: test/test-reports/python-pytest/dynamo.test_backends/dynamo.test_backends-2bf853a9ce3082ab.xml (deflated 28%) 2025-12-04T10:11:07.1013000Z adding: test/test-reports/python-pytest/dynamo.test_functions/dynamo.test_functions-891767433e56c016.xml (deflated 28%) 2025-12-04T10:11:07.1013724Z adding: test/test-reports/python-pytest/dynamo.test_functions/dynamo.test_functions-74b8016243633845.xml (deflated 28%) 2025-12-04T10:11:07.1014487Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-adb9352cbf30e883.xml (deflated 28%) 2025-12-04T10:11:07.1015304Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7d44f20bb1c46c3f.xml (deflated 98%) 2025-12-04T10:11:07.1016059Z adding: test/test-reports/python-pytest/inductor.test_padding/inductor.test_padding-2d400f1d2457cdfd.xml (deflated 28%) 2025-12-04T10:11:07.1016750Z adding: test/test-reports/python-pytest/inductor.test_padding/inductor.test_padding-764f2c6729c80dee.xml (deflated 28%) 2025-12-04T10:11:07.1017558Z adding: test/test-reports/python-pytest/dynamo.test_dicts/dynamo.test_dicts-4379238afe4ffe3d.xml (deflated 28%) 2025-12-04T10:11:07.1018196Z adding: test/test-reports/python-pytest/dynamo.test_dicts/dynamo.test_dicts-2ba8c58b73d80421.xml (deflated 28%) 2025-12-04T10:11:07.1018871Z adding: test/test-reports/python-pytest/dynamo.test_aot_compile/dynamo.test_aot_compile-79eacfeb4f1e860c.xml (deflated 28%) 2025-12-04T10:11:07.1019578Z adding: test/test-reports/python-pytest/dynamo.test_aot_compile/dynamo.test_aot_compile-603f3f2f0f262772.xml (deflated 28%) 2025-12-04T10:11:07.1020240Z adding: test/test-reports/python-pytest/dynamo.test_sets/dynamo.test_sets-135ba238b392c31c.xml (deflated 27%) 2025-12-04T10:11:07.1020865Z adding: test/test-reports/python-pytest/dynamo.test_sets/dynamo.test_sets-03077bc01df815c9.xml (deflated 28%) 2025-12-04T10:11:07.1021515Z adding: test/test-reports/python-pytest/dynamo.test_callback/dynamo.test_callback-d90a836cf7319d29.xml (deflated 28%) 2025-12-04T10:11:07.1022194Z adding: test/test-reports/python-pytest/dynamo.test_callback/dynamo.test_callback-9aebdb8d0ae9ab09.xml (deflated 28%) 2025-12-04T10:11:07.1023048Z adding: test/test-reports/python-pytest/inductor.test_cudagraph_trees_expandable_segments/inductor.test_cudagraph_trees_expandable_segments-87db9d88f872c44a.xml (deflated 28%) 2025-12-04T10:11:07.1034068Z adding: test/test-reports/python-pytest/inductor.test_cudagraph_trees_expandable_segments/inductor.test_cudagraph_trees_expandable_segments-06ab583be2814889.xml (deflated 99%) 2025-12-04T10:11:07.1035247Z adding: test/test-reports/python-pytest/dynamo.test_exceptions/dynamo.test_exceptions-68332230799da90f.xml (deflated 28%) 2025-12-04T10:11:07.1036113Z adding: test/test-reports/python-pytest/dynamo.test_exceptions/dynamo.test_exceptions-d0179208da9e5b37.xml (deflated 28%) 2025-12-04T10:11:07.1044887Z adding: test/test-reports/python-pytest/test_cuda/test_cuda-0b3575165da88d35.xml (deflated 99%) 2025-12-04T10:11:07.1051231Z adding: test/test-reports/python-pytest/test_cuda/test_cuda-8b00a0d732cb0448.xml (deflated 98%) 2025-12-04T10:11:07.1051999Z adding: test/test-reports/python-pytest/test_transformers/test_transformers-0aff76852e040d8e.xml (deflated 28%) 2025-12-04T10:11:07.1052804Z adding: test/test-reports/python-pytest/test_transformers/test_transformers-d83a44de2640a948.xml (deflated 27%) 2025-12-04T10:11:07.1053783Z adding: test/test-reports/python-pytest/higher_order_ops.test_local_map/higher_order_ops.test_local_map-c325eb7c48529fe1.xml (deflated 28%) 2025-12-04T10:11:07.1054769Z adding: test/test-reports/python-pytest/higher_order_ops.test_local_map/higher_order_ops.test_local_map-218c693b3a0f40d8.xml (deflated 28%) 2025-12-04T10:11:07.1055663Z adding: test/test-reports/python-pytest/test_dataloader/test_dataloader-c3ab56b023ec83c9.xml (deflated 28%) 2025-12-04T10:11:07.1056442Z adding: test/test-reports/python-pytest/test_dataloader/test_dataloader-ce202fcb501e6304.xml (deflated 28%) 2025-12-04T10:11:07.1057167Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-1d76070ccdaa412b.xml (deflated 28%) 2025-12-04T10:11:07.1057868Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-d3940d286152db68.xml (deflated 27%) 2025-12-04T10:11:07.1058556Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-7a98f01e940d9313.xml (deflated 27%) 2025-12-04T10:11:07.1059242Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-21da0f73c4549c1e.xml (deflated 28%) 2025-12-04T10:11:07.1059925Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-ac6cc9ee3eedc42e.xml (deflated 27%) 2025-12-04T10:11:07.1060618Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-866538190db3f9f9.xml (deflated 28%) 2025-12-04T10:11:07.1061298Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-5ef04743f41caed7.xml (deflated 28%) 2025-12-04T10:11:07.1062094Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-26241f1d9fe6a84a.xml (deflated 27%) 2025-12-04T10:11:07.1062636Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-3733592a1d6e9f3d.xml (deflated 28%) 2025-12-04T10:11:07.1063183Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-a49ca6736f66875e.xml (deflated 90%) 2025-12-04T10:11:07.1063723Z adding: test/test-reports/python-pytest/test_ops/test_ops-6e21a7eebf8ff2d6.xml (deflated 27%) 2025-12-04T10:11:07.1064245Z adding: test/test-reports/python-pytest/test_ops/test_ops-3e23a1ec37f874be.xml (deflated 27%) 2025-12-04T10:11:07.1064767Z adding: test/test-reports/python-pytest/test_ops/test_ops-5e4cfed161e0a0a1.xml (deflated 28%) 2025-12-04T10:11:07.1220522Z adding: test/test-reports/python-pytest/test_ops/test_ops-3bbe4a7adb52bac4.xml (deflated 99%) 2025-12-04T10:11:07.1221375Z adding: test/test-reports/python-pytest/functorch.test_dims/functorch.test_dims-46a4cc4d3955dce2.xml (deflated 28%) 2025-12-04T10:11:07.1231799Z adding: test/test-reports/python-pytest/functorch.test_dims/functorch.test_dims-09de3759aef74478.xml (deflated 99%) 2025-12-04T10:11:07.1232894Z adding: test/test-reports/python-pytest/test_torchfuzz_repros/test_torchfuzz_repros-8e06ea0d070c14c7.xml (deflated 28%) 2025-12-04T10:11:07.1233755Z adding: test/test-reports/python-pytest/test_torchfuzz_repros/test_torchfuzz_repros-184e97a40463bc74.xml (deflated 28%) 2025-12-04T10:11:07.1234820Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-9fcfc42987163c90.xml (deflated 28%) 2025-12-04T10:11:07.1235785Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-6af558fb08ec07a5.xml (deflated 28%) 2025-12-04T10:11:07.1458880Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-dbac887b32dd50ab.xml (deflated 98%) 2025-12-04T10:11:07.2095663Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-22efee56a99319ab.xml (deflated 98%) 2025-12-04T10:11:07.2096706Z adding: test/test-reports/python-pytest/inductor.test_kernel_benchmark/inductor.test_kernel_benchmark-1b89aea89d8a1e3b.xml (deflated 28%) 2025-12-04T10:11:07.2123321Z adding: test/test-reports/python-pytest/inductor.test_kernel_benchmark/inductor.test_kernel_benchmark-8244c39f5fb36d0d.xml (deflated 99%) 2025-12-04T10:11:07.2124579Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-c8a0664bdf5f16cf.xml (deflated 28%) 2025-12-04T10:11:07.2125671Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-4be68e197594ecf9.xml (deflated 28%) 2025-12-04T10:11:07.2126979Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-4814c98fafa7f915.xml (deflated 28%) 2025-12-04T10:11:07.2128208Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-e6a790fd2716c9ce.xml (deflated 28%) 2025-12-04T10:11:07.2129614Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-b56e595397fe8756.xml (deflated 99%) 2025-12-04T10:11:07.2131236Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-0b2c1e92820d7d68.xml (deflated 98%) 2025-12-04T10:11:07.2132296Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-05457fa7ee6b6307.xml (deflated 98%) 2025-12-04T10:11:07.2133130Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-6f7410ff779c9f27.xml (deflated 28%) 2025-12-04T10:11:07.2134022Z adding: test/test-reports/python-pytest/inductor.test_layout_optim/inductor.test_layout_optim-0bd77606ca486567.xml (deflated 28%) 2025-12-04T10:11:07.2134771Z adding: test/test-reports/python-pytest/inductor.test_layout_optim/inductor.test_layout_optim-081f5af62216a830.xml (deflated 28%) 2025-12-04T10:11:07.2135726Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor_arrayref/inductor.test_aot_inductor_arrayref-f384b91aa578f585.xml (deflated 28%) 2025-12-04T10:11:07.2136581Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor_arrayref/inductor.test_aot_inductor_arrayref-51f2220a835161fe.xml (deflated 28%) 2025-12-04T10:11:07.2137458Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_strided_blocks/inductor.test_torchinductor_strided_blocks-ea8814c1f4889964.xml (deflated 28%) 2025-12-04T10:11:07.2156550Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_strided_blocks/inductor.test_torchinductor_strided_blocks-9737d64a17925152.xml (deflated 99%) 2025-12-04T10:11:07.2157942Z adding: test/test-reports/python-pytest/test_custom_ops/test_custom_ops-efe082e4ab0a44ae.xml (deflated 28%) 2025-12-04T10:11:07.2159232Z adding: test/test-reports/python-pytest/test_custom_ops/test_custom_ops-c87b73c0dc85593b.xml (deflated 98%) 2025-12-04T10:11:07.2160561Z adding: test/test-reports/python-pytest/test_content_store/test_content_store-e26a61a7c77ed768.xml (deflated 28%) 2025-12-04T10:11:07.2161866Z adding: test/test-reports/python-pytest/test_content_store/test_content_store-a61adca807bbc325.xml (deflated 28%) 2025-12-04T10:11:07.2163108Z adding: test/test-reports/python-pytest/inductor.test_flex_decoding/inductor.test_flex_decoding-7745b7464b29f22e.xml (deflated 28%) 2025-12-04T10:11:07.2164376Z adding: test/test-reports/python-pytest/inductor.test_flex_decoding/inductor.test_flex_decoding-ba88f8b8af9ad9df.xml (deflated 28%) 2025-12-04T10:11:07.2165631Z adding: test/test-reports/python-pytest/inductor.test_flex_decoding/inductor.test_flex_decoding-59c9d120af700297.xml (deflated 28%) 2025-12-04T10:11:07.2166807Z adding: test/test-reports/python-pytest/inductor.test_flex_decoding/inductor.test_flex_decoding-4bc3b9a2985b9ca4.xml (deflated 28%) 2025-12-04T10:11:07.2167984Z adding: test/test-reports/python-pytest/inductor.test_deterministic/inductor.test_deterministic-5cb12759f42ba3ba.xml (deflated 28%) 2025-12-04T10:11:07.2169129Z adding: test/test-reports/python-pytest/inductor.test_deterministic/inductor.test_deterministic-46ebc55a5876c541.xml (deflated 28%) 2025-12-04T10:11:07.2170244Z adding: test/test-reports/python-pytest/inductor.test_b2b_gemm/inductor.test_b2b_gemm-2ff610447b04c70b.xml (deflated 28%) 2025-12-04T10:11:07.2171305Z adding: test/test-reports/python-pytest/inductor.test_b2b_gemm/inductor.test_b2b_gemm-f3a614691e2e3a64.xml (deflated 28%) 2025-12-04T10:11:07.2172344Z adding: test/test-reports/python-pytest/export.test_tree_utils/export.test_tree_utils-05f2e9a7741495f8.xml (deflated 28%) 2025-12-04T10:11:07.2173424Z adding: test/test-reports/python-pytest/export.test_tree_utils/export.test_tree_utils-6c4b1c3428f21437.xml (deflated 28%) 2025-12-04T10:11:07.2174335Z adding: test/test-reports/python-pytest/inductor.test_triton_wrapper/inductor.test_triton_wrapper-c2ec4261f482965d.xml (deflated 28%) 2025-12-04T10:11:07.2175107Z adding: test/test-reports/python-pytest/inductor.test_triton_wrapper/inductor.test_triton_wrapper-24ddba239b87c5fd.xml (deflated 28%) 2025-12-04T10:11:07.2175909Z adding: test/test-reports/python-pytest/inductor.test_static_cuda_launcher/inductor.test_static_cuda_launcher-49f744a2d7796139.xml (deflated 28%) 2025-12-04T10:11:07.2176737Z adding: test/test-reports/python-pytest/inductor.test_static_cuda_launcher/inductor.test_static_cuda_launcher-add7e1c387913324.xml (deflated 28%) 2025-12-04T10:11:07.2177520Z adding: test/test-reports/python-pytest/export.test_dynamic_shapes/export.test_dynamic_shapes-66fe2c3481907496.xml (deflated 28%) 2025-12-04T10:11:07.2178255Z adding: test/test-reports/python-pytest/export.test_dynamic_shapes/export.test_dynamic_shapes-80cb687c234e2e79.xml (deflated 28%) 2025-12-04T10:11:07.2179017Z adding: test/test-reports/python-pytest/dynamo.test_sdpa/dynamo.test_sdpa-bdc7e48df97c3ca8.xml (deflated 28%) 2025-12-04T10:11:07.2180122Z adding: test/test-reports/python-pytest/dynamo.test_sdpa/dynamo.test_sdpa-49a3e779aba88f83.xml (deflated 28%) 2025-12-04T10:11:07.2181297Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor_package/inductor.test_aot_inductor_package-3ed92bfad98c98b2.xml (deflated 28%) 2025-12-04T10:11:07.2182549Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor_package/inductor.test_aot_inductor_package-b829514935b5c71f.xml (deflated 28%) 2025-12-04T10:11:07.2183812Z adding: test/test-reports/python-pytest/inductor.test_compiled_optimizers/inductor.test_compiled_optimizers-fc3fdb63f48e8ded.xml (deflated 28%) 2025-12-04T10:11:07.2184988Z adding: test/test-reports/python-pytest/inductor.test_compiled_optimizers/inductor.test_compiled_optimizers-bccb353119aa7e57.xml (deflated 28%) 2025-12-04T10:11:07.2185804Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor_utils/inductor.test_aot_inductor_utils-6fb3d946f8c4ffd3.xml (deflated 28%) 2025-12-04T10:11:07.2186773Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor_utils/inductor.test_aot_inductor_utils-d291c39c695faa5d.xml (deflated 28%) 2025-12-04T10:11:07.2187787Z adding: test/test-reports/python-pytest/inductor.test_control_flow/inductor.test_control_flow-8be7394d621c6360.xml (deflated 28%) 2025-12-04T10:11:07.2188896Z adding: test/test-reports/python-pytest/inductor.test_control_flow/inductor.test_control_flow-11a5ec05df23ef24.xml (deflated 28%) 2025-12-04T10:11:07.2189910Z adding: test/test-reports/python-pytest/test_mkl_verbose/test_mkl_verbose-13dd2bd43e71ccf7.xml (deflated 28%) 2025-12-04T10:11:07.2190592Z adding: test/test-reports/python-pytest/test_mkl_verbose/test_mkl_verbose-f26bddd178ca728a.xml (deflated 28%) 2025-12-04T10:11:07.2191592Z adding: test/test-reports/python-pytest/test_comparison_utils/test_comparison_utils-54963d4d516ccde5.xml (deflated 28%) 2025-12-04T10:11:07.2192600Z adding: test/test-reports/python-pytest/test_comparison_utils/test_comparison_utils-d3a1a704cb9d5665.xml (deflated 28%) 2025-12-04T10:11:07.2193690Z adding: test/test-reports/python-pytest/functorch.test_ac_logging/functorch.test_ac_logging-2dd9f98043af7cf7.xml (deflated 28%) 2025-12-04T10:11:07.2194792Z adding: test/test-reports/python-pytest/functorch.test_ac_logging/functorch.test_ac_logging-161e2c94aebfad09.xml (deflated 28%) 2025-12-04T10:11:07.2195672Z adding: test/test-reports/python-pytest/test_mkldnn_verbose/test_mkldnn_verbose-d7761a02a830b7c0.xml (deflated 28%) 2025-12-04T10:11:07.2196666Z adding: test/test-reports/python-pytest/test_mkldnn_verbose/test_mkldnn_verbose-43202a21dc76fbd5.xml (deflated 28%) 2025-12-04T10:11:07.2197348Z adding: test/test-reports/python-pytest/test_utils_config_module/test_utils_config_module-472c75b33322e8a7.xml (deflated 28%) 2025-12-04T10:11:07.2198051Z adding: test/test-reports/python-pytest/test_utils_config_module/test_utils_config_module-262aead0bb49178e.xml (deflated 28%) 2025-12-04T10:11:07.2198698Z adding: test/test-reports/python-pytest/test_hop_infra/test_hop_infra-2ef307420583b1a3.xml (deflated 28%) 2025-12-04T10:11:07.2199280Z adding: test/test-reports/python-pytest/test_hop_infra/test_hop_infra-d73563c119a036ea.xml (deflated 28%) 2025-12-04T10:11:07.2199988Z adding: test/test-reports/python-pytest/test_appending_byte_serializer/test_appending_byte_serializer-7b18d6c4f7206cf9.xml (deflated 28%) 2025-12-04T10:11:07.2200768Z adding: test/test-reports/python-pytest/test_appending_byte_serializer/test_appending_byte_serializer-4e76a1b178e5d348.xml (deflated 28%) 2025-12-04T10:11:07.2201443Z adding: test/test-reports/python-pytest/test_license/test_license-787c69b7e1011532.xml (deflated 28%) 2025-12-04T10:11:07.2202009Z adding: test/test-reports/python-pytest/test_license/test_license-1d2b2367c347e96e.xml (deflated 28%) 2025-12-04T10:11:07.2202602Z adding: test/test-reports/python-pytest/test_ao_sparsity/test_ao_sparsity-ce7eaf33f7dbaec9.xml (deflated 28%) 2025-12-04T10:11:07.2203339Z adding: test/test-reports/python-pytest/test_ao_sparsity/test_ao_sparsity-08192502e661e963.xml (deflated 28%) 2025-12-04T10:11:07.2204256Z adding: test/test-reports/python-pytest/test_autoload/test_autoload-e264c515ca0f679c.xml (deflated 28%) 2025-12-04T10:11:07.2205372Z adding: test/test-reports/python-pytest/test_autoload/test_autoload-f8c034a1e389653e.xml (deflated 28%) 2025-12-04T10:11:07.2206396Z adding: test/test-reports/python-pytest/nn.attention.test_open_registry/nn.attention.test_open_registry-96e12afd5c2c5834.xml (deflated 28%) 2025-12-04T10:11:07.2207509Z adding: test/test-reports/python-pytest/nn.attention.test_open_registry/nn.attention.test_open_registry-7b14a30ba5e47451.xml (deflated 28%) 2025-12-04T10:11:07.2208430Z adding: test/test-reports/python-pytest/test_as_strided/test_as_strided-498f7726c4fd5386.xml (deflated 28%) 2025-12-04T10:11:07.2209052Z adding: test/test-reports/python-pytest/test_as_strided/test_as_strided-902df6e2e575efe4.xml (deflated 28%) 2025-12-04T10:11:07.2209978Z adding: test/test-reports/python-pytest/test_foreach/test_foreach-7dea0e9af6e9ef83.xml (deflated 28%) 2025-12-04T10:11:07.2210850Z adding: test/test-reports/python-pytest/test_foreach/test_foreach-014b7fdc7e3e1cf2.xml (deflated 98%) 2025-12-04T10:11:07.2211735Z adding: test/test-reports/python-pytest/xpu.test_gemm/xpu.test_gemm-e12f724e557e6d57.xml (deflated 28%) 2025-12-04T10:11:07.2212606Z adding: test/test-reports/python-pytest/xpu.test_gemm/xpu.test_gemm-3a4dcb0c73b9942b.xml (deflated 28%) 2025-12-04T10:11:07.2213796Z adding: test/test-reports/python-pytest/higher_order_ops.test_print/higher_order_ops.test_print-4bc80d64e1280602.xml (deflated 28%) 2025-12-04T10:11:07.2214988Z adding: test/test-reports/python-pytest/higher_order_ops.test_print/higher_order_ops.test_print-8bf101c57cf0f824.xml (deflated 28%) 2025-12-04T10:11:07.2216017Z adding: test/test-reports/python-pytest/test_per_overload_api/test_per_overload_api-21ff1c6aab58c94d.xml (deflated 28%) 2025-12-04T10:11:07.2216692Z adding: test/test-reports/python-pytest/test_per_overload_api/test_per_overload_api-7ad42b77bba6a3a2.xml (deflated 28%) 2025-12-04T10:11:07.2217720Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_einsum/torch_np.numpy_tests.core.test_einsum-e842edbf10eb231f.xml (deflated 28%) 2025-12-04T10:11:07.2218963Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_einsum/torch_np.numpy_tests.core.test_einsum-7f41ab7e4542d2ff.xml (deflated 28%) 2025-12-04T10:11:07.2219714Z adding: test/test-reports/python-pytest/test_out_dtype_op/test_out_dtype_op-000e0deed8d272bf.xml (deflated 28%) 2025-12-04T10:11:07.2220589Z adding: test/test-reports/python-pytest/test_out_dtype_op/test_out_dtype_op-00e91616e649df2c.xml (deflated 28%) 2025-12-04T10:11:07.2221608Z adding: test/test-reports/python-pytest/torch_np.test_ufuncs_basic/torch_np.test_ufuncs_basic-f7d6a1f02b6919cc.xml (deflated 28%) 2025-12-04T10:11:07.2222486Z adding: test/test-reports/python-pytest/torch_np.test_ufuncs_basic/torch_np.test_ufuncs_basic-c8fa7ec0bb170cfd.xml (deflated 28%) 2025-12-04T10:11:07.2223202Z adding: test/test-reports/python-pytest/lazy.test_step_closures/lazy.test_step_closures-b9a26ea3a1d388cc.xml (deflated 28%) 2025-12-04T10:11:07.2223902Z adding: test/test-reports/python-pytest/lazy.test_step_closures/lazy.test_step_closures-ed58f93fb889064c.xml (deflated 28%) 2025-12-04T10:11:07.2224649Z adding: test/test-reports/python-pytest/functorch.dim.test_getsetitem/functorch.dim.test_getsetitem-54dfd23e9b8e2df1.xml (deflated 28%) 2025-12-04T10:11:07.2225433Z adding: test/test-reports/python-pytest/functorch.dim.test_getsetitem/functorch.dim.test_getsetitem-674883abaf9ddbc7.xml (deflated 28%) 2025-12-04T10:11:07.2226171Z adding: test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-4bb6f0d5c3842f89.xml (deflated 28%) 2025-12-04T10:11:07.2226853Z adding: test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-bc1bcd823497d313.xml (deflated 28%) 2025-12-04T10:11:07.2227746Z adding: test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-5472f3d31fbed190.xml (deflated 98%) 2025-12-04T10:11:07.2228927Z adding: test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-8f921b290e79acad.xml (deflated 98%) 2025-12-04T10:11:07.2229822Z adding: test/test-reports/python-pytest/test_meta/test_meta-b802ca42a17a2408.xml (deflated 27%) 2025-12-04T10:11:07.2230362Z adding: test/test-reports/python-pytest/test_meta/test_meta-95507d9e0beebf05.xml (deflated 28%) 2025-12-04T10:11:07.2230947Z adding: test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-49efc715bdff07e3.xml (deflated 28%) 2025-12-04T10:11:07.2231586Z adding: test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-66fecba3f80d413a.xml (deflated 28%) 2025-12-04T10:11:07.2232225Z adding: test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-2defe60eca797d43.xml (deflated 27%) 2025-12-04T10:11:07.2232859Z adding: test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-f3843a4cb154c125.xml (deflated 28%) 2025-12-04T10:11:07.2233465Z adding: test/test-reports/python-pytest/test_modules/test_modules-dedf31149329b85c.xml (deflated 28%) 2025-12-04T10:11:07.2234033Z adding: test/test-reports/python-pytest/test_modules/test_modules-e4b5d71a8ab0cbdf.xml (deflated 98%) 2025-12-04T10:11:07.2234626Z adding: test/test-reports/python-pytest/test_tensorboard/test_tensorboard-2f994ca1a93f319e.xml (deflated 28%) 2025-12-04T10:11:07.2235318Z adding: test/test-reports/python-pytest/test_tensorboard/test_tensorboard-81b720b0e25291b8.xml (deflated 28%) 2025-12-04T10:11:07.2236104Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_indexing/torch_np.numpy_tests.core.test_indexing-273e2720b4753f9b.xml (deflated 28%) 2025-12-04T10:11:07.2236963Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_indexing/torch_np.numpy_tests.core.test_indexing-1a695318c456148e.xml (deflated 28%) 2025-12-04T10:11:07.2237682Z adding: test/test-reports/python-pytest/test_futures/test_futures-e1ab9c636dcc5536.xml (deflated 28%) 2025-12-04T10:11:07.2238248Z adding: test/test-reports/python-pytest/test_futures/test_futures-eaeb3a3297f2467a.xml (deflated 28%) 2025-12-04T10:11:07.2238827Z adding: test/test-reports/python-pytest/nn.test_dropout/nn.test_dropout-3a6a2fbf43a8e461.xml (deflated 28%) 2025-12-04T10:11:07.2239418Z adding: test/test-reports/python-pytest/nn.test_dropout/nn.test_dropout-d1d94a3c4456c153.xml (deflated 28%) 2025-12-04T10:11:07.2240079Z adding: test/test-reports/python-pytest/functorch.dim.test_split/functorch.dim.test_split-8adecf644c26bf14.xml (deflated 28%) 2025-12-04T10:11:07.2240820Z adding: test/test-reports/python-pytest/functorch.dim.test_split/functorch.dim.test_split-ac52710017bc0ec1.xml (deflated 28%) 2025-12-04T10:11:07.2241608Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.lib.test_type_check/torch_np.numpy_tests.lib.test_type_check-8efd39d4632ebedf.xml (deflated 28%) 2025-12-04T10:11:07.2242472Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.lib.test_type_check/torch_np.numpy_tests.lib.test_type_check-b366c3bac00b0d6b.xml (deflated 28%) 2025-12-04T10:11:07.2243335Z adding: test/test-reports/python-pytest/cpp_extensions.test_libtorch_agnostic/cpp_extensions.test_libtorch_agnostic-9de1e21a730c2263.xml (deflated 28%) 2025-12-04T10:11:07.2244199Z adding: test/test-reports/python-pytest/cpp_extensions.test_libtorch_agnostic/cpp_extensions.test_libtorch_agnostic-8a9c37e2e1da6c28.xml (deflated 96%) 2025-12-04T10:11:07.2245031Z adding: test/test-reports/python-pytest/profiler.test_execution_trace/profiler.test_execution_trace-852e4822f8230313.xml (deflated 28%) 2025-12-04T10:11:07.2245803Z adding: test/test-reports/python-pytest/profiler.test_execution_trace/profiler.test_execution_trace-2d1f3ef34837b171.xml (deflated 28%) 2025-12-04T10:11:07.2246452Z adding: test/test-reports/python-pytest/test_jit/test_jit-711c2ccc279f7210.xml (deflated 28%) 2025-12-04T10:11:07.2247077Z adding: test/test-reports/python-pytest/test_jit/test_jit-4dc671ed81b7895e.xml (deflated 99%) 2025-12-04T10:11:07.2247643Z adding: test/test-reports/python-pytest/test_datapipe/test_datapipe-0e1964a956d0a480.xml (deflated 28%) 2025-12-04T10:11:07.2248224Z adding: test/test-reports/python-pytest/test_datapipe/test_datapipe-82b13967e74e19a2.xml (deflated 28%) 2025-12-04T10:11:07.2248862Z adding: test/test-reports/python-pytest/test_numba_integration/test_numba_integration-844fbe0ccaacabee.xml (deflated 28%) 2025-12-04T10:11:07.2249558Z adding: test/test-reports/python-pytest/test_numba_integration/test_numba_integration-6bf5efdb972a9a8f.xml (deflated 28%) 2025-12-04T10:11:07.2250257Z adding: test/test-reports/python-pytest/test_functional_optim/test_functional_optim-182d833e6572f213.xml (deflated 28%) 2025-12-04T10:11:07.2250935Z adding: test/test-reports/python-pytest/test_functional_optim/test_functional_optim-c14b0e455295dcf4.xml (deflated 28%) 2025-12-04T10:11:07.2251588Z adding: test/test-reports/python-pytest/test_maskedtensor/test_maskedtensor-2e832fb0545eaaab.xml (deflated 28%) 2025-12-04T10:11:07.2252252Z adding: test/test-reports/python-pytest/test_maskedtensor/test_maskedtensor-d6bf47e2c29f29e6.xml (deflated 28%) 2025-12-04T10:11:07.2253005Z adding: test/test-reports/python-pytest/benchmark_utils.test_benchmark_utils/benchmark_utils.test_benchmark_utils-af96c8a6347bcdb1.xml (deflated 28%) 2025-12-04T10:11:07.2254001Z adding: test/test-reports/python-pytest/benchmark_utils.test_benchmark_utils/benchmark_utils.test_benchmark_utils-f962c5399d90fb14.xml (deflated 28%) 2025-12-04T10:11:07.2254918Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_scalarmath/torch_np.numpy_tests.core.test_scalarmath-b099d6ab8284837f.xml (deflated 28%) 2025-12-04T10:11:07.2255799Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_scalarmath/torch_np.numpy_tests.core.test_scalarmath-6b29237e24c20b32.xml (deflated 28%) 2025-12-04T10:11:07.2256588Z adding: test/test-reports/python-pytest/test_scaled_matmul_cuda/test_scaled_matmul_cuda-5972d960cfe7ec4e.xml (deflated 28%) 2025-12-04T10:11:07.2257277Z adding: test/test-reports/python-pytest/test_scaled_matmul_cuda/test_scaled_matmul_cuda-6e739739e0574a47.xml (deflated 98%) 2025-12-04T10:11:07.2258042Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_shape_base/torch_np.numpy_tests.core.test_shape_base-ea32bc17692c1ce8.xml (deflated 28%) 2025-12-04T10:11:07.2258917Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_shape_base/torch_np.numpy_tests.core.test_shape_base-36c86d2bfdd67ddd.xml (deflated 28%) 2025-12-04T10:11:07.2259637Z adding: test/test-reports/python-pytest/test_vulkan/test_vulkan-8b9c4d8ede98ee3d.xml (deflated 28%) 2025-12-04T10:11:07.2260193Z adding: test/test-reports/python-pytest/test_vulkan/test_vulkan-e1a521ce9cf70bfb.xml (deflated 28%) 2025-12-04T10:11:07.2260789Z adding: test/test-reports/python-pytest/lazy.test_generator/lazy.test_generator-3b115040434e772f.xml (deflated 28%) 2025-12-04T10:11:07.2261447Z adding: test/test-reports/python-pytest/lazy.test_generator/lazy.test_generator-5262beb36f04167a.xml (deflated 28%) 2025-12-04T10:11:07.2262235Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.linalg.test_linalg/torch_np.numpy_tests.linalg.test_linalg-40abae860675e681.xml (deflated 28%) 2025-12-04T10:11:07.2263131Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.linalg.test_linalg/torch_np.numpy_tests.linalg.test_linalg-fa5a137aca79e3bc.xml (deflated 28%) 2025-12-04T10:11:07.2263992Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_dtype/torch_np.numpy_tests.core.test_dtype-fa2df09a508d13dc.xml (deflated 28%) 2025-12-04T10:11:07.2264812Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_dtype/torch_np.numpy_tests.core.test_dtype-6960413498a3b4a7.xml (deflated 28%) 2025-12-04T10:11:07.2265673Z adding: test/test-reports/python-pytest/lazy.test_debug_util/lazy.test_debug_util-4dae4a778d170bc1.xml (deflated 28%) 2025-12-04T10:11:07.2266330Z adding: test/test-reports/python-pytest/lazy.test_debug_util/lazy.test_debug_util-24c3e47e3d2b7323.xml (deflated 28%) 2025-12-04T10:11:07.2266982Z adding: test/test-reports/python-pytest/nn.test_load_state_dict/nn.test_load_state_dict-91536496375f24b6.xml (deflated 28%) 2025-12-04T10:11:07.2267643Z adding: test/test-reports/python-pytest/nn.test_load_state_dict/nn.test_load_state_dict-b877f1ca494eb282.xml (deflated 28%) 2025-12-04T10:11:07.2268278Z adding: test/test-reports/python-pytest/test_shape_ops/test_shape_ops-012d0fcc882a88fd.xml (deflated 28%) 2025-12-04T10:11:07.2268868Z adding: test/test-reports/python-pytest/test_shape_ops/test_shape_ops-feb6212001ee076f.xml (deflated 28%) 2025-12-04T10:11:07.2269489Z adding: test/test-reports/python-pytest/nn.test_module_hooks/nn.test_module_hooks-46ca7ec97ec9318c.xml (deflated 28%) 2025-12-04T10:11:07.2270125Z adding: test/test-reports/python-pytest/nn.test_module_hooks/nn.test_module_hooks-8ff29e8b05880501.xml (deflated 28%) 2025-12-04T10:11:07.2270906Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.lib.test_twodim_base/torch_np.numpy_tests.lib.test_twodim_base-2accadefabd28fef.xml (deflated 28%) 2025-12-04T10:11:07.2271788Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.lib.test_twodim_base/torch_np.numpy_tests.lib.test_twodim_base-fd2b4839449b1b20.xml (deflated 28%) 2025-12-04T10:11:07.2272675Z adding: test/test-reports/python-pytest/profiler.test_memory_profiler/profiler.test_memory_profiler-d6dc2fc9f3dab1f2.xml (deflated 28%) 2025-12-04T10:11:07.2273511Z adding: test/test-reports/python-pytest/profiler.test_memory_profiler/profiler.test_memory_profiler-89519a5c938e684c.xml (deflated 98%) 2025-12-04T10:11:07.2274217Z adding: test/test-reports/python-pytest/test_jit_llga_fuser/test_jit_llga_fuser-644d80dd48236bd4.xml (deflated 28%) 2025-12-04T10:11:07.2274846Z adding: test/test-reports/python-pytest/test_jit_llga_fuser/test_jit_llga_fuser-d79703873568858a.xml (deflated 28%) 2025-12-04T10:11:07.2275491Z adding: test/test-reports/python-pytest/test_serialization/test_serialization-9d6f2802fe66a8b4.xml (deflated 29%) 2025-12-04T10:11:07.2276138Z adding: test/test-reports/python-pytest/test_serialization/test_serialization-d1a17d69bf98ee36.xml (deflated 28%) 2025-12-04T10:11:07.2276760Z adding: test/test-reports/python-pytest/test_sparse_csr/test_sparse_csr-1e945673d9d0e64f.xml (deflated 28%) 2025-12-04T10:11:07.2277362Z adding: test/test-reports/python-pytest/test_sparse_csr/test_sparse_csr-68a0a0e21687dd8c.xml (deflated 98%) 2025-12-04T10:11:07.2277984Z adding: test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-258fb62f6ed3ad5d.xml (deflated 28%) 2025-12-04T10:11:07.2278618Z adding: test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-a4489a5f35bbfe70.xml (deflated 28%) 2025-12-04T10:11:07.2279258Z adding: test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-63eca64add6314a4.xml (deflated 28%) 2025-12-04T10:11:07.2279900Z adding: test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-4a923aaed07d8de4.xml (deflated 28%) 2025-12-04T10:11:07.2280501Z adding: test/test-reports/python-pytest/test_mkldnn/test_mkldnn-96ab41951b468da7.xml (deflated 28%) 2025-12-04T10:11:07.2281061Z adding: test/test-reports/python-pytest/test_mkldnn/test_mkldnn-3fd399b85d51ccfc.xml (deflated 28%) 2025-12-04T10:11:07.2305507Z ##[group]Run # Remove any previous usage logs if they exist 2025-12-04T10:11:07.2305832Z # Remove any previous usage logs if they exist 2025-12-04T10:11:07.2306098Z rm -f logs-*.zip 2025-12-04T10:11:07.2306351Z zip "logs-${FILE_SUFFIX}.zip" 'usage_log.txt' || true 2025-12-04T10:11:07.2306714Z zip -r "logs-${FILE_SUFFIX}.zip" test/test-reports -i '*.log' || true 2025-12-04T10:11:07.2313942Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:11:07.2314217Z env: 2025-12-04T10:11:07.2314379Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:07.2314686Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:07.2314928Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:07.2315324Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:07.2315674Z DEVICE_NAME: 2025-12-04T10:11:07.2315831Z DEVICE_TYPE: 2025-12-04T10:11:07.2316136Z FILE_SUFFIX: test-default-4-5-lf.linux.g6.4xlarge.experimental.nvidia.gpu_57116084869 2025-12-04T10:11:07.2316496Z ##[endgroup] 2025-12-04T10:11:07.2366634Z adding: usage_log.txt (deflated 58%) 2025-12-04T10:11:07.2429024Z adding: test/test-reports/test_hop_infra_1.1_3ccb8c957d7bb463_.log (deflated 48%) 2025-12-04T10:11:07.2429691Z adding: test/test-reports/inductor.test_aot_inductor_3.4_4414340a188212de_.log (deflated 49%) 2025-12-04T10:11:07.2430337Z adding: test/test-reports/test_mkldnn_verbose_1.1_c89e68a75cbfee9d_.log (deflated 49%) 2025-12-04T10:11:07.2430969Z adding: test/test-reports/dynamo.test_model_output_1.1_4617a3ec054d38e3_.log (deflated 50%) 2025-12-04T10:11:07.2432230Z adding: test/test-reports/test_decomp_18.22_6d2264971e288acd_.log (deflated 91%) 2025-12-04T10:11:07.2432858Z adding: test/test-reports/inductor.test_extension_backend_1.1_31c714ecf5a69f35_.log (deflated 49%) 2025-12-04T10:11:07.2434171Z adding: test/test-reports/inductor.test_torchinductor_opinfo_7.13_ec766f1d46ff24f6_.log (deflated 95%) 2025-12-04T10:11:07.2435031Z adding: test/test-reports/inductor.test_cooperative_reductions_1.1_2b00296025758482_.log (deflated 54%) 2025-12-04T10:11:07.2435669Z adding: test/test-reports/test_content_store_1.1_8b370b99c7663fbf_.log (deflated 49%) 2025-12-04T10:11:07.2436364Z adding: test/test-reports/dynamo.test_fx_graph_runnable_1.1_d62e5b0a2f03d2b8_.log (deflated 51%) 2025-12-04T10:11:07.2436973Z adding: test/test-reports/test_autoload_1.1_bf261471dad7ddf1_.log (deflated 48%) 2025-12-04T10:11:07.2437545Z adding: test/test-reports/dynamo.test_backends_1.1_dee2ca4cb0571aff_.log (deflated 49%) 2025-12-04T10:11:07.2438146Z adding: test/test-reports/test_per_overload_api_1.1_d71ebb1277eaeb76_.log (deflated 49%) 2025-12-04T10:11:07.2438727Z adding: test/test-reports/dynamo.test_functions_1.1_01cba0571253972c_.log (deflated 50%) 2025-12-04T10:11:07.2439337Z adding: test/test-reports/inductor.test_layout_optim_1.1_05e4855def927acc_.log (deflated 50%) 2025-12-04T10:11:07.2440002Z adding: test/test-reports/inductor.test_mix_order_reduction_1.1_b3c0a6fce4a8dfd0_.log (deflated 51%) 2025-12-04T10:11:07.2440630Z adding: test/test-reports/test_license_1.1_b93bab9da6159df7_.log (deflated 48%) 2025-12-04T10:11:07.2441201Z adding: test/test-reports/inductor.test_padding_1.1_df8aa049a65f2ea4_.log (deflated 50%) 2025-12-04T10:11:07.2441782Z adding: test/test-reports/test_serialization_1.1_30115278e988556a_.log (deflated 63%) 2025-12-04T10:11:07.2442346Z adding: test/test-reports/dynamo.test_dicts_1.1_dabfcd9b1024dd1a_.log (deflated 49%) 2025-12-04T10:11:07.2442812Z adding: test/test-reports/test_ao_sparsity_1.1_baacf12029e08a3c_.log (deflated 48%) 2025-12-04T10:11:07.2443274Z adding: test/test-reports/dynamo.test_aot_compile_1.1_52a2ed661618cde4_.log (deflated 51%) 2025-12-04T10:11:07.2443744Z adding: test/test-reports/test_sparse_csr_1.3_999a94d874ab59b6_.log (deflated 95%) 2025-12-04T10:11:07.2444185Z adding: test/test-reports/dynamo.test_sets_1.1_0a6d988f4dd8b837_.log (deflated 49%) 2025-12-04T10:11:07.2444619Z adding: test/test-reports/xpu.test_gemm_1.1_5c80f370f4543dcf_.log (deflated 48%) 2025-12-04T10:11:07.2445070Z adding: test/test-reports/dynamo.test_callback_1.1_3b05c69589f6a657_.log (deflated 49%) 2025-12-04T10:11:07.2445554Z adding: test/test-reports/dynamo.test_backends_1.1_732b26763f9c8444_.log (deflated 49%) 2025-12-04T10:11:07.2446111Z adding: test/test-reports/inductor.test_cudagraph_trees_expandable_segments_1.1_261e345fe4595b61_.log (deflated 54%) 2025-12-04T10:11:07.2446646Z adding: test/test-reports/test_as_strided_1.1_5667cb6d1805af19_.log (deflated 48%) 2025-12-04T10:11:07.2447219Z adding: test/test-reports/dynamo.test_exceptions_1.1_27ab688a8f5f578d_.log (deflated 50%) 2025-12-04T10:11:07.2462311Z adding: test/test-reports/test_cuda_1.1_c907a69791bbc315_.log (deflated 99%) 2025-12-04T10:11:07.2462884Z adding: test/test-reports/functorch.test_ops_6.9_7aa027fe1b42b488_.log (deflated 49%) 2025-12-04T10:11:07.2463455Z adding: test/test-reports/test_transformers_1.1_553a713ded2a82f8_.log (deflated 49%) 2025-12-04T10:11:07.2464198Z adding: test/test-reports/test_custom_ops_1.1_80f9a34920161ad5_.log (deflated 94%) 2025-12-04T10:11:07.2464802Z adding: test/test-reports/higher_order_ops.test_local_map_1.1_251d3402b699ff52_.log (deflated 51%) 2025-12-04T10:11:07.2465403Z adding: test/test-reports/test_mkldnn_1.1_cf74075bbf0a4f72_.log (deflated 48%) 2025-12-04T10:11:07.2465932Z adding: test/test-reports/test_dataloader_1.1_8a6dced23b850537_.log (deflated 73%) 2025-12-04T10:11:07.2466468Z adding: test/test-reports/test_decomp_6.22_772db1786e1c2ca3_.log (deflated 48%) 2025-12-04T10:11:07.2466993Z adding: test/test-reports/test_decomp_8.22_2bb5d34331689c22_.log (deflated 48%) 2025-12-04T10:11:07.2467515Z adding: test/test-reports/test_decomp_12.22_09d4a096062ecff8_.log (deflated 48%) 2025-12-04T10:11:07.2468035Z adding: test/test-reports/test_decomp_18.22_b7c40422afc58245_.log (deflated 48%) 2025-12-04T10:11:07.2468553Z adding: test/test-reports/test_decomp_20.22_2b2eda3179b882c5_.log (deflated 48%) 2025-12-04T10:11:07.2469143Z adding: test/test-reports/test_ops_2.9_5593faa30198642f_.log (deflated 48%) 2025-12-04T10:11:07.2469636Z adding: test/test-reports/test_ops_8.9_4a7c51b6bf787854_.log (deflated 48%) 2025-12-04T10:11:07.2470227Z adding: test/test-reports/functorch.test_dims_1.1_b36a72e7d9c2bfba_.log (deflated 50%) 2025-12-04T10:11:07.2474161Z adding: test/test-reports/test_foreach_1.1_216741b645f3a84c_.log (deflated 98%) 2025-12-04T10:11:07.2474623Z adding: test/test-reports/test_torchfuzz_repros_1.1_9f3f619c966168b6_.log (deflated 50%) 2025-12-04T10:11:07.2475115Z adding: test/test-reports/lazy.test_step_closures_1.1_9ba2d3a408de6eb8_.log (deflated 50%) 2025-12-04T10:11:07.2475618Z adding: test/test-reports/inductor.test_torchinductor_1.2_0c18f3f955f52c94_.log (deflated 49%) 2025-12-04T10:11:07.2476106Z adding: test/test-reports/test_nestedtensor_4.4_f06c8dae8e05ce7f_.log (deflated 49%) 2025-12-04T10:11:07.2476599Z adding: test/test-reports/inductor.test_torchinductor_2.2_88a9d2ce75af7d9b_.log (deflated 49%) 2025-12-04T10:11:07.2477117Z adding: test/test-reports/inductor.test_flex_decoding_1.3_d0db0cd9a9640154_.log (deflated 51%) 2025-12-04T10:11:07.2477637Z adding: test/test-reports/inductor.test_kernel_benchmark_1.1_3e5069f4126a6335_.log (deflated 49%) 2025-12-04T10:11:07.2478165Z adding: test/test-reports/inductor.test_flex_decoding_3.3_cda88ebe36ff4298_.log (deflated 50%) 2025-12-04T10:11:07.2478710Z adding: test/test-reports/inductor.test_torchinductor_opinfo_1.13_8df0b8edc083f301_.log (deflated 50%) 2025-12-04T10:11:07.2479253Z adding: test/test-reports/inductor.test_deterministic_3.3_49cc6b9d77892b3f_.log (deflated 48%) 2025-12-04T10:11:07.2479788Z adding: test/test-reports/inductor.test_torchinductor_opinfo_5.13_31f48df8ddc8d1c5_.log (deflated 50%) 2025-12-04T10:11:07.2480308Z adding: test/test-reports/inductor.test_b2b_gemm_1.1_e64cb8635d438c4d_.log (deflated 50%) 2025-12-04T10:11:07.2480827Z adding: test/test-reports/inductor.test_torchinductor_opinfo_7.13_ce10eb09f3c1ae50_.log (deflated 50%) 2025-12-04T10:11:07.2481390Z adding: test/test-reports/inductor.test_mix_order_reduction_1.1_737090c6800dd0c8_.log (deflated 95%) 2025-12-04T10:11:07.2481945Z adding: test/test-reports/inductor.test_torchinductor_opinfo_13.13_23e880810105a276_.log (deflated 50%) 2025-12-04T10:11:07.2482450Z adding: test/test-reports/test_out_dtype_op_1.1_0c60bd107333eda1_.log (deflated 49%) 2025-12-04T10:11:07.2482925Z adding: test/test-reports/inductor.test_layout_optim_1.1_01519e00b3bd1c88_.log (deflated 50%) 2025-12-04T10:11:07.2483534Z adding: test/test-reports/inductor.test_aot_inductor_utils_1.1_e3aa7bd30346a26c_.log (deflated 51%) 2025-12-04T10:11:07.2484106Z adding: test/test-reports/inductor.test_aot_inductor_arrayref_2.2_8efae0aa65170a95_.log (deflated 50%) 2025-12-04T10:11:07.2484598Z adding: test/test-reports/test_decomp_12.22_5f45f685300755c6_.log (deflated 48%) 2025-12-04T10:11:07.2485118Z adding: test/test-reports/inductor.test_torchinductor_strided_blocks_1.1_79b6d86fa53da74b_.log (deflated 51%) 2025-12-04T10:11:07.2485642Z adding: test/test-reports/test_custom_ops_1.1_57e8b64e81978c4f_.log (deflated 49%) 2025-12-04T10:11:07.2486102Z adding: test/test-reports/test_content_store_1.1_abe40a032272748f_.log (deflated 49%) 2025-12-04T10:11:07.2486589Z adding: test/test-reports/torch_np.test_ufuncs_basic_1.1_cd7199ad4e0a518a_.log (deflated 50%) 2025-12-04T10:11:07.2487104Z adding: test/test-reports/inductor.test_flex_decoding_1.3_30d64424b74ff5a2_.log (deflated 50%) 2025-12-04T10:11:07.2487606Z adding: test/test-reports/functorch.dim.test_split_1.1_14b064ffd62da833_.log (deflated 50%) 2025-12-04T10:11:07.2488115Z adding: test/test-reports/inductor.test_flex_decoding_3.3_27519a12c6ab0e36_.log (deflated 50%) 2025-12-04T10:11:07.2488611Z adding: test/test-reports/test_ops_fwd_gradients_1.3_86a0d9916377e0c1_.log (deflated 95%) 2025-12-04T10:11:07.2489103Z adding: test/test-reports/inductor.test_deterministic_3.3_344aea65313e5096_.log (deflated 49%) 2025-12-04T10:11:07.2490003Z adding: test/test-reports/test_ops_fwd_gradients_2.3_24e5b94522df8eef_.log (deflated 95%) 2025-12-04T10:11:07.2490485Z adding: test/test-reports/inductor.test_b2b_gemm_1.1_fd238e41e950af5f_.log (deflated 50%) 2025-12-04T10:11:07.2490997Z adding: test/test-reports/test_nestedtensor_3.4_b0204203c06b9fb4_.log (deflated 49%) 2025-12-04T10:11:07.2491466Z adding: test/test-reports/export.test_tree_utils_1.1_763b13f8fc426b15_.log (deflated 50%) 2025-12-04T10:11:07.2491974Z adding: test/test-reports/inductor.test_compiled_optimizers_3.3_a39d9905fe0a7e96_.log (deflated 49%) 2025-12-04T10:11:07.2492524Z adding: test/test-reports/inductor.test_triton_wrapper_1.1_1526c6966d209d73_.log (deflated 51%) 2025-12-04T10:11:07.2493039Z adding: test/test-reports/inductor.test_triton_wrapper_1.1_d176b02f5f219154_.log (deflated 50%) 2025-12-04T10:11:07.2493662Z adding: test/test-reports/inductor.test_static_cuda_launcher_1.1_c765f37b7d99c783_.log (deflated 51%) 2025-12-04T10:11:07.2494244Z adding: test/test-reports/test_modules_2.2_5235b4f398c2f08b_.log (deflated 95%) 2025-12-04T10:11:07.2494711Z adding: test/test-reports/export.test_dynamic_shapes_1.1_09e3ef98478a54cd_.log (deflated 50%) 2025-12-04T10:11:07.2495193Z adding: test/test-reports/dynamo.test_sdpa_1.1_6e4183b888ceacf8_.log (deflated 49%) 2025-12-04T10:11:07.2495692Z adding: test/test-reports/inductor.test_static_cuda_launcher_1.1_9b61891d2c79064c_.log (deflated 51%) 2025-12-04T10:11:07.2496234Z adding: test/test-reports/inductor.test_aot_inductor_package_1.1_2e32c0b0033ee866_.log (deflated 52%) 2025-12-04T10:11:07.2496765Z adding: test/test-reports/export.test_dynamic_shapes_1.1_0acf0e3fcb724802_.log (deflated 50%) 2025-12-04T10:11:07.2497308Z adding: test/test-reports/inductor.test_compiled_optimizers_3.3_e91eedd79d0b2aa7_.log (deflated 49%) 2025-12-04T10:11:07.2497807Z adding: test/test-reports/dynamo.test_sdpa_1.1_a67282d1f1c8872a_.log (deflated 48%) 2025-12-04T10:11:07.2498297Z adding: test/test-reports/inductor.test_aot_inductor_utils_1.1_3c60e8e2309ff072_.log (deflated 51%) 2025-12-04T10:11:07.2498788Z adding: test/test-reports/test_tensorboard_1.1_9b09def11169383b_.log (deflated 49%) 2025-12-04T10:11:07.2499263Z adding: test/test-reports/inductor.test_control_flow_3.5_33c5414d09e1fd98_.log (deflated 50%) 2025-12-04T10:11:07.2499736Z adding: test/test-reports/test_mkl_verbose_1.1_4ffe3bbc2c20413d_.log (deflated 49%) 2025-12-04T10:11:07.2500161Z adding: test/test-reports/test_futures_1.1_f17fce10eca39347_.log (deflated 48%) 2025-12-04T10:11:07.2500722Z adding: test/test-reports/test_comparison_utils_1.1_03659744aead51ab_.log (deflated 50%) 2025-12-04T10:11:07.2501186Z adding: test/test-reports/nn.test_dropout_1.1_5b2ab44291c4879d_.log (deflated 48%) 2025-12-04T10:11:07.2501660Z adding: test/test-reports/functorch.test_ac_logging_1.1_fe60cab7003c0ff7_.log (deflated 50%) 2025-12-04T10:11:07.2502137Z adding: test/test-reports/test_mkldnn_verbose_1.1_24beac6fb8e8b85c_.log (deflated 49%) 2025-12-04T10:11:07.2506844Z adding: test/test-reports/test_jit_1.1_335f426c034a81ab_.log (deflated 98%) 2025-12-04T10:11:07.2507667Z adding: test/test-reports/test_utils_config_module_1.1_2fb5daf1593c0cf5_.log (deflated 50%) 2025-12-04T10:11:07.2508513Z adding: test/test-reports/test_hop_infra_1.1_8ee2aa3c7e32f90d_.log (deflated 49%) 2025-12-04T10:11:07.2509394Z adding: test/test-reports/inductor.test_aot_inductor_package_1.1_e6c86135fe5c29f7_.log (deflated 51%) 2025-12-04T10:11:07.2510361Z adding: test/test-reports/test_appending_byte_serializer_1.1_64533d4e3bf3cda4_.log (deflated 51%) 2025-12-04T10:11:07.2511212Z adding: test/test-reports/test_license_1.1_702ac49c8c00b783_.log (deflated 49%) 2025-12-04T10:11:07.2511979Z adding: test/test-reports/test_ao_sparsity_1.1_fc9d1e285ec3ec39_.log (deflated 49%) 2025-12-04T10:11:07.2512741Z adding: test/test-reports/test_autoload_1.1_b3134511ed88d0ac_.log (deflated 49%) 2025-12-04T10:11:07.2513570Z adding: test/test-reports/inductor.test_control_flow_3.5_00fe48f3237e41c2_.log (deflated 50%) 2025-12-04T10:11:07.2514591Z adding: test/test-reports/nn.attention.test_open_registry_1.1_da6b3f4cb0dc7340_.log (deflated 51%) 2025-12-04T10:11:07.2515509Z adding: test/test-reports/test_as_strided_1.1_96a2fa8a410b6049_.log (deflated 49%) 2025-12-04T10:11:07.2516254Z adding: test/test-reports/test_foreach_1.1_4b853a43bd6e725e_.log (deflated 48%) 2025-12-04T10:11:07.2516995Z adding: test/test-reports/xpu.test_gemm_1.1_709617076e2e7b44_.log (deflated 48%) 2025-12-04T10:11:07.2517718Z adding: test/test-reports/test_meta_4.4_2bd0d20115e5ff88_.log (deflated 49%) 2025-12-04T10:11:07.2518526Z adding: test/test-reports/higher_order_ops.test_print_1.1_afd6752b8d2cc0f6_.log (deflated 51%) 2025-12-04T10:11:07.2519403Z adding: test/test-reports/test_scaled_matmul_cuda_1.1_d6e96ad4cd7f3119_.log (deflated 94%) 2025-12-04T10:11:07.2520245Z adding: test/test-reports/test_per_overload_api_1.1_d91a23c6a925aea0_.log (deflated 50%) 2025-12-04T10:11:07.2521060Z adding: test/test-reports/dynamo.test_dicts_1.1_de613a189e3e1504_.log (deflated 49%) 2025-12-04T10:11:07.2521967Z adding: test/test-reports/torch_np.numpy_tests.core.test_einsum_1.1_720e264a46a74885_.log (deflated 52%) 2025-12-04T10:11:07.2522859Z adding: test/test-reports/test_out_dtype_op_1.1_f2a6f18e42db3cfb_.log (deflated 49%) 2025-12-04T10:11:07.2523620Z adding: test/test-reports/test_vulkan_1.1_4680f5183d531af0_.log (deflated 48%) 2025-12-04T10:11:07.2524430Z adding: test/test-reports/torch_np.test_ufuncs_basic_1.1_bfaee78e6ce1e6e9_.log (deflated 50%) 2025-12-04T10:11:07.2525247Z adding: test/test-reports/test_datapipe_1.1_8084ffe5984ad57f_.log (deflated 48%) 2025-12-04T10:11:07.2526054Z adding: test/test-reports/lazy.test_step_closures_1.1_868a341d9b1e0a2d_.log (deflated 50%) 2025-12-04T10:11:07.2526867Z adding: test/test-reports/test_mkl_verbose_1.1_3691733155bb2907_.log (deflated 48%) 2025-12-04T10:11:07.2527717Z adding: test/test-reports/functorch.dim.test_getsetitem_1.1_38009191dd39b63e_.log (deflated 51%) 2025-12-04T10:11:07.2528601Z adding: test/test-reports/test_numba_integration_1.1_22b9b4055cfb8752_.log (deflated 49%) 2025-12-04T10:11:07.2529437Z adding: test/test-reports/test_ops_fwd_gradients_1.3_f370258098a4a3d8_.log (deflated 49%) 2025-12-04T10:11:07.2530278Z adding: test/test-reports/test_functional_optim_1.1_13d5a8983a5027f5_.log (deflated 49%) 2025-12-04T10:11:07.2531090Z adding: test/test-reports/test_ops_fwd_gradients_2.3_222494559a713d2f_.log (deflated 50%) 2025-12-04T10:11:07.2531561Z adding: test/test-reports/test_meta_4.4_42623e8fef4d35a7_.log (deflated 49%) 2025-12-04T10:11:07.2532151Z adding: test/test-reports/test_nestedtensor_3.4_7d556cc06aefdeb1_.log (deflated 49%) 2025-12-04T10:11:07.2532645Z adding: test/test-reports/test_nestedtensor_4.4_1ca35917d202dd6d_.log (deflated 49%) 2025-12-04T10:11:07.2533088Z adding: test/test-reports/test_modules_2.2_8119bdbc5009a9d0_.log (deflated 48%) 2025-12-04T10:11:07.2533631Z adding: test/test-reports/test_tensorboard_1.1_cb20da95004ddae1_.log (deflated 49%) 2025-12-04T10:11:07.2534105Z adding: test/test-reports/dynamo.test_aot_compile_1.1_d4d7319297f0b64f_.log (deflated 51%) 2025-12-04T10:11:07.2534636Z adding: test/test-reports/torch_np.numpy_tests.core.test_indexing_1.1_ae4b7321e401d129_.log (deflated 52%) 2025-12-04T10:11:07.2535132Z adding: test/test-reports/test_futures_1.1_ab361dc5625ed4a3_.log (deflated 49%) 2025-12-04T10:11:07.2535560Z adding: test/test-reports/nn.test_dropout_1.1_397a2054f8bff0d6_.log (deflated 49%) 2025-12-04T10:11:07.2536023Z adding: test/test-reports/test_maskedtensor_1.1_86a8621c8ea58e71_.log (deflated 49%) 2025-12-04T10:11:07.2536502Z adding: test/test-reports/functorch.dim.test_split_1.1_822ccb8ad215b24b_.log (deflated 50%) 2025-12-04T10:11:07.2536967Z adding: test/test-reports/dynamo.test_sets_1.1_11315a33e392b1af_.log (deflated 48%) 2025-12-04T10:11:07.2537482Z adding: test/test-reports/torch_np.numpy_tests.lib.test_type_check_1.1_ccc3d1df2287267e_.log (deflated 52%) 2025-12-04T10:11:07.2538034Z adding: test/test-reports/test_decomp_20.22_f3a68dcf10860f63_.log (deflated 48%) 2025-12-04T10:11:07.2538531Z adding: test/test-reports/cpp_extensions.test_libtorch_agnostic_1.1_db47f5251d8f5d77_.log (deflated 92%) 2025-12-04T10:11:07.2539107Z adding: test/test-reports/functorch.test_ac_logging_1.1_6ca2b350550f1b46_.log (deflated 50%) 2025-12-04T10:11:07.2539772Z adding: test/test-reports/profiler.test_execution_trace_1.1_6be3fb6173e553da_.log (deflated 51%) 2025-12-04T10:11:07.2540256Z adding: test/test-reports/test_jit_1.1_235772887b92014d_.log (deflated 48%) 2025-12-04T10:11:07.2540687Z adding: test/test-reports/test_datapipe_1.1_eeeeff3cb1cf07c0_.log (deflated 49%) 2025-12-04T10:11:07.2541131Z adding: test/test-reports/lazy.test_generator_1.1_647ef07d1063ea00_.log (deflated 49%) 2025-12-04T10:11:07.2541603Z adding: test/test-reports/test_numba_integration_1.1_4717bde06fdb8313_.log (deflated 50%) 2025-12-04T10:11:07.2542082Z adding: test/test-reports/functorch.test_ops_8.9_bb4596b740c061de_.log (deflated 49%) 2025-12-04T10:11:07.2542545Z adding: test/test-reports/test_functional_optim_1.1_f4eff2a83f847c20_.log (deflated 50%) 2025-12-04T10:11:07.2543011Z adding: test/test-reports/test_maskedtensor_1.1_b12b8ab6e2307bad_.log (deflated 49%) 2025-12-04T10:11:07.2543439Z adding: test/test-reports/test_ops_2.9_146c8473bb4b76bc_.log (deflated 47%) 2025-12-04T10:11:07.2543923Z adding: test/test-reports/benchmark_utils.test_benchmark_utils_1.1_772184f487491c3a_.log (deflated 52%) 2025-12-04T10:11:07.2544410Z adding: test/test-reports/test_ops_8.9_047a11a4b058dbbd_.log (deflated 94%) 2025-12-04T10:11:07.2544905Z adding: test/test-reports/torch_np.numpy_tests.core.test_scalarmath_1.1_6cf1169adf67ed5b_.log (deflated 52%) 2025-12-04T10:11:07.2545431Z adding: test/test-reports/nn.test_module_hooks_1.1_af6a85c4218349c4_.log (deflated 49%) 2025-12-04T10:11:07.2545906Z adding: test/test-reports/test_scaled_matmul_cuda_1.1_e3dc30c781dc2515_.log (deflated 50%) 2025-12-04T10:11:07.2553765Z adding: test/test-reports/functorch.test_dims_1.1_99eaaa239e58b576_.log (deflated 98%) 2025-12-04T10:11:07.2554440Z adding: test/test-reports/torch_np.numpy_tests.core.test_shape_base_1.1_fdd19f4d65e5ee90_.log (deflated 52%) 2025-12-04T10:11:07.2555076Z adding: test/test-reports/test_vulkan_1.1_610e90772bff2bbd_.log (deflated 48%) 2025-12-04T10:11:07.2555619Z adding: test/test-reports/lazy.test_generator_1.1_08428f43a23dddc2_.log (deflated 50%) 2025-12-04T10:11:07.2556188Z adding: test/test-reports/test_torchfuzz_repros_1.1_e32a95fbe8787815_.log (deflated 49%) 2025-12-04T10:11:07.2556955Z adding: test/test-reports/torch_np.numpy_tests.linalg.test_linalg_1.1_2ae6c059e7041367_.log (deflated 52%) 2025-12-04T10:11:07.2557653Z adding: test/test-reports/inductor.test_torchinductor_1.2_140eeae6a9700c9e_.log (deflated 95%) 2025-12-04T10:11:07.2558321Z adding: test/test-reports/torch_np.numpy_tests.core.test_dtype_1.1_ce81285daa6e127c_.log (deflated 52%) 2025-12-04T10:11:07.2558957Z adding: test/test-reports/lazy.test_debug_util_1.1_5381255485755d96_.log (deflated 49%) 2025-12-04T10:11:07.2559537Z adding: test/test-reports/lazy.test_debug_util_1.1_ce59c472ce92b8ab_.log (deflated 50%) 2025-12-04T10:11:07.2560118Z adding: test/test-reports/nn.test_load_state_dict_1.1_5a4907470fe975cf_.log (deflated 50%) 2025-12-04T10:11:07.2560697Z adding: test/test-reports/nn.test_load_state_dict_1.1_74b8c756583770f2_.log (deflated 50%) 2025-12-04T10:11:07.2561246Z adding: test/test-reports/test_shape_ops_1.1_55833b361d0f983c_.log (deflated 49%) 2025-12-04T10:11:07.2561784Z adding: test/test-reports/test_shape_ops_1.1_aa3f7cc31cd86f20_.log (deflated 48%) 2025-12-04T10:11:07.2562307Z adding: test/test-reports/nn.test_module_hooks_1.1_69958f2e5e5f5e6f_.log (deflated 50%) 2025-12-04T10:11:07.2564152Z adding: test/test-reports/inductor.test_torchinductor_2.2_d10e54041d6e8139_.log (deflated 96%) 2025-12-04T10:11:07.2564717Z adding: test/test-reports/torch_np.numpy_tests.lib.test_twodim_base_1.1_af867f8373093dbc_.log (deflated 52%) 2025-12-04T10:11:07.2565317Z adding: test/test-reports/test_utils_config_module_1.1_6e4b2011ed24ec12_.log (deflated 50%) 2025-12-04T10:11:07.2566006Z adding: test/test-reports/profiler.test_memory_profiler_1.1_ea25b5650767ab1b_.log (deflated 51%) 2025-12-04T10:11:07.2566517Z adding: test/test-reports/test_jit_llga_fuser_1.1_7b4085049eb12443_.log (deflated 49%) 2025-12-04T10:11:07.2566959Z adding: test/test-reports/test_serialization_1.1_9875b20eb7f175eb_.log (deflated 62%) 2025-12-04T10:11:07.2567414Z adding: test/test-reports/test_sparse_csr_1.3_feec296a088cb196_.log (deflated 49%) 2025-12-04T10:11:07.2567867Z adding: test/test-reports/functorch.test_ops_6.9_a93c0a53b4ef93ed_.log (deflated 49%) 2025-12-04T10:11:07.2568331Z adding: test/test-reports/functorch.test_ops_8.9_8bf2c8a6adbd0bd1_.log (deflated 49%) 2025-12-04T10:11:07.2568764Z adding: test/test-reports/test_mkldnn_1.1_e0499fc9c340a32c_.log (deflated 48%) 2025-12-04T10:11:07.2569265Z adding: test/test-reports/inductor.test_cooperative_reductions_1.1_07ed6b958ddfd437_.log (deflated 54%) 2025-12-04T10:11:07.2569794Z adding: test/test-reports/inductor.test_padding_1.1_dd27bd85980e083b_.log (deflated 50%) 2025-12-04T10:11:07.2570412Z adding: test/test-reports/dynamo.test_model_output_1.1_4f24511bc5d419f7_.log (deflated 94%) 2025-12-04T10:11:07.2602731Z adding: test/test-reports/inductor.test_kernel_benchmark_1.1_4dd3a5ee490d965b_.log (deflated 99%) 2025-12-04T10:11:07.2603458Z adding: test/test-reports/inductor.test_aot_inductor_3.4_67cbdcc1e4ecfe70_.log (deflated 48%) 2025-12-04T10:11:07.2609862Z adding: test/test-reports/inductor.test_extension_backend_1.1_c635b12d7c8d4f3e_.log (deflated 97%) 2025-12-04T10:11:07.2610533Z adding: test/test-reports/dynamo.test_functions_1.1_e9fec77e4c6ac578_.log (deflated 49%) 2025-12-04T10:11:07.2611110Z adding: test/test-reports/test_decomp_8.22_fb867f77ecfd23a1_.log (deflated 48%) 2025-12-04T10:11:07.2611680Z adding: test/test-reports/dynamo.test_callback_1.1_46f26d924e4a214f_.log (deflated 49%) 2025-12-04T10:11:07.2612277Z adding: test/test-reports/dynamo.test_exceptions_1.1_c4596b16d2ddbb72_.log (deflated 50%) 2025-12-04T10:11:07.2612991Z adding: test/test-reports/inductor.test_cudagraph_trees_expandable_segments_1.1_944dfa8853de3b32_.log (deflated 93%) 2025-12-04T10:11:07.2619402Z adding: test/test-reports/test_cuda_1.1_1d7d8eb7c1053f49_.log (deflated 97%) 2025-12-04T10:11:07.2619968Z adding: test/test-reports/test_comparison_utils_1.1_6dc4c17d20e9f5c0_.log (deflated 49%) 2025-12-04T10:11:07.2620791Z adding: test/test-reports/higher_order_ops.test_local_map_1.1_cb27141dc638196e_.log (deflated 51%) 2025-12-04T10:11:07.2621457Z adding: test/test-reports/test_dataloader_1.1_a3d69a47396d20d9_.log (deflated 73%) 2025-12-04T10:11:07.2622049Z adding: test/test-reports/test_transformers_1.1_6b85ae180d2e1277_.log (deflated 49%) 2025-12-04T10:11:07.2622637Z adding: test/test-reports/export.test_tree_utils_1.1_a150d77e4196bf39_.log (deflated 50%) 2025-12-04T10:11:07.2623106Z adding: test/test-reports/test_decomp_6.22_9c63b20df2fb93f2_.log (deflated 48%) 2025-12-04T10:11:07.2624316Z adding: test/test-reports/inductor.test_torchinductor_opinfo_1.13_5617d11ad1b4a636_.log (deflated 97%) 2025-12-04T10:11:07.2625459Z adding: test/test-reports/inductor.test_torchinductor_opinfo_5.13_ab6bb32b2389ca26_.log (deflated 95%) 2025-12-04T10:11:07.2626028Z adding: test/test-reports/inductor.test_torchinductor_opinfo_13.13_1a71b2cb0cb0ac84_.log (deflated 49%) 2025-12-04T10:11:07.2626601Z adding: test/test-reports/inductor.test_aot_inductor_arrayref_2.2_ddb845b598b6b717_.log (deflated 49%) 2025-12-04T10:11:07.2630205Z adding: test/test-reports/inductor.test_torchinductor_strided_blocks_1.1_e98c1bd43c2c27cb_.log (deflated 98%) 2025-12-04T10:11:07.2630819Z adding: test/test-reports/test_appending_byte_serializer_1.1_38e7f5cc083c9bad_.log (deflated 50%) 2025-12-04T10:11:07.2631363Z adding: test/test-reports/nn.attention.test_open_registry_1.1_c0a1d8e6882b9f0d_.log (deflated 51%) 2025-12-04T10:11:07.2631984Z adding: test/test-reports/higher_order_ops.test_print_1.1_56d44039c67711d6_.log (deflated 50%) 2025-12-04T10:11:07.2632546Z adding: test/test-reports/torch_np.numpy_tests.core.test_einsum_1.1_55ea97b8b3e11159_.log (deflated 52%) 2025-12-04T10:11:07.2633178Z adding: test/test-reports/functorch.dim.test_getsetitem_1.1_189c9e5b965e450c_.log (deflated 50%) 2025-12-04T10:11:07.2633730Z adding: test/test-reports/torch_np.numpy_tests.core.test_indexing_1.1_dec9824650356edd_.log (deflated 52%) 2025-12-04T10:11:07.2634320Z adding: test/test-reports/torch_np.numpy_tests.lib.test_type_check_1.1_772dbb7849e86426_.log (deflated 52%) 2025-12-04T10:11:07.2634885Z adding: test/test-reports/profiler.test_execution_trace_1.1_9cc4d20396616452_.log (deflated 51%) 2025-12-04T10:11:07.2638753Z adding: test/test-reports/cpp_extensions.test_libtorch_agnostic_1.1_75653c2be51b730e_.log (deflated 93%) 2025-12-04T10:11:07.2639355Z adding: test/test-reports/benchmark_utils.test_benchmark_utils_1.1_6343f18ab9650ba6_.log (deflated 52%) 2025-12-04T10:11:07.2639936Z adding: test/test-reports/torch_np.numpy_tests.core.test_scalarmath_1.1_de9f81d561ae89ba_.log (deflated 52%) 2025-12-04T10:11:07.2640525Z adding: test/test-reports/torch_np.numpy_tests.core.test_shape_base_1.1_6b4f426779600108_.log (deflated 52%) 2025-12-04T10:11:07.2641108Z adding: test/test-reports/torch_np.numpy_tests.linalg.test_linalg_1.1_5ada73b70ce3b922_.log (deflated 52%) 2025-12-04T10:11:07.2641673Z adding: test/test-reports/torch_np.numpy_tests.core.test_dtype_1.1_c041cb87250c2eda_.log (deflated 51%) 2025-12-04T10:11:07.2642236Z adding: test/test-reports/torch_np.numpy_tests.lib.test_twodim_base_1.1_96deb5da66224442_.log (deflated 52%) 2025-12-04T10:11:07.2642801Z adding: test/test-reports/profiler.test_memory_profiler_1.1_d170501bd0fd49c4_.log (deflated 96%) 2025-12-04T10:11:07.2643299Z adding: test/test-reports/test_jit_llga_fuser_1.1_490d6d2a811ac6f0_.log (deflated 49%) 2025-12-04T10:11:07.2647306Z adding: test/test-reports/dynamo.test_fx_graph_runnable_1.1_3eb668d8af36329e_.log (deflated 96%) 2025-12-04T10:11:07.2691801Z ##[group]Run # Remove any previous debugging artifacts if they exist 2025-12-04T10:11:07.2692203Z # Remove any previous debugging artifacts if they exist 2025-12-04T10:11:07.2692485Z rm -f debug-*.zip 2025-12-04T10:11:07.2692691Z if [ -d 'test/debug' ]; then 2025-12-04T10:11:07.2692942Z  zip -r "debug-${FILE_SUFFIX}.zip" test/debug 2025-12-04T10:11:07.2693279Z fi 2025-12-04T10:11:07.2699948Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:11:07.2700229Z env: 2025-12-04T10:11:07.2700386Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:07.2700578Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:07.2700802Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:07.2701182Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:07.2701542Z DEVICE_NAME: 2025-12-04T10:11:07.2701707Z DEVICE_TYPE: 2025-12-04T10:11:07.2702011Z FILE_SUFFIX: test-default-4-5-lf.linux.g6.4xlarge.experimental.nvidia.gpu_57116084869 2025-12-04T10:11:07.2702380Z ##[endgroup] 2025-12-04T10:11:07.2790680Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T10:11:07.2790921Z with: 2025-12-04T10:11:07.2791075Z s3-bucket: gha-artifacts 2025-12-04T10:11:07.2791308Z s3-prefix: pytorch/pytorch/19922768520/1/artifact 2025-12-04T10:11:07.2791552Z retention-days: 14 2025-12-04T10:11:07.2791726Z if-no-files-found: warn 2025-12-04T10:11:07.2791920Z path: test-jsons-*.zip 2025-12-04T10:11:07.2792091Z name: artifact 2025-12-04T10:11:07.2792246Z region: us-east-1 2025-12-04T10:11:07.2792401Z env: 2025-12-04T10:11:07.2792554Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:07.2792743Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:07.2792969Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:07.2793354Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:07.2793796Z DEVICE_NAME: 2025-12-04T10:11:07.2793953Z DEVICE_TYPE: 2025-12-04T10:11:07.2794102Z ##[endgroup] 2025-12-04T10:11:07.8270673Z NOTE: s3-prefix specified, ignoring name parameter 2025-12-04T10:11:07.8271117Z With the provided path, there will be 1 file uploaded 2025-12-04T10:11:07.8271526Z Uploading to s3 prefix: pytorch/pytorch/19922768520/1/artifact 2025-12-04T10:11:07.8324444Z Starting upload of test-jsons-test-default-4-5-lf.linux.g6.4xlarge.experimental.nvidia.gpu_57116084869.zip 2025-12-04T10:11:07.9589997Z Finished upload of test-jsons-test-default-4-5-lf.linux.g6.4xlarge.experimental.nvidia.gpu_57116084869.zip 2025-12-04T10:11:07.9823098Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T10:11:07.9823329Z with: 2025-12-04T10:11:07.9823491Z s3-bucket: gha-artifacts 2025-12-04T10:11:07.9823726Z s3-prefix: pytorch/pytorch/19922768520/1/artifact 2025-12-04T10:11:07.9823974Z retention-days: 14 2025-12-04T10:11:07.9824158Z if-no-files-found: error 2025-12-04T10:11:07.9824349Z path: test-reports-*.zip 2025-12-04T10:11:07.9824527Z name: artifact 2025-12-04T10:11:07.9824680Z region: us-east-1 2025-12-04T10:11:07.9824849Z env: 2025-12-04T10:11:07.9824995Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:07.9825176Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:07.9825408Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:07.9825824Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:07.9826172Z DEVICE_NAME: 2025-12-04T10:11:07.9826332Z DEVICE_TYPE: 2025-12-04T10:11:07.9826498Z ##[endgroup] 2025-12-04T10:11:08.2716691Z NOTE: s3-prefix specified, ignoring name parameter 2025-12-04T10:11:08.2717125Z With the provided path, there will be 1 file uploaded 2025-12-04T10:11:08.2717535Z Uploading to s3 prefix: pytorch/pytorch/19922768520/1/artifact 2025-12-04T10:11:08.2769655Z Starting upload of test-reports-test-default-4-5-lf.linux.g6.4xlarge.experimental.nvidia.gpu_57116084869.zip 2025-12-04T10:11:08.4066369Z Finished upload of test-reports-test-default-4-5-lf.linux.g6.4xlarge.experimental.nvidia.gpu_57116084869.zip 2025-12-04T10:11:08.4289648Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T10:11:08.4289879Z with: 2025-12-04T10:11:08.4290041Z s3-bucket: gha-artifacts 2025-12-04T10:11:08.4290286Z s3-prefix: pytorch/pytorch/19922768520/1/artifact 2025-12-04T10:11:08.4290533Z retention-days: 14 2025-12-04T10:11:08.4290706Z if-no-files-found: ignore 2025-12-04T10:11:08.4290893Z path: logs-*.zip 2025-12-04T10:11:08.4291056Z name: artifact 2025-12-04T10:11:08.4291373Z region: us-east-1 2025-12-04T10:11:08.4291553Z env: 2025-12-04T10:11:08.4291699Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:08.4291879Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:08.4292103Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:08.4292489Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:08.4292840Z DEVICE_NAME: 2025-12-04T10:11:08.4293000Z DEVICE_TYPE: 2025-12-04T10:11:08.4293154Z ##[endgroup] 2025-12-04T10:11:08.7165531Z NOTE: s3-prefix specified, ignoring name parameter 2025-12-04T10:11:08.7166215Z With the provided path, there will be 1 file uploaded 2025-12-04T10:11:08.7166789Z Uploading to s3 prefix: pytorch/pytorch/19922768520/1/artifact 2025-12-04T10:11:08.7218760Z Starting upload of logs-test-default-4-5-lf.linux.g6.4xlarge.experimental.nvidia.gpu_57116084869.zip 2025-12-04T10:11:08.8585896Z Finished upload of logs-test-default-4-5-lf.linux.g6.4xlarge.experimental.nvidia.gpu_57116084869.zip 2025-12-04T10:11:08.9110522Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T10:11:08.9110761Z with: 2025-12-04T10:11:08.9110920Z s3-bucket: gha-artifacts 2025-12-04T10:11:08.9111167Z s3-prefix: pytorch/pytorch/19922768520/1/artifact 2025-12-04T10:11:08.9111418Z retention-days: 14 2025-12-04T10:11:08.9111597Z if-no-files-found: ignore 2025-12-04T10:11:08.9111910Z path: debug-*.zip 2025-12-04T10:11:08.9112081Z name: artifact 2025-12-04T10:11:08.9112237Z region: us-east-1 2025-12-04T10:11:08.9112401Z env: 2025-12-04T10:11:08.9112571Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:08.9112844Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:08.9113080Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:08.9113488Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:08.9113833Z DEVICE_NAME: 2025-12-04T10:11:08.9113999Z DEVICE_TYPE: 2025-12-04T10:11:08.9114153Z ##[endgroup] 2025-12-04T10:11:09.1952293Z No files were found with the provided path: debug-*.zip. No artifacts will be uploaded. 2025-12-04T10:11:09.2190521Z ##[group]Run # shellcheck disable=SC2156 2025-12-04T10:11:09.2190831Z # shellcheck disable=SC2156 2025-12-04T10:11:09.2191267Z find . -iname "core.[1-9]*" -exec docker exec "${DOCKER_CONTAINER_ID}" sh -c "gdb python {} -ex 'bt' -ex 'q'" \; 2025-12-04T10:11:09.2199044Z shell: /usr/bin/bash -e {0} 2025-12-04T10:11:09.2199241Z env: 2025-12-04T10:11:09.2199409Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:09.2199595Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:09.2199830Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:09.2200228Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:09.2200573Z DEVICE_NAME: 2025-12-04T10:11:09.2200726Z DEVICE_TYPE: 2025-12-04T10:11:09.2200881Z ##[endgroup] 2025-12-04T10:11:09.5628793Z Prepare all required actions 2025-12-04T10:11:09.5629144Z Getting action download info 2025-12-04T10:11:09.8403270Z Download action repository 'actions/setup-python@v6' (SHA:83679a892e2d95755f2dac6acb0bfd1e9ac5d548) 2025-12-04T10:11:11.5815058Z ##[group]Run ./.github/actions/upload-utilization-stats 2025-12-04T10:11:11.5815337Z with: 2025-12-04T10:11:11.5815489Z job_id: 57116084869 2025-12-04T10:11:11.5815925Z job_name: linux-jammy-cuda12.8-py3.10-gcc11 / test (default, 4, 5, lf.linux.g6.4xlarge.experimental.nvidia.gpu, rerun_disabled_tests) 2025-12-04T10:11:11.5816419Z workflow_name: trunk 2025-12-04T10:11:11.5816595Z workflow_run_id: 19922768520 2025-12-04T10:11:11.5816798Z workflow_attempt: 1 2025-12-04T10:11:11.5816963Z env: 2025-12-04T10:11:11.5817104Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:11.5817307Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:11.5817527Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:11.5818079Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:11.5818496Z DEVICE_NAME: 2025-12-04T10:11:11.5818897Z DEVICE_TYPE: 2025-12-04T10:11:11.5819164Z ##[endgroup] 2025-12-04T10:11:11.5931240Z ##[group]Run actions/setup-python@v6 2025-12-04T10:11:11.5931642Z with: 2025-12-04T10:11:11.5931900Z python-version: 3.10 2025-12-04T10:11:11.5932147Z check-latest: false 2025-12-04T10:11:11.5932660Z token: *** 2025-12-04T10:11:11.5932932Z update-environment: true 2025-12-04T10:11:11.5933295Z allow-prereleases: false 2025-12-04T10:11:11.5933653Z freethreaded: false 2025-12-04T10:11:11.5933913Z env: 2025-12-04T10:11:11.5934159Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:11.5934552Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:11.5934864Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:11.5935346Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:11.5935813Z DEVICE_NAME: 2025-12-04T10:11:11.5936071Z DEVICE_TYPE: 2025-12-04T10:11:11.5936314Z ##[endgroup] 2025-12-04T10:11:11.9092141Z ##[group]Installed versions 2025-12-04T10:11:11.9100748Z Version 3.10 was not found in the local cache 2025-12-04T10:11:11.9246753Z (node:156560) [DEP0040] DeprecationWarning: The `punycode` module is deprecated. Please use a userland alternative instead. 2025-12-04T10:11:11.9247579Z (Use `node --trace-deprecation ...` to show where the warning was created) 2025-12-04T10:11:12.3400107Z ##[error]The version '3.10' with architecture 'x64' was not found for this operating system. The list of all available versions can be found here: https://raw.githubusercontent.com/actions/python-versions/main/versions-manifest.json 2025-12-04T10:11:12.3560240Z ##[group]Run pytorch/test-infra/.github/actions/teardown-linux@main 2025-12-04T10:11:12.3560591Z with: 2025-12-04T10:11:12.3560747Z env: 2025-12-04T10:11:12.3560899Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:12.3561093Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:12.3561334Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:12.3561729Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:12.3562079Z DEVICE_NAME: 2025-12-04T10:11:12.3562235Z DEVICE_TYPE: 2025-12-04T10:11:12.3562388Z ##[endgroup] 2025-12-04T10:11:12.3658669Z ##[group]Run set -eou pipefail 2025-12-04T10:11:12.3658910Z set -eou pipefail 2025-12-04T10:11:12.3659088Z  2025-12-04T10:11:12.3659357Z echo "Holding runner for 2 hours until all ssh sessions have logged out" 2025-12-04T10:11:12.3659693Z for _ in $(seq 1440); do 2025-12-04T10:11:12.3659955Z  # Break if no ssh session exists anymore 2025-12-04T10:11:12.3660215Z  if [ "$(who)" = "" ]; then 2025-12-04T10:11:12.3660446Z  break 2025-12-04T10:11:12.3660604Z  fi 2025-12-04T10:11:12.3660763Z  echo "." 2025-12-04T10:11:12.3660934Z  sleep 5 2025-12-04T10:11:12.3661095Z done 2025-12-04T10:11:12.3668557Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:11:12.3668831Z env: 2025-12-04T10:11:12.3668986Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:12.3669169Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:12.3669393Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:12.3669789Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:12.3670131Z DEVICE_NAME: 2025-12-04T10:11:12.3670300Z DEVICE_TYPE: 2025-12-04T10:11:12.3670458Z ##[endgroup] 2025-12-04T10:11:12.3698377Z Holding runner for 2 hours until all ssh sessions have logged out 2025-12-04T10:11:12.4217207Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T10:11:12.4217623Z # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T10:11:12.4217933Z # shellcheck disable=SC2046 2025-12-04T10:11:12.4218184Z docker stop $(docker ps -q) || true 2025-12-04T10:11:12.4218437Z # Prune all of the docker images 2025-12-04T10:11:12.4218675Z docker system prune -af 2025-12-04T10:11:12.4225657Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:11:12.4225931Z env: 2025-12-04T10:11:12.4226090Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:11:12.4226285Z HAS_NVIDIA_GPU: true 2025-12-04T10:11:12.4226508Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T10:11:12.4226909Z DOCKER_CONTAINER_ID: 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:12.4227255Z DEVICE_NAME: 2025-12-04T10:11:12.4227407Z DEVICE_TYPE: 2025-12-04T10:11:12.4227559Z ##[endgroup] 2025-12-04T10:11:33.5378935Z 364ec7eaeaf9 2025-12-04T10:11:34.6701762Z Deleted Containers: 2025-12-04T10:11:34.6702207Z 364ec7eaeaf9029d46cc3e7596ee62016baf33050267b20b0d03707821ee5e76 2025-12-04T10:11:34.6702522Z 2025-12-04T10:11:46.4254682Z Deleted Images: 2025-12-04T10:11:46.4255999Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T10:11:46.4257590Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image@sha256:ba21003510dba4bdeed83df81a56fa468e0ee1b612a9445ae1f402a280804f97 2025-12-04T10:11:46.4260527Z deleted: sha256:add7313791033822205cdb3cf32096534b2cfaa4855bd48119b59000bfe00301 2025-12-04T10:11:46.4261008Z deleted: sha256:85a76b7bf29ad34eb76cce6f46af5d49a58b6272f80f983d5c769e82c7749301 2025-12-04T10:11:46.4261729Z deleted: sha256:0882f3ce59ff5ae30195ee4b059fc713e13eda107a3a7814a4616ac9058a30a4 2025-12-04T10:11:46.4262410Z deleted: sha256:64ba5b9344c11a3e4729136076830b90ac4cf1554046edb1bd4f0784b66ebd9b 2025-12-04T10:11:46.4262962Z deleted: sha256:88213c59cf461a65ab9b6cb07b4195dc9d41b5241c152daa002c7b3112e09124 2025-12-04T10:11:46.4263406Z deleted: sha256:4c0f83afa802ffbc05ebaf1aa50e48a2447c7c295549a6dded80ac63437906ca 2025-12-04T10:11:46.4263849Z deleted: sha256:6f7ec74460e8fb070c8209949095ea3be5f4e2fd69c9f750cd39ac4093f5e64b 2025-12-04T10:11:46.4264277Z deleted: sha256:d6928b0d1021b31942fdcb64e5eb4a34682de66e959dd424ed6ed02c29cd706d 2025-12-04T10:11:46.4264708Z deleted: sha256:4e9fbcb1705a6351bb34dd320558752614308636b94fd9ae6f26063e3deadc0a 2025-12-04T10:11:46.4265130Z deleted: sha256:43aabd0201f48712f21758071352dea029b4de37be08b2e2197706856a9ecbf2 2025-12-04T10:11:46.4265547Z deleted: sha256:940a98dec78303f0548beb1033242a45e9097607ef3e55c8b949b69b73d1b95e 2025-12-04T10:11:46.4265986Z deleted: sha256:d2849fa0e0411cf66e4408831d70e38838afb55b11a80c1c4d8aa0ae7dc9ca40 2025-12-04T10:11:46.4266422Z deleted: sha256:14f40d23c20c7e562623f89deb376520296758bc39dd3c77284049b84ebd8a31 2025-12-04T10:11:46.4266860Z deleted: sha256:a8ccba61f90ca097cb391d0f4fbed0d9f821d06b00e28f7332e9e2dcfcbac4ca 2025-12-04T10:11:46.4267289Z deleted: sha256:91b2060d290547d3b517d4a11d994bbe23f4560b5546cb91918ca1828dde6be1 2025-12-04T10:11:46.4267713Z deleted: sha256:b42a184755715dcfead7fad655a127433541d316d9628f5f730ff17ad5f8071c 2025-12-04T10:11:46.4268147Z deleted: sha256:aa5b4f3c9169061dc3c6da0e677e8a86f11ecb0a3f9fb4861ab3d8c04379775c 2025-12-04T10:11:46.4268579Z deleted: sha256:b4dcf450081a48d77fea0a21b8d810a69c03608a595e754fe7d365058d0579b7 2025-12-04T10:11:46.4269017Z deleted: sha256:4f7fe12d3d4f5bf890c7ada4ce16f17a105472aa6509a778f917dcce2f28174b 2025-12-04T10:11:46.4269452Z deleted: sha256:2d1d5a74182594f9a8553df00fdcfc809dba407bcd6700d667f862cbe9d555ce 2025-12-04T10:11:46.4269899Z deleted: sha256:d901e2f5d449aeed16b727bdcc11fc0e0f6c30c8fc5c39ac7eeac8a74d9d176c 2025-12-04T10:11:46.4270339Z deleted: sha256:a04df2603bd12372c6632469a9a81ebc4a8d677452c250672b9692884fa6a452 2025-12-04T10:11:46.4270767Z deleted: sha256:f438a6b52273a552dc3820a55c74c53a62a0eae9f2a7d21b37125add7d71639f 2025-12-04T10:11:46.4271195Z deleted: sha256:d4b09517e9518d709ac98b0ae6f8446ec9ac51688253607b1fca67aa2c87b3f4 2025-12-04T10:11:46.4271622Z deleted: sha256:c1fa38335237f5e7263e39d3d3de98215bcfbbb12b826955c02e149bf68efd13 2025-12-04T10:11:46.4272047Z deleted: sha256:c898d20a30de901fca74d7611663b17ab48e1726a11e031e40548ed16ee81877 2025-12-04T10:11:46.4272474Z deleted: sha256:3baceec7096518fcc10696feba551639d698b3145c2fc09cac927bb60c0fd751 2025-12-04T10:11:46.4272908Z deleted: sha256:5245aaaa3d5c3a19f76b9a6c920bd82d1a0ff5289f87c8c109652089709d9b3b 2025-12-04T10:11:46.4273329Z deleted: sha256:f05cc789b95246938c377f474c41187965b89ceac0250e7d5124bec32153f447 2025-12-04T10:11:46.4273797Z deleted: sha256:07ec4fc008de4e7a2c794ec7094cc72e0d287c04c8b2156163aee0bae147fe2d 2025-12-04T10:11:46.4274238Z deleted: sha256:c6302601ad5fde573c1f8c900250478fca7fdc6907d8fd4fae651b94b4d9264d 2025-12-04T10:11:46.4274672Z deleted: sha256:cc5e955ee1dc54931f02606c5ea87aae14f03b5d764492be611480ab041f2882 2025-12-04T10:11:46.4275106Z deleted: sha256:f21c03518996d98452338f4e80bcfd9b139a1dab155f4830be0d3f623035269f 2025-12-04T10:11:46.4275529Z deleted: sha256:519ca6f1279f7886f25f0005527cfa627deebbc5b7d7cdbfa7ef962bcfc4c26d 2025-12-04T10:11:46.4275954Z deleted: sha256:0ef990495216807d0175b192045be3f617e72331bc373b3434807f41bf69168d 2025-12-04T10:11:46.4276376Z deleted: sha256:7093edf7319e1f0e01654c3224e32c8dede5b948d106e0b9b03cbf0bb1091e33 2025-12-04T10:11:46.4276795Z deleted: sha256:c478161e058e2f4041555c3e880b95ee1ee047938dc58549a3a88135740996ae 2025-12-04T10:11:46.4277223Z deleted: sha256:9bb853b0d938cd7c36a80ce8ee40653f2c0ff92719209b11beb03acc8855ce3e 2025-12-04T10:11:46.4277657Z deleted: sha256:fdf2ace71a78ce6910ef9c4b073c195531da47022443b606bb92dcd6499b6afc 2025-12-04T10:11:46.4278172Z deleted: sha256:576c2b3770d871937d3cfb7014328bcb4bd1aed0c28bc438764b3bfdac4c1ac2 2025-12-04T10:11:46.4278954Z deleted: sha256:878e92b9cb82de09ac14a9d5f3f7bc2411a799b6f54d0d64b78c2bb4d1fdc0fc 2025-12-04T10:11:46.4279612Z deleted: sha256:85c8c3b98b65a6695f988a10cc66c981d73a3ef03eda15b8e14d227b50b56300 2025-12-04T10:11:46.4280071Z deleted: sha256:ce2ab3ba07794f9ee95d6ea7de6dcd3d2aed96561f9a79192dd56ca5bf29313a 2025-12-04T10:11:46.4280522Z deleted: sha256:37a6e12976ca957286977e696e63012ab9821214b0483fe1a48d29dcb280508a 2025-12-04T10:11:46.4280959Z deleted: sha256:cd1d5d3dd7038144ca6fe961c0d4c8e705625ae0c36190ba8b3e9602abedad19 2025-12-04T10:11:46.4281396Z deleted: sha256:0e707276e0be2e0008b86d594fadc0d16444d66c4fb7227c56f144cbb3c2affd 2025-12-04T10:11:46.4281826Z deleted: sha256:22d4aad6a2ada91b341c1225a0f314042b8aeabef7568c5c019709b058bf070b 2025-12-04T10:11:46.4282257Z deleted: sha256:ee4adacf4e0933131d0275eddad406b3c8147e6cf07a292b99f1aff4b5355f33 2025-12-04T10:11:46.4282699Z deleted: sha256:43da0b9e7c0e18403dcb834e53628dc7c970ccb2dbd091878c0d7c0170dbc97f 2025-12-04T10:11:46.4283134Z deleted: sha256:00571684bdcd75beda15eb7d4e79b5458bc914350f9bb4d87fcdc97ad15e0da1 2025-12-04T10:11:46.4283564Z deleted: sha256:41615f09950259f1d75e82ef35b6fc53b18fe71ebff143744cfd51009d04349e 2025-12-04T10:11:46.4283993Z deleted: sha256:75ab34d2eed3c7915467a506ab6dab2711918fbabe94add2fb5c62780221ab0c 2025-12-04T10:11:46.4284429Z deleted: sha256:0a39ef2bebf44c1c3893d1e5fb42dad48b8fac7ca673141267ee967f85455e89 2025-12-04T10:11:46.4284867Z deleted: sha256:9b7d024e48ba1f9824a54597621b1b062cbc4aa41a77d81ca538d6b5c24a612c 2025-12-04T10:11:46.4285282Z deleted: sha256:392257172de6434c271bd93394218a91e9aa86d7c18abc2f2759317b9d5fb6de 2025-12-04T10:11:46.4285695Z deleted: sha256:6c3232860b930866a463a356124fc392c7e5f04895695229257e8c3e8a02711d 2025-12-04T10:11:46.4286122Z deleted: sha256:63dd55b807215e2fa6c715419ac0c5072d02dddc848dbf74bb7e77b906b5eaed 2025-12-04T10:11:46.4286549Z deleted: sha256:07a8738c1b4584db72ed9aa60f5274321eb0ba16263450da3a75df8326ebc25f 2025-12-04T10:11:46.4286966Z deleted: sha256:053fe2965b01281d12040ec1893e0d1aa77362a49ea9a1067402272c69dad9f5 2025-12-04T10:11:46.4287393Z deleted: sha256:7857fb5eb181c4e80262ecab60bdd3c266cf3d1409ceb76c05882609b416a8d3 2025-12-04T10:11:46.4287829Z deleted: sha256:752528477fc99089de3bd2c6da7b30cf34f2e901fe06d8fcfe685b411461e883 2025-12-04T10:11:46.4288259Z deleted: sha256:cce0210e2f4b042601813df03aa294a86b0c668fcfc75f4c63f6fa12b2952e15 2025-12-04T10:11:46.4288682Z deleted: sha256:f2bb405a26705ecd12d21380d26d9355d01db3a2175080fbdb468f2b5a25a76c 2025-12-04T10:11:46.4289120Z deleted: sha256:ad430120d4ffbaf97cd8d6de6ea8eefa4a8f80ec45f0b176c6b26bff0970fd33 2025-12-04T10:11:46.4289561Z deleted: sha256:225a4910baea7cc540ed43eeac75046293800ab0b8e0192b51e991c8cb50bcf3 2025-12-04T10:11:46.4289989Z deleted: sha256:a259945b0c3507f049fbac10fb3d3ffe43d45e83c91b80ae8cd1dafb855ad83c 2025-12-04T10:11:46.4290418Z deleted: sha256:862a98881b1d5adad5c21d01602773b894794097de80964ef8f47bcaadb43255 2025-12-04T10:11:46.4290841Z deleted: sha256:1cf6d3c8b6c2694b79a2d08719594903811c330a36a4c7a8a7153a350b53d292 2025-12-04T10:11:46.4291272Z deleted: sha256:232a1ae8b0fee817ff7838bb5986a2f38377d3b1dbbf5217b576af0f953b0844 2025-12-04T10:11:46.4291704Z deleted: sha256:c72c5705dabd6314423dd7d4fb260a20d5d9886b2ebce60d19e9d78c4a2335c2 2025-12-04T10:11:46.4292129Z deleted: sha256:296734cf81fd92c913884d058908598424ffe072676e38de289bbab83768c7bd 2025-12-04T10:11:46.4292557Z deleted: sha256:7c76040481b889847a1804021aeff07547eaa4ee706d6137db218d497a8fd9c1 2025-12-04T10:11:46.4292985Z deleted: sha256:d5e293f5b354e8cbcc6de893ea72cc632b02d8fdfbb08ec3127c4e9662f3ebff 2025-12-04T10:11:46.4293526Z deleted: sha256:f35a64e429c88e249645090f21fbe7dae108d98e0ab4ea13184f24b3fd66c315 2025-12-04T10:11:46.4293959Z deleted: sha256:ce6ae8d595c8e69115c51b1ce4f9a9158795d7b863b1cb53f21c39a87974d41b 2025-12-04T10:11:46.4294389Z deleted: sha256:8941abaee59400fb9b3a60765fea4a1fc2a6a447467a6d983e84c7f72494a450 2025-12-04T10:11:46.4294876Z deleted: sha256:ef53c29a9a2c2bc80ffdb9bfaf92842436b5755ec1ce828b9d11e5e27d656ea1 2025-12-04T10:11:46.4295327Z deleted: sha256:7a347fb0acb43f1c814f8c8ff21185e8b5cf64d7bc5988cea060f77d906e08b5 2025-12-04T10:11:46.4296547Z deleted: sha256:cc855dc9be79496e15175569dced2d13477e50b077a5fd3945f9bf50018880c1 2025-12-04T10:11:46.4296986Z deleted: sha256:f7a9946ada3d4786658bc0b643808bb32a9a45e4e90e30dc43ee19e2dbe24024 2025-12-04T10:11:46.4297420Z deleted: sha256:c22a9215f62812c1d2e32827f5221ff556c5b6702aadbdab6b87b8293f19635e 2025-12-04T10:11:46.4297855Z deleted: sha256:959a56746620012e37c1def1a83c5afb1e7c0adc59b021a28beb53c24df98032 2025-12-04T10:11:46.4298287Z deleted: sha256:31a0fff0695bf6100c17954be72eab2095b466d559c75c3faf2a17d8c41e6ebe 2025-12-04T10:11:46.4298705Z deleted: sha256:c15e2b5241b9e55af1b2593e544391b4b44d0505e6528e8f12425136e93b424c 2025-12-04T10:11:46.4299144Z deleted: sha256:73974f74b436f39a2fdb6461b1e3f7c3e41c73325776fa71d16b942a5b4a365b 2025-12-04T10:11:46.4299784Z untagged: public.ecr.aws/docker/library/python:3.13 2025-12-04T10:11:46.4300611Z untagged: public.ecr.aws/docker/library/python@sha256:3f986299a7b8b44b0d8cf9bda2b22361ce5c3058ef5d7cb17fb7452506680ab0 2025-12-04T10:11:46.4301553Z deleted: sha256:44438aecfedf7b6086fce506dae0db5ba7fc0027f9b743f1a75a6b5cbc7de70a 2025-12-04T10:11:46.4302026Z deleted: sha256:6f09a1f5d8a107c2532fbd116e75116cb75fa77b1a7d72d3bdf1ac12de152acd 2025-12-04T10:11:46.4302473Z deleted: sha256:fe5f3ac0be086125eb1e3cd10cc33e8e426f4e079381f7ce5a987b626e99fa67 2025-12-04T10:11:46.4302909Z deleted: sha256:79dd2061a22cf919cfc4f1f02704bfda09afadb017265e670ee54441d296c06c 2025-12-04T10:11:46.4303348Z deleted: sha256:9447ad402aafdbee17e999b0ec84ad89c2646dbebf054d469d4f8bee77f66212 2025-12-04T10:11:46.4303781Z deleted: sha256:7a4909f3c1975be52292f53107495ee1b41c17494918767ccedf1cf1688ae318 2025-12-04T10:11:46.4304202Z deleted: sha256:3474923d97f1f498237650a7d51bd4aea37d5e6b9d8a778777920584af5dd560 2025-12-04T10:11:46.4304897Z deleted: sha256:683afd1773444401a9cbd24842ee5d9154a11abb4fab63ddea5c03df788597ee 2025-12-04T10:11:46.4305183Z 2025-12-04T10:11:46.4305275Z Total reclaimed space: 35.6GB 2025-12-04T10:11:46.4364191Z Post job cleanup. 2025-12-04T10:11:46.4393468Z Post job cleanup. 2025-12-04T10:11:46.5601986Z (node:156721) [DEP0040] DeprecationWarning: The `punycode` module is deprecated. Please use a userland alternative instead. 2025-12-04T10:11:46.5602771Z (Use `node --trace-deprecation ...` to show where the warning was created) 2025-12-04T10:11:46.5750129Z Post job cleanup. 2025-12-04T10:11:46.5786392Z Post job cleanup. 2025-12-04T10:11:46.6650199Z [command]/usr/bin/git version 2025-12-04T10:11:46.6689151Z git version 2.50.1 2025-12-04T10:11:46.6722285Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/9046f717-3cf4-45de-8fe2-845e3d9a3a71/.gitconfig' 2025-12-04T10:11:46.6731498Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/9046f717-3cf4-45de-8fe2-845e3d9a3a71' before making global git config changes 2025-12-04T10:11:46.6732417Z Adding repository directory to the temporary git global config as a safe directory 2025-12-04T10:11:46.6736346Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T10:11:46.6776415Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-12-04T10:11:46.6814860Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-12-04T10:11:46.7144890Z Entering 'android/libs/fbjni' 2025-12-04T10:11:46.7208100Z Entering 'third_party/FP16' 2025-12-04T10:11:46.7273191Z Entering 'third_party/FXdiv' 2025-12-04T10:11:46.7337009Z Entering 'third_party/NNPACK' 2025-12-04T10:11:46.7403056Z Entering 'third_party/NVTX' 2025-12-04T10:11:46.7474041Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T10:11:46.7543068Z Entering 'third_party/XNNPACK' 2025-12-04T10:11:46.7620419Z Entering 'third_party/aiter' 2025-12-04T10:11:46.7684569Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T10:11:46.7756678Z Entering 'third_party/benchmark' 2025-12-04T10:11:46.7818461Z Entering 'third_party/composable_kernel' 2025-12-04T10:11:46.7890788Z Entering 'third_party/cpp-httplib' 2025-12-04T10:11:46.7959345Z Entering 'third_party/cpuinfo' 2025-12-04T10:11:46.8024359Z Entering 'third_party/cudnn_frontend' 2025-12-04T10:11:46.8087314Z Entering 'third_party/cutlass' 2025-12-04T10:11:46.8161887Z Entering 'third_party/fbgemm' 2025-12-04T10:11:46.8229939Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T10:11:46.8292798Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T10:11:46.8366986Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T10:11:46.8435003Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T10:11:46.8512594Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T10:11:46.8578446Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T10:11:46.8646692Z Entering 'third_party/fbgemm/external/json' 2025-12-04T10:11:46.8719328Z Entering 'third_party/flash-attention' 2025-12-04T10:11:46.8783118Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T10:11:46.8852390Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T10:11:46.8930342Z Entering 'third_party/flatbuffers' 2025-12-04T10:11:46.8998311Z Entering 'third_party/fmt' 2025-12-04T10:11:46.9063163Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T10:11:46.9127137Z Entering 'third_party/gloo' 2025-12-04T10:11:46.9193482Z Entering 'third_party/googletest' 2025-12-04T10:11:46.9257185Z Entering 'third_party/ideep' 2025-12-04T10:11:46.9321440Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T10:11:46.9392016Z Entering 'third_party/ittapi' 2025-12-04T10:11:46.9457824Z Entering 'third_party/kineto' 2025-12-04T10:11:46.9522278Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T10:11:46.9587269Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T10:11:46.9654649Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T10:11:46.9724748Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T10:11:46.9787666Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T10:11:46.9846604Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T10:11:46.9915928Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T10:11:46.9979013Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T10:11:47.0043356Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T10:11:47.0108785Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T10:11:47.0173706Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T10:11:47.0235175Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:11:47.0300884Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:11:47.0371713Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T10:11:47.0434815Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T10:11:47.0500855Z Entering 'third_party/kleidiai' 2025-12-04T10:11:47.0565660Z Entering 'third_party/mimalloc' 2025-12-04T10:11:47.0634390Z Entering 'third_party/nlohmann' 2025-12-04T10:11:47.0698793Z Entering 'third_party/onnx' 2025-12-04T10:11:47.0777464Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T10:11:47.0847567Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T10:11:47.0914715Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T10:11:47.0977382Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T10:11:47.1037676Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T10:11:47.1099172Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T10:11:47.1163884Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T10:11:47.1227448Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T10:11:47.1295315Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T10:11:47.1361061Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:11:47.1426312Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:11:47.1496030Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T10:11:47.1581583Z Entering 'third_party/pocketfft' 2025-12-04T10:11:47.1644537Z Entering 'third_party/protobuf' 2025-12-04T10:11:47.1711915Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T10:11:47.1773624Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T10:11:47.1840000Z Entering 'third_party/psimd' 2025-12-04T10:11:47.1908418Z Entering 'third_party/pthreadpool' 2025-12-04T10:11:47.1976987Z Entering 'third_party/pybind11' 2025-12-04T10:11:47.2042972Z Entering 'third_party/python-peachpy' 2025-12-04T10:11:47.2107101Z Entering 'third_party/sleef' 2025-12-04T10:11:47.2172104Z Entering 'third_party/tensorpipe' 2025-12-04T10:11:47.2233465Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T10:11:47.2296410Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T10:11:47.2356980Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T10:11:47.2418734Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T10:11:47.2481601Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T10:11:47.2580837Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-12-04T10:11:47.2603577Z http.https://github.com/.extraheader 2025-12-04T10:11:47.2613671Z [command]/usr/bin/git config --local --unset-all http.https://github.com/.extraheader 2025-12-04T10:11:47.2646805Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-12-04T10:11:47.2979404Z Entering 'android/libs/fbjni' 2025-12-04T10:11:47.3022191Z http.https://github.com/.extraheader 2025-12-04T10:11:47.3059643Z Entering 'third_party/FP16' 2025-12-04T10:11:47.3101861Z http.https://github.com/.extraheader 2025-12-04T10:11:47.3142849Z Entering 'third_party/FXdiv' 2025-12-04T10:11:47.3183852Z http.https://github.com/.extraheader 2025-12-04T10:11:47.3227541Z Entering 'third_party/NNPACK' 2025-12-04T10:11:47.3269842Z http.https://github.com/.extraheader 2025-12-04T10:11:47.3309241Z Entering 'third_party/NVTX' 2025-12-04T10:11:47.3350936Z http.https://github.com/.extraheader 2025-12-04T10:11:47.3392350Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T10:11:47.3432000Z http.https://github.com/.extraheader 2025-12-04T10:11:47.3473190Z Entering 'third_party/XNNPACK' 2025-12-04T10:11:47.3514124Z http.https://github.com/.extraheader 2025-12-04T10:11:47.3567460Z Entering 'third_party/aiter' 2025-12-04T10:11:47.3611460Z http.https://github.com/.extraheader 2025-12-04T10:11:47.3648821Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T10:11:47.3690598Z http.https://github.com/.extraheader 2025-12-04T10:11:47.3739033Z Entering 'third_party/benchmark' 2025-12-04T10:11:47.3781383Z http.https://github.com/.extraheader 2025-12-04T10:11:47.3820181Z Entering 'third_party/composable_kernel' 2025-12-04T10:11:47.3861493Z http.https://github.com/.extraheader 2025-12-04T10:11:47.3914851Z Entering 'third_party/cpp-httplib' 2025-12-04T10:11:47.3956557Z http.https://github.com/.extraheader 2025-12-04T10:11:47.3996291Z Entering 'third_party/cpuinfo' 2025-12-04T10:11:47.4036355Z http.https://github.com/.extraheader 2025-12-04T10:11:47.4075944Z Entering 'third_party/cudnn_frontend' 2025-12-04T10:11:47.4121909Z http.https://github.com/.extraheader 2025-12-04T10:11:47.4161181Z Entering 'third_party/cutlass' 2025-12-04T10:11:47.4202792Z http.https://github.com/.extraheader 2025-12-04T10:11:47.4252110Z Entering 'third_party/fbgemm' 2025-12-04T10:11:47.4295165Z http.https://github.com/.extraheader 2025-12-04T10:11:47.4336805Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T10:11:47.4378927Z http.https://github.com/.extraheader 2025-12-04T10:11:47.4418259Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T10:11:47.4460798Z http.https://github.com/.extraheader 2025-12-04T10:11:47.4514917Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T10:11:47.4556622Z http.https://github.com/.extraheader 2025-12-04T10:11:47.4596406Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T10:11:47.4638614Z http.https://github.com/.extraheader 2025-12-04T10:11:47.4685205Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T10:11:47.4731960Z http.https://github.com/.extraheader 2025-12-04T10:11:47.4772937Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T10:11:47.4813501Z http.https://github.com/.extraheader 2025-12-04T10:11:47.4853487Z Entering 'third_party/fbgemm/external/json' 2025-12-04T10:11:47.4894391Z http.https://github.com/.extraheader 2025-12-04T10:11:47.4939834Z Entering 'third_party/flash-attention' 2025-12-04T10:11:47.4983612Z http.https://github.com/.extraheader 2025-12-04T10:11:47.5024235Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T10:11:47.5065223Z http.https://github.com/.extraheader 2025-12-04T10:11:47.5111589Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T10:11:47.5152180Z http.https://github.com/.extraheader 2025-12-04T10:11:47.5202395Z Entering 'third_party/flatbuffers' 2025-12-04T10:11:47.5243764Z http.https://github.com/.extraheader 2025-12-04T10:11:47.5287592Z Entering 'third_party/fmt' 2025-12-04T10:11:47.5331355Z http.https://github.com/.extraheader 2025-12-04T10:11:47.5377878Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T10:11:47.5421025Z http.https://github.com/.extraheader 2025-12-04T10:11:47.5460968Z Entering 'third_party/gloo' 2025-12-04T10:11:47.5502502Z http.https://github.com/.extraheader 2025-12-04T10:11:47.5546356Z Entering 'third_party/googletest' 2025-12-04T10:11:47.5589311Z http.https://github.com/.extraheader 2025-12-04T10:11:47.5629249Z Entering 'third_party/ideep' 2025-12-04T10:11:47.5670672Z http.https://github.com/.extraheader 2025-12-04T10:11:47.5710905Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T10:11:47.5751273Z http.https://github.com/.extraheader 2025-12-04T10:11:47.5803779Z Entering 'third_party/ittapi' 2025-12-04T10:11:47.5845781Z http.https://github.com/.extraheader 2025-12-04T10:11:47.5885777Z Entering 'third_party/kineto' 2025-12-04T10:11:47.5932706Z http.https://github.com/.extraheader 2025-12-04T10:11:47.5976360Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T10:11:47.6021578Z http.https://github.com/.extraheader 2025-12-04T10:11:47.6065332Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T10:11:47.6108085Z http.https://github.com/.extraheader 2025-12-04T10:11:47.6149475Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T10:11:47.6191815Z http.https://github.com/.extraheader 2025-12-04T10:11:47.6234705Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T10:11:47.6276138Z http.https://github.com/.extraheader 2025-12-04T10:11:47.6316948Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T10:11:47.6357429Z http.https://github.com/.extraheader 2025-12-04T10:11:47.6396600Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T10:11:47.6438888Z http.https://github.com/.extraheader 2025-12-04T10:11:47.6480909Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T10:11:47.6522638Z http.https://github.com/.extraheader 2025-12-04T10:11:47.6568421Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T10:11:47.6611066Z http.https://github.com/.extraheader 2025-12-04T10:11:47.6669870Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T10:11:47.6702283Z http.https://github.com/.extraheader 2025-12-04T10:11:47.6745989Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T10:11:47.6786559Z http.https://github.com/.extraheader 2025-12-04T10:11:47.6828026Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T10:11:47.6870643Z http.https://github.com/.extraheader 2025-12-04T10:11:47.6909121Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:11:47.6951996Z http.https://github.com/.extraheader 2025-12-04T10:11:47.6996668Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:11:47.7037273Z http.https://github.com/.extraheader 2025-12-04T10:11:47.7083085Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T10:11:47.7123900Z http.https://github.com/.extraheader 2025-12-04T10:11:47.7164441Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T10:11:47.7206020Z http.https://github.com/.extraheader 2025-12-04T10:11:47.7249714Z Entering 'third_party/kleidiai' 2025-12-04T10:11:47.7293884Z http.https://github.com/.extraheader 2025-12-04T10:11:47.7335125Z Entering 'third_party/mimalloc' 2025-12-04T10:11:47.7381804Z http.https://github.com/.extraheader 2025-12-04T10:11:47.7427707Z Entering 'third_party/nlohmann' 2025-12-04T10:11:47.7470341Z http.https://github.com/.extraheader 2025-12-04T10:11:47.7522871Z Entering 'third_party/onnx' 2025-12-04T10:11:47.7570453Z http.https://github.com/.extraheader 2025-12-04T10:11:47.7630398Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T10:11:47.7671220Z http.https://github.com/.extraheader 2025-12-04T10:11:47.7717395Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T10:11:47.7760371Z http.https://github.com/.extraheader 2025-12-04T10:11:47.7799513Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T10:11:47.7841287Z http.https://github.com/.extraheader 2025-12-04T10:11:47.7887062Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T10:11:47.7927640Z http.https://github.com/.extraheader 2025-12-04T10:11:47.7966653Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T10:11:47.8012439Z http.https://github.com/.extraheader 2025-12-04T10:11:47.8057484Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T10:11:47.8102568Z http.https://github.com/.extraheader 2025-12-04T10:11:47.8144944Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T10:11:47.8185892Z http.https://github.com/.extraheader 2025-12-04T10:11:47.8225744Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T10:11:47.8268924Z http.https://github.com/.extraheader 2025-12-04T10:11:47.8308281Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T10:11:47.8349782Z http.https://github.com/.extraheader 2025-12-04T10:11:47.8386885Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:11:47.8429534Z http.https://github.com/.extraheader 2025-12-04T10:11:47.8470253Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:11:47.8512508Z http.https://github.com/.extraheader 2025-12-04T10:11:47.8558936Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T10:11:47.8601157Z http.https://github.com/.extraheader 2025-12-04T10:11:47.8659488Z Entering 'third_party/pocketfft' 2025-12-04T10:11:47.8703870Z http.https://github.com/.extraheader 2025-12-04T10:11:47.8745374Z Entering 'third_party/protobuf' 2025-12-04T10:11:47.8791610Z http.https://github.com/.extraheader 2025-12-04T10:11:47.8831342Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T10:11:47.8871800Z http.https://github.com/.extraheader 2025-12-04T10:11:47.8916390Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T10:11:47.8956574Z http.https://github.com/.extraheader 2025-12-04T10:11:47.8999380Z Entering 'third_party/psimd' 2025-12-04T10:11:47.9041730Z http.https://github.com/.extraheader 2025-12-04T10:11:47.9082637Z Entering 'third_party/pthreadpool' 2025-12-04T10:11:47.9123758Z http.https://github.com/.extraheader 2025-12-04T10:11:47.9163688Z Entering 'third_party/pybind11' 2025-12-04T10:11:47.9205515Z http.https://github.com/.extraheader 2025-12-04T10:11:47.9245708Z Entering 'third_party/python-peachpy' 2025-12-04T10:11:47.9291537Z http.https://github.com/.extraheader 2025-12-04T10:11:47.9330902Z Entering 'third_party/sleef' 2025-12-04T10:11:47.9371754Z http.https://github.com/.extraheader 2025-12-04T10:11:47.9413872Z Entering 'third_party/tensorpipe' 2025-12-04T10:11:47.9455672Z http.https://github.com/.extraheader 2025-12-04T10:11:47.9499692Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T10:11:47.9541744Z http.https://github.com/.extraheader 2025-12-04T10:11:47.9579648Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T10:11:47.9621485Z http.https://github.com/.extraheader 2025-12-04T10:11:47.9662073Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T10:11:47.9702568Z http.https://github.com/.extraheader 2025-12-04T10:11:47.9745690Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T10:11:47.9786333Z http.https://github.com/.extraheader 2025-12-04T10:11:47.9824350Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T10:11:47.9866292Z http.https://github.com/.extraheader 2025-12-04T10:11:47.9932187Z [command]/usr/bin/git config --local --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:47.9962954Z [command]/usr/bin/git submodule foreach --recursive git config --local --show-origin --name-only --get-regexp remote.origin.url 2025-12-04T10:11:48.0297716Z Entering 'android/libs/fbjni' 2025-12-04T10:11:48.0325820Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-12-04T10:11:48.0345853Z Entering 'third_party/FP16' 2025-12-04T10:11:48.0374886Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-12-04T10:11:48.0394879Z Entering 'third_party/FXdiv' 2025-12-04T10:11:48.0426985Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-12-04T10:11:48.0446541Z Entering 'third_party/NNPACK' 2025-12-04T10:11:48.0474671Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-12-04T10:11:48.0494956Z Entering 'third_party/NVTX' 2025-12-04T10:11:48.0525739Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-12-04T10:11:48.0547145Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T10:11:48.0575937Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-12-04T10:11:48.0596174Z Entering 'third_party/XNNPACK' 2025-12-04T10:11:48.0627489Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-12-04T10:11:48.0660782Z Entering 'third_party/aiter' 2025-12-04T10:11:48.0687961Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-12-04T10:11:48.0707749Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T10:11:48.0735504Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-12-04T10:11:48.0763754Z Entering 'third_party/benchmark' 2025-12-04T10:11:48.0792457Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-12-04T10:11:48.0814230Z Entering 'third_party/composable_kernel' 2025-12-04T10:11:48.0843008Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-12-04T10:11:48.0871599Z Entering 'third_party/cpp-httplib' 2025-12-04T10:11:48.0899173Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-12-04T10:11:48.0919663Z Entering 'third_party/cpuinfo' 2025-12-04T10:11:48.0947088Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-12-04T10:11:48.0967617Z Entering 'third_party/cudnn_frontend' 2025-12-04T10:11:48.0995404Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-12-04T10:11:48.1016900Z Entering 'third_party/cutlass' 2025-12-04T10:11:48.1044984Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-12-04T10:11:48.1073787Z Entering 'third_party/fbgemm' 2025-12-04T10:11:48.1102748Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-12-04T10:11:48.1124575Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T10:11:48.1152592Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-12-04T10:11:48.1172350Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T10:11:48.1199668Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-12-04T10:11:48.1227064Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T10:11:48.1254534Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-12-04T10:11:48.1274735Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T10:11:48.1303079Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-12-04T10:11:48.1333654Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T10:11:48.1361737Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-12-04T10:11:48.1379864Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T10:11:48.1409243Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-12-04T10:11:48.1427857Z Entering 'third_party/fbgemm/external/json' 2025-12-04T10:11:48.1455299Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-12-04T10:11:48.1481873Z Entering 'third_party/flash-attention' 2025-12-04T10:11:48.1512172Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-12-04T10:11:48.1532410Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T10:11:48.1560559Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-12-04T10:11:48.1585071Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T10:11:48.1613246Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-12-04T10:11:48.1644126Z Entering 'third_party/flatbuffers' 2025-12-04T10:11:48.1673086Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-12-04T10:11:48.1697294Z Entering 'third_party/fmt' 2025-12-04T10:11:48.1724954Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-12-04T10:11:48.1745847Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T10:11:48.1774352Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-12-04T10:11:48.1795044Z Entering 'third_party/gloo' 2025-12-04T10:11:48.1827057Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-12-04T10:11:48.1847377Z Entering 'third_party/googletest' 2025-12-04T10:11:48.1875485Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:11:48.1896756Z Entering 'third_party/ideep' 2025-12-04T10:11:48.1924011Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-12-04T10:11:48.1942935Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T10:11:48.1968963Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-12-04T10:11:48.1996193Z Entering 'third_party/ittapi' 2025-12-04T10:11:48.2027866Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-12-04T10:11:48.2047740Z Entering 'third_party/kineto' 2025-12-04T10:11:48.2076033Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-12-04T10:11:48.2095514Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T10:11:48.2123951Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-12-04T10:11:48.2142694Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T10:11:48.2170871Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-12-04T10:11:48.2190838Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T10:11:48.2220026Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-12-04T10:11:48.2238864Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T10:11:48.2266443Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-12-04T10:11:48.2286387Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T10:11:48.2315146Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-12-04T10:11:48.2333539Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T10:11:48.2362147Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-12-04T10:11:48.2383475Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T10:11:48.2410622Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-12-04T10:11:48.2429437Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T10:11:48.2457296Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:11:48.2477471Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T10:11:48.2505391Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-12-04T10:11:48.2527200Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T10:11:48.2557788Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-12-04T10:11:48.2577107Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T10:11:48.2606228Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T10:11:48.2624440Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:11:48.2653448Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T10:11:48.2675718Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:11:48.2707448Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T10:11:48.2732557Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T10:11:48.2759628Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-12-04T10:11:48.2778632Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T10:11:48.2804170Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-12-04T10:11:48.2827097Z Entering 'third_party/kleidiai' 2025-12-04T10:11:48.2855418Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-12-04T10:11:48.2876502Z Entering 'third_party/mimalloc' 2025-12-04T10:11:48.2908282Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-12-04T10:11:48.2928270Z Entering 'third_party/nlohmann' 2025-12-04T10:11:48.2956930Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-12-04T10:11:48.2978542Z Entering 'third_party/onnx' 2025-12-04T10:11:48.3007443Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-12-04T10:11:48.3041573Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T10:11:48.3068325Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-12-04T10:11:48.3092194Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T10:11:48.3120110Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-12-04T10:11:48.3140032Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T10:11:48.3167200Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-12-04T10:11:48.3186736Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T10:11:48.3212589Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:11:48.3232693Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T10:11:48.3259278Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-12-04T10:11:48.3278098Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T10:11:48.3304276Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-12-04T10:11:48.3325262Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T10:11:48.3356357Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-12-04T10:11:48.3375787Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T10:11:48.3401531Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-12-04T10:11:48.3419923Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T10:11:48.3446559Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T10:11:48.3464476Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:11:48.3490233Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T10:11:48.3511237Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:11:48.3537607Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T10:11:48.3559022Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T10:11:48.3584906Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-12-04T10:11:48.3622789Z Entering 'third_party/pocketfft' 2025-12-04T10:11:48.3651609Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-12-04T10:11:48.3672873Z Entering 'third_party/protobuf' 2025-12-04T10:11:48.3700158Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-12-04T10:11:48.3721212Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T10:11:48.3748611Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-12-04T10:11:48.3767325Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T10:11:48.3794877Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:11:48.3817852Z Entering 'third_party/psimd' 2025-12-04T10:11:48.3845714Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-12-04T10:11:48.3866169Z Entering 'third_party/pthreadpool' 2025-12-04T10:11:48.3895067Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-12-04T10:11:48.3916455Z Entering 'third_party/pybind11' 2025-12-04T10:11:48.3944890Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-12-04T10:11:48.3965537Z Entering 'third_party/python-peachpy' 2025-12-04T10:11:48.3996933Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-12-04T10:11:48.4018415Z Entering 'third_party/sleef' 2025-12-04T10:11:48.4046004Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-12-04T10:11:48.4066317Z Entering 'third_party/tensorpipe' 2025-12-04T10:11:48.4094808Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-12-04T10:11:48.4114343Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T10:11:48.4146616Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:11:48.4165741Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T10:11:48.4193357Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-12-04T10:11:48.4215255Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T10:11:48.4242799Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-12-04T10:11:48.4267340Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T10:11:48.4294547Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-12-04T10:11:48.4312712Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T10:11:48.4339138Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-12-04T10:11:48.4383564Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4412420Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4438452Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4464093Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4489688Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4514827Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4541363Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4564988Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4590826Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4617500Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4642936Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4670487Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4695015Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4720864Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4747411Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4772681Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4798684Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4823912Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4848583Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4874369Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4900907Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4926234Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4957584Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.4982503Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5009180Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5034002Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5059137Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5082686Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5109395Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5132978Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5158026Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5182134Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5209234Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5235159Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5261488Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5288942Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5315057Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5341596Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5365122Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5390958Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5416998Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5442545Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5470766Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5496408Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5531090Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5554646Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5580070Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5604769Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5630453Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5656132Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5680361Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5706550Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5730689Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5753314Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5778919Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5802533Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5828703Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5852771Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5878318Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5903232Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5930215Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5955130Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.5980924Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6008436Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6033820Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6058598Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6083885Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6110165Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6136847Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6162888Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6188898Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6213913Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6239919Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6264779Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6290412Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6316373Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6342193Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6368820Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6392812Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6419444Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6444269Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:11:48.6555758Z A job completed hook has been configured by the self-hosted runner administrator 2025-12-04T10:11:48.6573989Z ##[group]Run '/home/ec2-user/runner-scripts/after_job.sh' 2025-12-04T10:11:48.6580568Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:11:48.6580854Z ##[endgroup] 2025-12-04T10:11:48.6673968Z [!ALERT!] Swap in detected! [!ALERT!] 2025-12-04T10:11:58.1485624Z [!ALERT!] Swap out detected [!ALERT!] 2025-12-04T10:12:14.5873079Z Cleaning up orphan processes